[js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.527+0000 Starting JSTest jstests/concurrency/fsm_all_sharded_replication_legacy_config_servers_with_balancer.js... ./mongo --eval MongoRunner.dataDir = "/data/db/job0/mongorunner"; TestData = new Object(); TestData.wiredTigerEngineConfigString = ""; TestData.maxPort = 20249; TestData.wiredTigerIndexConfigString = ""; TestData.noJournal = false; TestData.testName = "fsm_all_sharded_replication_legacy_config_servers_with_balancer"; TestData.storageEngine = "wiredTiger"; TestData.minPort = 20010; TestData.noJournalPrealloc = true; TestData.wiredTigerCollectionConfigString = ""; MongoRunner.dataPath = "/data/db/job0/mongorunner/" --readMode commands --nodb jstests/concurrency/fsm_all_sharded_replication_legacy_config_servers_with_balancer.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.537+0000 JSTest jstests/concurrency/fsm_all_sharded_replication_legacy_config_servers_with_balancer.js started with pid 12867. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.554+0000 MongoDB shell version: 3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.595+0000 /data/db/job0/mongorunner/ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.788+0000 Replica set test! [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.788+0000 ReplSetTest Starting Set [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.789+0000 ReplSetTest n is : 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.789+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.789+0000 "useHostName" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "oplogSize" : 1024, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "keyFile" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "port" : 20010, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "noprealloc" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "smallfiles" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "replSet" : "test-rs0", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "dbpath" : "$set-$node", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.790+0000 "useHostname" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "noJournalPrealloc" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "pathOpts" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "testName" : "test", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "shard" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "node" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "set" : "test-rs0" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "verbose" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 "restart" : undefined [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.791+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.792+0000 ReplSetTest Starting.... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.793+0000 Resetting db path '/data/db/job0/mongorunner/test-rs0-0' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.799+0000 2015-11-23T18:44:03.799+0000 I - [thread1] shell: started program (sh12880): /data/mci/src/mongod --oplogSize 1024 --port 20010 --noprealloc --smallfiles --replSet test-rs0 --dbpath /data/db/job0/mongorunner/test-rs0-0 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.801+0000 2015-11-23T18:44:03.800+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.829+0000 d20010| note: noprealloc may hurt performance in many applications [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.839+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] MongoDB starting : pid=12880 port=20010 dbpath=/data/db/job0/mongorunner/test-rs0-0 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.839+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.840+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.840+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.840+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.841+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.841+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.841+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.841+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.841+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.842+0000 d20010| 2015-11-23T18:44:03.839+0000 I CONTROL [initandlisten] options: { net: { port: 20010 }, nopreallocj: true, replication: { oplogSizeMB: 1024, replSet: "test-rs0" }, setParameter: { enableTestCommands: "1" }, storage: { dbPath: "/data/db/job0/mongorunner/test-rs0-0", engine: "wiredTiger", mmapv1: { preallocDataFiles: false, smallFiles: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:03.860+0000 d20010| 2015-11-23T18:44:03.859+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:04.002+0000 2015-11-23T18:44:04.002+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:04.203+0000 2015-11-23T18:44:04.203+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:04.405+0000 2015-11-23T18:44:04.404+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:04.606+0000 2015-11-23T18:44:04.605+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:04.807+0000 2015-11-23T18:44:04.806+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.007+0000 2015-11-23T18:44:05.007+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.208+0000 2015-11-23T18:44:05.207+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.398+0000 d20010| 2015-11-23T18:44:05.398+0000 W STORAGE [initandlisten] Detected configuration for non-active storage engine mmapv1 when current storage engine is wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.399+0000 d20010| 2015-11-23T18:44:05.398+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.399+0000 d20010| 2015-11-23T18:44:05.398+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.399+0000 d20010| 2015-11-23T18:44:05.398+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.400+0000 d20010| 2015-11-23T18:44:05.398+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.409+0000 2015-11-23T18:44:05.409+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.610+0000 2015-11-23T18:44:05.610+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.811+0000 2015-11-23T18:44:05.811+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.916+0000 d20010| 2015-11-23T18:44:05.915+0000 I REPL [initandlisten] Did not find local voted for document at startup; NoMatchingDocument Did not find replica set lastVote document in local.replset.election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.916+0000 d20010| 2015-11-23T18:44:05.915+0000 I REPL [initandlisten] Did not find local replica set configuration document at startup; NoMatchingDocument Did not find replica set configuration document in local.system.replset [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.917+0000 d20010| 2015-11-23T18:44:05.916+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-rs0-0/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:05.917+0000 d20010| 2015-11-23T18:44:05.916+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.012+0000 2015-11-23T18:44:06.012+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.213+0000 2015-11-23T18:44:06.212+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.413+0000 2015-11-23T18:44:06.413+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.416+0000 d20010| 2015-11-23T18:44:06.416+0000 I NETWORK [initandlisten] waiting for connections on port 20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.614+0000 d20010| 2015-11-23T18:44:06.614+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:34887 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.615+0000 [ connection to ip-10-99-163-247:20010 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.615+0000 ReplSetTest n is : 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.617+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.617+0000 "useHostName" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "oplogSize" : 1024, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "keyFile" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "port" : 20011, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "noprealloc" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "smallfiles" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "replSet" : "test-rs0", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "dbpath" : "$set-$node", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "useHostname" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "noJournalPrealloc" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.618+0000 "pathOpts" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 "testName" : "test", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 "shard" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 "node" : 1, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 "set" : "test-rs0" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 "verbose" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 "restart" : undefined [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.619+0000 ReplSetTest Starting.... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.620+0000 Resetting db path '/data/db/job0/mongorunner/test-rs0-1' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.622+0000 2015-11-23T18:44:06.621+0000 I - [thread1] shell: started program (sh12919): /data/mci/src/mongod --oplogSize 1024 --port 20011 --noprealloc --smallfiles --replSet test-rs0 --dbpath /data/db/job0/mongorunner/test-rs0-1 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.623+0000 2015-11-23T18:44:06.623+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.651+0000 d20011| note: noprealloc may hurt performance in many applications [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.661+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] MongoDB starting : pid=12919 port=20011 dbpath=/data/db/job0/mongorunner/test-rs0-1 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.661+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.662+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.662+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.662+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.663+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.663+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.663+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.664+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.664+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.665+0000 d20011| 2015-11-23T18:44:06.661+0000 I CONTROL [initandlisten] options: { net: { port: 20011 }, nopreallocj: true, replication: { oplogSizeMB: 1024, replSet: "test-rs0" }, setParameter: { enableTestCommands: "1" }, storage: { dbPath: "/data/db/job0/mongorunner/test-rs0-1", engine: "wiredTiger", mmapv1: { preallocDataFiles: false, smallFiles: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.682+0000 d20011| 2015-11-23T18:44:06.681+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:06.825+0000 2015-11-23T18:44:06.824+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:07.025+0000 2015-11-23T18:44:07.025+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:07.226+0000 2015-11-23T18:44:07.226+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:07.427+0000 2015-11-23T18:44:07.427+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:07.628+0000 2015-11-23T18:44:07.628+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:07.829+0000 2015-11-23T18:44:07.828+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.029+0000 2015-11-23T18:44:08.029+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.117+0000 d20011| 2015-11-23T18:44:08.117+0000 W STORAGE [initandlisten] Detected configuration for non-active storage engine mmapv1 when current storage engine is wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.118+0000 d20011| 2015-11-23T18:44:08.117+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.118+0000 d20011| 2015-11-23T18:44:08.117+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.118+0000 d20011| 2015-11-23T18:44:08.117+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.119+0000 d20011| 2015-11-23T18:44:08.117+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.230+0000 2015-11-23T18:44:08.229+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.430+0000 2015-11-23T18:44:08.430+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.528+0000 d20011| 2015-11-23T18:44:08.528+0000 I REPL [initandlisten] Did not find local voted for document at startup; NoMatchingDocument Did not find replica set lastVote document in local.replset.election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.529+0000 d20011| 2015-11-23T18:44:08.528+0000 I REPL [initandlisten] Did not find local replica set configuration document at startup; NoMatchingDocument Did not find replica set configuration document in local.system.replset [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.529+0000 d20011| 2015-11-23T18:44:08.528+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-rs0-1/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.530+0000 d20011| 2015-11-23T18:44:08.528+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.631+0000 2015-11-23T18:44:08.631+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.831+0000 2015-11-23T18:44:08.831+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:08.911+0000 d20011| 2015-11-23T18:44:08.911+0000 I NETWORK [initandlisten] waiting for connections on port 20011 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.032+0000 d20011| 2015-11-23T18:44:09.032+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:39276 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.033+0000 [ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.033+0000 connection to ip-10-99-163-247:20010, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.033+0000 connection to ip-10-99-163-247:20011 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.033+0000 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.033+0000 ReplSetTest n is : 2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.033+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.033+0000 "useHostName" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "oplogSize" : 1024, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "keyFile" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "port" : 20012, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "noprealloc" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "smallfiles" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "replSet" : "test-rs0", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "dbpath" : "$set-$node", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "useHostname" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.034+0000 "noJournalPrealloc" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 "pathOpts" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 "testName" : "test", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 "shard" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 "node" : 2, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 "set" : "test-rs0" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 "verbose" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 "restart" : undefined [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.035+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.036+0000 ReplSetTest Starting.... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.036+0000 Resetting db path '/data/db/job0/mongorunner/test-rs0-2' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.040+0000 2015-11-23T18:44:09.039+0000 I - [thread1] shell: started program (sh12956): /data/mci/src/mongod --oplogSize 1024 --port 20012 --noprealloc --smallfiles --replSet test-rs0 --dbpath /data/db/job0/mongorunner/test-rs0-2 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.041+0000 2015-11-23T18:44:09.040+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20012, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.068+0000 d20012| note: noprealloc may hurt performance in many applications [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.078+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] MongoDB starting : pid=12956 port=20012 dbpath=/data/db/job0/mongorunner/test-rs0-2 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.079+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.079+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.079+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.080+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.080+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.080+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.080+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.080+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.081+0000 d20012| 2015-11-23T18:44:09.078+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.081+0000 d20012| 2015-11-23T18:44:09.079+0000 I CONTROL [initandlisten] options: { net: { port: 20012 }, nopreallocj: true, replication: { oplogSizeMB: 1024, replSet: "test-rs0" }, setParameter: { enableTestCommands: "1" }, storage: { dbPath: "/data/db/job0/mongorunner/test-rs0-2", engine: "wiredTiger", mmapv1: { preallocDataFiles: false, smallFiles: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.099+0000 d20012| 2015-11-23T18:44:09.098+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.242+0000 2015-11-23T18:44:09.241+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20012, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.443+0000 2015-11-23T18:44:09.442+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20012, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.643+0000 2015-11-23T18:44:09.643+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20012, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.653+0000 d20012| 2015-11-23T18:44:09.652+0000 W STORAGE [initandlisten] Detected configuration for non-active storage engine mmapv1 when current storage engine is wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.653+0000 d20012| 2015-11-23T18:44:09.652+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.653+0000 d20012| 2015-11-23T18:44:09.653+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.653+0000 d20012| 2015-11-23T18:44:09.653+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.654+0000 d20012| 2015-11-23T18:44:09.653+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.664+0000 d20012| 2015-11-23T18:44:09.664+0000 I REPL [initandlisten] Did not find local voted for document at startup; NoMatchingDocument Did not find replica set lastVote document in local.replset.election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.665+0000 d20012| 2015-11-23T18:44:09.664+0000 I REPL [initandlisten] Did not find local replica set configuration document at startup; NoMatchingDocument Did not find replica set configuration document in local.system.replset [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.665+0000 d20012| 2015-11-23T18:44:09.664+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-rs0-2/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.666+0000 d20012| 2015-11-23T18:44:09.664+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.678+0000 d20012| 2015-11-23T18:44:09.677+0000 I NETWORK [initandlisten] waiting for connections on port 20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.844+0000 d20012| 2015-11-23T18:44:09.843+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:55048 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.845+0000 [ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.845+0000 connection to ip-10-99-163-247:20010, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.845+0000 connection to ip-10-99-163-247:20011, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 connection to ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 "replSetInitiate" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 "_id" : "test-rs0", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 "members" : [ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 "_id" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 "host" : "ip-10-99-163-247:20010" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.846+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.848+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.848+0000 "_id" : 1, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.848+0000 "host" : "ip-10-99-163-247:20011" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.848+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.848+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.848+0000 "_id" : 2, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.848+0000 "host" : "ip-10-99-163-247:20012" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.849+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.849+0000 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.849+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.849+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.849+0000 d20010| 2015-11-23T18:44:09.846+0000 I REPL [conn1] replSetInitiate admin command received from client [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.849+0000 d20011| 2015-11-23T18:44:09.847+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49607 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.850+0000 d20011| 2015-11-23T18:44:09.847+0000 I NETWORK [conn2] end connection 10.99.163.247:49607 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.850+0000 d20012| 2015-11-23T18:44:09.848+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:45850 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.850+0000 d20010| 2015-11-23T18:44:09.848+0000 I REPL [conn1] replSetInitiate config object with 3 members parses ok [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.851+0000 d20012| 2015-11-23T18:44:09.848+0000 I NETWORK [conn2] end connection 10.99.163.247:45850 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.851+0000 d20011| 2015-11-23T18:44:09.849+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49609 #3 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.852+0000 d20012| 2015-11-23T18:44:09.849+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:45852 #3 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.852+0000 d20010| 2015-11-23T18:44:09.849+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20011 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.852+0000 d20010| 2015-11-23T18:44:09.849+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.852+0000 d20010| 2015-11-23T18:44:09.850+0000 I REPL [conn1] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.853+0000 d20010| 2015-11-23T18:44:09.850+0000 I REPL [conn1] creating replication oplog of size: 1024MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.853+0000 d20010| 2015-11-23T18:44:09.851+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38653 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.853+0000 d20010| 2015-11-23T18:44:09.851+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38654 #3 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.853+0000 d20012| 2015-11-23T18:44:09.851+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.854+0000 d20011| 2015-11-23T18:44:09.852+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.854+0000 d20010| 2015-11-23T18:44:09.854+0000 I STORAGE [conn1] Starting WiredTigerRecordStoreThread local.oplog.rs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.855+0000 d20010| 2015-11-23T18:44:09.854+0000 I STORAGE [conn1] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.903+0000 d20010| 2015-11-23T18:44:09.903+0000 I REPL [conn1] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.914+0000 d20010| 2015-11-23T18:44:09.912+0000 I REPL [ReplicationExecutor] New replica set config in use: { _id: "test-rs0", version: 1, protocolVersion: 1, members: [ { _id: 0, host: "ip-10-99-163-247:20010", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 1, host: "ip-10-99-163-247:20011", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 2, host: "ip-10-99-163-247:20012", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 } ], settings: { chainingAllowed: true, heartbeatIntervalMillis: 2000, heartbeatTimeoutSecs: 10, electionTimeoutMillis: 10000, getLastErrorModes: {}, getLastErrorDefaults: { w: 1, wtimeout: 0 } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.915+0000 d20010| 2015-11-23T18:44:09.912+0000 I REPL [ReplicationExecutor] This node is ip-10-99-163-247:20010 in the config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.915+0000 d20010| 2015-11-23T18:44:09.912+0000 I REPL [ReplicationExecutor] transition to STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.915+0000 d20010| 2015-11-23T18:44:09.912+0000 I REPL [conn1] Starting replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.916+0000 d20010| 2015-11-23T18:44:09.912+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20011 is now in state STARTUP [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.916+0000 d20010| 2015-11-23T18:44:09.913+0000 I REPL [ReplicationExecutor] transition to RECOVERING [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.916+0000 d20010| 2015-11-23T18:44:09.914+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20012 is now in state STARTUP [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:09.916+0000 d20010| 2015-11-23T18:44:09.915+0000 I REPL [ReplicationExecutor] transition to SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.854+0000 d20010| 2015-11-23T18:44:11.853+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38655 #4 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.855+0000 d20010| 2015-11-23T18:44:11.854+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38656 #5 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.855+0000 d20010| 2015-11-23T18:44:11.854+0000 I NETWORK [conn4] end connection 10.99.163.247:38655 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.855+0000 d20011| 2015-11-23T18:44:11.854+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49615 #4 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.856+0000 d20010| 2015-11-23T18:44:11.854+0000 I NETWORK [conn5] end connection 10.99.163.247:38656 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.856+0000 d20011| 2015-11-23T18:44:11.855+0000 I NETWORK [conn4] end connection 10.99.163.247:49615 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.856+0000 d20012| 2015-11-23T18:44:11.855+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:45858 #4 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.857+0000 d20012| 2015-11-23T18:44:11.856+0000 I NETWORK [conn4] end connection 10.99.163.247:45858 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.873+0000 d20012| 2015-11-23T18:44:11.873+0000 I REPL [replExecDBWorker-0] Starting replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.873+0000 d20011| 2015-11-23T18:44:11.873+0000 I REPL [replExecDBWorker-0] Starting replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.874+0000 d20012| 2015-11-23T18:44:11.873+0000 W REPL [rsSync] did not receive a valid config yet [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.874+0000 d20011| 2015-11-23T18:44:11.873+0000 W REPL [rsSync] did not receive a valid config yet [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.875+0000 d20012| 2015-11-23T18:44:11.873+0000 I REPL [ReplicationExecutor] New replica set config in use: { _id: "test-rs0", version: 1, protocolVersion: 1, members: [ { _id: 0, host: "ip-10-99-163-247:20010", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 1, host: "ip-10-99-163-247:20011", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 2, host: "ip-10-99-163-247:20012", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 } ], settings: { chainingAllowed: true, heartbeatIntervalMillis: 2000, heartbeatTimeoutSecs: 10, electionTimeoutMillis: 10000, getLastErrorModes: {}, getLastErrorDefaults: { w: 1, wtimeout: 0 } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.876+0000 d20012| 2015-11-23T18:44:11.873+0000 I REPL [ReplicationExecutor] This node is ip-10-99-163-247:20012 in the config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.876+0000 d20012| 2015-11-23T18:44:11.873+0000 I REPL [ReplicationExecutor] transition to STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.876+0000 d20012| 2015-11-23T18:44:11.874+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20010 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.877+0000 d20011| 2015-11-23T18:44:11.874+0000 I REPL [ReplicationExecutor] New replica set config in use: { _id: "test-rs0", version: 1, protocolVersion: 1, members: [ { _id: 0, host: "ip-10-99-163-247:20010", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 1, host: "ip-10-99-163-247:20011", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 2, host: "ip-10-99-163-247:20012", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 } ], settings: { chainingAllowed: true, heartbeatIntervalMillis: 2000, heartbeatTimeoutSecs: 10, electionTimeoutMillis: 10000, getLastErrorModes: {}, getLastErrorDefaults: { w: 1, wtimeout: 0 } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.878+0000 d20011| 2015-11-23T18:44:11.874+0000 I REPL [ReplicationExecutor] This node is ip-10-99-163-247:20011 in the config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.878+0000 d20011| 2015-11-23T18:44:11.874+0000 I REPL [ReplicationExecutor] transition to STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.878+0000 d20011| 2015-11-23T18:44:11.874+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49617 #5 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.878+0000 d20011| 2015-11-23T18:44:11.874+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20010 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.879+0000 d20012| 2015-11-23T18:44:11.874+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20011 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.879+0000 d20012| 2015-11-23T18:44:11.875+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:45860 #5 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.879+0000 d20012| 2015-11-23T18:44:11.875+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20011 is now in state STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.879+0000 d20011| 2015-11-23T18:44:11.875+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.880+0000 d20011| 2015-11-23T18:44:11.875+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20012 is now in state STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.916+0000 d20010| 2015-11-23T18:44:11.916+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20011 is now in state STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:11.916+0000 d20010| 2015-11-23T18:44:11.916+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20012 is now in state STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.874+0000 d20012| 2015-11-23T18:44:12.873+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.875+0000 d20012| 2015-11-23T18:44:12.873+0000 I REPL [rsSync] creating replication oplog of size: 1024MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.875+0000 d20011| 2015-11-23T18:44:12.873+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.876+0000 d20011| 2015-11-23T18:44:12.873+0000 I REPL [rsSync] creating replication oplog of size: 1024MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.879+0000 d20011| 2015-11-23T18:44:12.879+0000 I STORAGE [rsSync] Starting WiredTigerRecordStoreThread local.oplog.rs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.879+0000 d20012| 2015-11-23T18:44:12.879+0000 I STORAGE [rsSync] Starting WiredTigerRecordStoreThread local.oplog.rs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.880+0000 d20012| 2015-11-23T18:44:12.879+0000 I STORAGE [rsSync] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.880+0000 d20011| 2015-11-23T18:44:12.879+0000 I STORAGE [rsSync] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.911+0000 d20012| 2015-11-23T18:44:12.911+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.911+0000 d20012| 2015-11-23T18:44:12.911+0000 I REPL [rsSync] initial sync pending [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.912+0000 d20011| 2015-11-23T18:44:12.911+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.912+0000 d20011| 2015-11-23T18:44:12.911+0000 I REPL [rsSync] initial sync pending [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.917+0000 d20012| 2015-11-23T18:44:12.916+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.918+0000 d20010| 2015-11-23T18:44:12.917+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38661 #6 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.918+0000 d20011| 2015-11-23T18:44:12.917+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.919+0000 d20010| 2015-11-23T18:44:12.918+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38662 #7 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.921+0000 d20012| 2015-11-23T18:44:12.920+0000 I REPL [rsSync] initial sync drop all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.921+0000 d20012| 2015-11-23T18:44:12.920+0000 I STORAGE [rsSync] dropAllDatabasesExceptLocal 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.921+0000 d20012| 2015-11-23T18:44:12.920+0000 I REPL [rsSync] initial sync clone all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.922+0000 d20012| 2015-11-23T18:44:12.921+0000 I REPL [rsSync] initial sync data copy, starting syncup [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.922+0000 d20012| 2015-11-23T18:44:12.921+0000 I REPL [rsSync] oplog sync 1 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.922+0000 d20012| 2015-11-23T18:44:12.921+0000 I REPL [rsSync] oplog sync 2 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.922+0000 d20011| 2015-11-23T18:44:12.922+0000 I REPL [rsSync] initial sync drop all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.922+0000 d20011| 2015-11-23T18:44:12.922+0000 I STORAGE [rsSync] dropAllDatabasesExceptLocal 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.923+0000 d20011| 2015-11-23T18:44:12.922+0000 I REPL [rsSync] initial sync clone all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.923+0000 d20012| 2015-11-23T18:44:12.923+0000 I REPL [rsSync] initial sync building indexes [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.923+0000 d20012| 2015-11-23T18:44:12.923+0000 I REPL [rsSync] oplog sync 3 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.924+0000 d20011| 2015-11-23T18:44:12.923+0000 I REPL [rsSync] initial sync data copy, starting syncup [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.924+0000 d20011| 2015-11-23T18:44:12.923+0000 I REPL [rsSync] oplog sync 1 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.924+0000 d20011| 2015-11-23T18:44:12.924+0000 I REPL [rsSync] oplog sync 2 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.925+0000 d20011| 2015-11-23T18:44:12.924+0000 I REPL [rsSync] initial sync building indexes [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.925+0000 d20011| 2015-11-23T18:44:12.924+0000 I REPL [rsSync] oplog sync 3 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.926+0000 d20012| 2015-11-23T18:44:12.925+0000 I REPL [rsSync] initial sync finishing up [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.926+0000 d20012| 2015-11-23T18:44:12.925+0000 I REPL [rsSync] set minValid=(term: -1, timestamp: Nov 23 18:44:09:1) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.927+0000 d20012| 2015-11-23T18:44:12.926+0000 I REPL [rsSync] initial sync done [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.927+0000 d20011| 2015-11-23T18:44:12.927+0000 I REPL [rsSync] initial sync finishing up [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.927+0000 d20011| 2015-11-23T18:44:12.927+0000 I REPL [rsSync] set minValid=(term: -1, timestamp: Nov 23 18:44:09:1) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.928+0000 d20011| 2015-11-23T18:44:12.928+0000 I REPL [rsSync] initial sync done [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.930+0000 d20012| 2015-11-23T18:44:12.930+0000 I REPL [ReplicationExecutor] transition to RECOVERING [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.930+0000 d20010| 2015-11-23T18:44:12.930+0000 I NETWORK [conn6] end connection 10.99.163.247:38661 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.931+0000 d20011| 2015-11-23T18:44:12.931+0000 I REPL [ReplicationExecutor] transition to RECOVERING [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.931+0000 d20010| 2015-11-23T18:44:12.931+0000 I NETWORK [conn7] end connection 10.99.163.247:38662 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.932+0000 d20012| 2015-11-23T18:44:12.932+0000 I REPL [ReplicationExecutor] transition to SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:12.933+0000 d20011| 2015-11-23T18:44:12.933+0000 I REPL [ReplicationExecutor] transition to SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:13.874+0000 d20012| 2015-11-23T18:44:13.874+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:13.874+0000 d20012| 2015-11-23T18:44:13.874+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20011 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:13.875+0000 d20011| 2015-11-23T18:44:13.875+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:13.876+0000 d20011| 2015-11-23T18:44:13.875+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20012 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:13.917+0000 d20010| 2015-11-23T18:44:13.916+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20011 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:13.917+0000 d20010| 2015-11-23T18:44:13.916+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20012 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.448+0000 d20010| 2015-11-23T18:44:20.447+0000 I REPL [ReplicationExecutor] Starting an election, since we've seen no PRIMARY in the past 10000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.448+0000 d20010| 2015-11-23T18:44:20.447+0000 I REPL [ReplicationExecutor] conducting a dry run election to see if we could be elected [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.459+0000 d20010| 2015-11-23T18:44:20.458+0000 I REPL [ReplicationExecutor] dry election run succeeded, running for election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.462+0000 d20011| 2015-11-23T18:44:20.462+0000 I NETWORK [conn3] end connection 10.99.163.247:49609 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.470+0000 d20010| 2015-11-23T18:44:20.470+0000 I REPL [ReplicationExecutor] election succeeded, assuming primary role in term 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.470+0000 d20010| 2015-11-23T18:44:20.470+0000 I REPL [ReplicationExecutor] transition to PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.471+0000 d20011| 2015-11-23T18:44:20.470+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49622 #6 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.471+0000 d20010| 2015-11-23T18:44:20.470+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20011 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.471+0000 d20011| 2015-11-23T18:44:20.471+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49623 #7 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.472+0000 d20010| 2015-11-23T18:44:20.471+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20011 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.918+0000 d20010| 2015-11-23T18:44:20.918+0000 I REPL [rsSync] transition to primary complete; database writes are now permitted [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.980+0000 Replica set test! [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.980+0000 ReplSetTest Starting Set [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.980+0000 ReplSetTest n is : 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.980+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.980+0000 "useHostName" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.980+0000 "oplogSize" : 1024, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.981+0000 "keyFile" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.981+0000 "port" : 20013, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.981+0000 "noprealloc" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.981+0000 "smallfiles" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.981+0000 "replSet" : "test-rs1", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.981+0000 "dbpath" : "$set-$node", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.981+0000 "useHostname" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.982+0000 "noJournalPrealloc" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.982+0000 "pathOpts" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.982+0000 "testName" : "test", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.982+0000 "shard" : 1, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.982+0000 "node" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.982+0000 "set" : "test-rs1" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.983+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.983+0000 "verbose" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.983+0000 "restart" : undefined [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.983+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.983+0000 ReplSetTest Starting.... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.983+0000 Resetting db path '/data/db/job0/mongorunner/test-rs1-0' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.986+0000 2015-11-23T18:44:20.985+0000 I - [thread1] shell: started program (sh13260): /data/mci/src/mongod --oplogSize 1024 --port 20013 --noprealloc --smallfiles --replSet test-rs1 --dbpath /data/db/job0/mongorunner/test-rs1-0 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:20.986+0000 2015-11-23T18:44:20.986+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.016+0000 d20013| note: noprealloc may hurt performance in many applications [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.026+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] MongoDB starting : pid=13260 port=20013 dbpath=/data/db/job0/mongorunner/test-rs1-0 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.026+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.027+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.027+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.027+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.027+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.028+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.028+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.028+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.028+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.029+0000 d20013| 2015-11-23T18:44:21.026+0000 I CONTROL [initandlisten] options: { net: { port: 20013 }, nopreallocj: true, replication: { oplogSizeMB: 1024, replSet: "test-rs1" }, setParameter: { enableTestCommands: "1" }, storage: { dbPath: "/data/db/job0/mongorunner/test-rs1-0", engine: "wiredTiger", mmapv1: { preallocDataFiles: false, smallFiles: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.046+0000 d20013| 2015-11-23T18:44:21.046+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.125+0000 d20013| 2015-11-23T18:44:21.124+0000 W STORAGE [initandlisten] Detected configuration for non-active storage engine mmapv1 when current storage engine is wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.125+0000 d20013| 2015-11-23T18:44:21.124+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.125+0000 d20013| 2015-11-23T18:44:21.124+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.125+0000 d20013| 2015-11-23T18:44:21.125+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.126+0000 d20013| 2015-11-23T18:44:21.125+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.129+0000 d20013| 2015-11-23T18:44:21.129+0000 I REPL [initandlisten] Did not find local voted for document at startup; NoMatchingDocument Did not find replica set lastVote document in local.replset.election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.130+0000 d20013| 2015-11-23T18:44:21.129+0000 I REPL [initandlisten] Did not find local replica set configuration document at startup; NoMatchingDocument Did not find replica set configuration document in local.system.replset [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.130+0000 d20013| 2015-11-23T18:44:21.129+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-rs1-0/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.131+0000 d20013| 2015-11-23T18:44:21.129+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.142+0000 d20013| 2015-11-23T18:44:21.142+0000 I NETWORK [initandlisten] waiting for connections on port 20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.188+0000 d20013| 2015-11-23T18:44:21.187+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:42676 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.188+0000 [ connection to ip-10-99-163-247:20013 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.188+0000 ReplSetTest n is : 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.189+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.189+0000 "useHostName" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.189+0000 "oplogSize" : 1024, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.189+0000 "keyFile" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.189+0000 "port" : 20014, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.189+0000 "noprealloc" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.190+0000 "smallfiles" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.190+0000 "replSet" : "test-rs1", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.190+0000 "dbpath" : "$set-$node", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.190+0000 "useHostname" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.190+0000 "noJournalPrealloc" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.191+0000 "pathOpts" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.191+0000 "testName" : "test", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.191+0000 "shard" : 1, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.191+0000 "node" : 1, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.191+0000 "set" : "test-rs1" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.191+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.191+0000 "verbose" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.192+0000 "restart" : undefined [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.192+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.192+0000 ReplSetTest Starting.... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.192+0000 Resetting db path '/data/db/job0/mongorunner/test-rs1-1' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.196+0000 2015-11-23T18:44:21.195+0000 I - [thread1] shell: started program (sh13286): /data/mci/src/mongod --oplogSize 1024 --port 20014 --noprealloc --smallfiles --replSet test-rs1 --dbpath /data/db/job0/mongorunner/test-rs1-1 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.197+0000 2015-11-23T18:44:21.196+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20014, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.225+0000 d20014| note: noprealloc may hurt performance in many applications [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.235+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] MongoDB starting : pid=13286 port=20014 dbpath=/data/db/job0/mongorunner/test-rs1-1 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.235+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.236+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.236+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.236+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.236+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.236+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.237+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.237+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.237+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.238+0000 d20014| 2015-11-23T18:44:21.235+0000 I CONTROL [initandlisten] options: { net: { port: 20014 }, nopreallocj: true, replication: { oplogSizeMB: 1024, replSet: "test-rs1" }, setParameter: { enableTestCommands: "1" }, storage: { dbPath: "/data/db/job0/mongorunner/test-rs1-1", engine: "wiredTiger", mmapv1: { preallocDataFiles: false, smallFiles: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.255+0000 d20014| 2015-11-23T18:44:21.255+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.329+0000 d20014| 2015-11-23T18:44:21.328+0000 W STORAGE [initandlisten] Detected configuration for non-active storage engine mmapv1 when current storage engine is wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.329+0000 d20014| 2015-11-23T18:44:21.328+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.329+0000 d20014| 2015-11-23T18:44:21.328+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.330+0000 d20014| 2015-11-23T18:44:21.329+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.330+0000 d20014| 2015-11-23T18:44:21.329+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.334+0000 d20014| 2015-11-23T18:44:21.333+0000 I REPL [initandlisten] Did not find local voted for document at startup; NoMatchingDocument Did not find replica set lastVote document in local.replset.election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.334+0000 d20014| 2015-11-23T18:44:21.333+0000 I REPL [initandlisten] Did not find local replica set configuration document at startup; NoMatchingDocument Did not find replica set configuration document in local.system.replset [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.335+0000 d20014| 2015-11-23T18:44:21.334+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-rs1-1/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.335+0000 d20014| 2015-11-23T18:44:21.334+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.340+0000 d20014| 2015-11-23T18:44:21.340+0000 I NETWORK [initandlisten] waiting for connections on port 20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.397+0000 d20014| 2015-11-23T18:44:21.397+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:47948 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.399+0000 [ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.399+0000 connection to ip-10-99-163-247:20013, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.399+0000 connection to ip-10-99-163-247:20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.399+0000 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.399+0000 ReplSetTest n is : 2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.399+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.400+0000 "useHostName" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.400+0000 "oplogSize" : 1024, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.400+0000 "keyFile" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.400+0000 "port" : 20015, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.400+0000 "noprealloc" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.401+0000 "smallfiles" : "", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.401+0000 "replSet" : "test-rs1", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.401+0000 "dbpath" : "$set-$node", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.401+0000 "useHostname" : true, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.401+0000 "noJournalPrealloc" : undefined, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.401+0000 "pathOpts" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 "testName" : "test", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 "shard" : 1, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 "node" : 2, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 "set" : "test-rs1" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 "verbose" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 "restart" : undefined [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.402+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.403+0000 ReplSetTest Starting.... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.403+0000 Resetting db path '/data/db/job0/mongorunner/test-rs1-2' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.407+0000 2015-11-23T18:44:21.406+0000 I - [thread1] shell: started program (sh13312): /data/mci/src/mongod --oplogSize 1024 --port 20015 --noprealloc --smallfiles --replSet test-rs1 --dbpath /data/db/job0/mongorunner/test-rs1-2 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.407+0000 2015-11-23T18:44:21.407+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20015, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.436+0000 d20015| note: noprealloc may hurt performance in many applications [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.445+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] MongoDB starting : pid=13312 port=20015 dbpath=/data/db/job0/mongorunner/test-rs1-2 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.446+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.446+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.446+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.447+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.447+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.447+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.447+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.447+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.448+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.448+0000 d20015| 2015-11-23T18:44:21.445+0000 I CONTROL [initandlisten] options: { net: { port: 20015 }, nopreallocj: true, replication: { oplogSizeMB: 1024, replSet: "test-rs1" }, setParameter: { enableTestCommands: "1" }, storage: { dbPath: "/data/db/job0/mongorunner/test-rs1-2", engine: "wiredTiger", mmapv1: { preallocDataFiles: false, smallFiles: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.466+0000 d20015| 2015-11-23T18:44:21.465+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.537+0000 d20015| 2015-11-23T18:44:21.537+0000 W STORAGE [initandlisten] Detected configuration for non-active storage engine mmapv1 when current storage engine is wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.538+0000 d20015| 2015-11-23T18:44:21.537+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.538+0000 d20015| 2015-11-23T18:44:21.537+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.538+0000 d20015| 2015-11-23T18:44:21.537+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.538+0000 d20015| 2015-11-23T18:44:21.538+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.542+0000 d20015| 2015-11-23T18:44:21.542+0000 I REPL [initandlisten] Did not find local voted for document at startup; NoMatchingDocument Did not find replica set lastVote document in local.replset.election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.543+0000 d20015| 2015-11-23T18:44:21.542+0000 I REPL [initandlisten] Did not find local replica set configuration document at startup; NoMatchingDocument Did not find replica set configuration document in local.system.replset [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.543+0000 d20015| 2015-11-23T18:44:21.542+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.543+0000 d20015| 2015-11-23T18:44:21.542+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-rs1-2/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.549+0000 d20015| 2015-11-23T18:44:21.548+0000 I NETWORK [initandlisten] waiting for connections on port 20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.608+0000 d20015| 2015-11-23T18:44:21.608+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:35712 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.609+0000 [ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 connection to ip-10-99-163-247:20013, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 connection to ip-10-99-163-247:20014, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 connection to ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 "replSetInitiate" : { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 "_id" : "test-rs1", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 "members" : [ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.610+0000 "_id" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 "host" : "ip-10-99-163-247:20013" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 "_id" : 1, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 "host" : "ip-10-99-163-247:20014" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 }, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.611+0000 "_id" : 2, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.612+0000 "host" : "ip-10-99-163-247:20015" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.612+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.612+0000 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.612+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.612+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.612+0000 d20013| 2015-11-23T18:44:21.610+0000 I REPL [conn1] replSetInitiate admin command received from client [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.612+0000 d20014| 2015-11-23T18:44:21.611+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37192 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.613+0000 d20014| 2015-11-23T18:44:21.612+0000 I NETWORK [conn2] end connection 10.99.163.247:37192 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.614+0000 d20015| 2015-11-23T18:44:21.612+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56525 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.614+0000 d20013| 2015-11-23T18:44:21.612+0000 I REPL [conn1] replSetInitiate config object with 3 members parses ok [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.615+0000 d20015| 2015-11-23T18:44:21.613+0000 I NETWORK [conn2] end connection 10.99.163.247:56525 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.615+0000 d20014| 2015-11-23T18:44:21.613+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37194 #3 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.616+0000 d20015| 2015-11-23T18:44:21.613+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56527 #3 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.616+0000 d20013| 2015-11-23T18:44:21.614+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.617+0000 d20013| 2015-11-23T18:44:21.614+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.617+0000 d20013| 2015-11-23T18:44:21.614+0000 I REPL [conn1] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.617+0000 d20013| 2015-11-23T18:44:21.615+0000 I REPL [conn1] creating replication oplog of size: 1024MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.617+0000 d20013| 2015-11-23T18:44:21.615+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36258 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.618+0000 d20014| 2015-11-23T18:44:21.615+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.618+0000 d20013| 2015-11-23T18:44:21.615+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36259 #3 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.618+0000 d20015| 2015-11-23T18:44:21.616+0000 I ASIO [NetworkInterfaceASIO-Replication-0] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.618+0000 d20013| 2015-11-23T18:44:21.617+0000 I STORAGE [conn1] Starting WiredTigerRecordStoreThread local.oplog.rs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.619+0000 d20013| 2015-11-23T18:44:21.617+0000 I STORAGE [conn1] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.635+0000 d20013| 2015-11-23T18:44:21.635+0000 I REPL [conn1] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.647+0000 d20013| 2015-11-23T18:44:21.645+0000 I REPL [ReplicationExecutor] New replica set config in use: { _id: "test-rs1", version: 1, protocolVersion: 1, members: [ { _id: 0, host: "ip-10-99-163-247:20013", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 1, host: "ip-10-99-163-247:20014", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 2, host: "ip-10-99-163-247:20015", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 } ], settings: { chainingAllowed: true, heartbeatIntervalMillis: 2000, heartbeatTimeoutSecs: 10, electionTimeoutMillis: 10000, getLastErrorModes: {}, getLastErrorDefaults: { w: 1, wtimeout: 0 } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.647+0000 d20013| 2015-11-23T18:44:21.645+0000 I REPL [ReplicationExecutor] This node is ip-10-99-163-247:20013 in the config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.648+0000 d20013| 2015-11-23T18:44:21.645+0000 I REPL [ReplicationExecutor] transition to STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.648+0000 d20013| 2015-11-23T18:44:21.645+0000 I REPL [conn1] Starting replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.648+0000 d20013| 2015-11-23T18:44:21.645+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20014 is now in state STARTUP [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.649+0000 d20013| 2015-11-23T18:44:21.645+0000 I REPL [ReplicationExecutor] transition to RECOVERING [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.649+0000 d20013| 2015-11-23T18:44:21.645+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20015 is now in state STARTUP [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.649+0000 d20013| 2015-11-23T18:44:21.648+0000 I REPL [ReplicationExecutor] transition to SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.877+0000 d20012| 2015-11-23T18:44:21.876+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20010 is now in state PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:21.877+0000 d20011| 2015-11-23T18:44:21.877+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20010 is now in state PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.876+0000 d20012| 2015-11-23T18:44:22.875+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.877+0000 d20010| 2015-11-23T18:44:22.876+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38678 #8 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.877+0000 d20011| 2015-11-23T18:44:22.877+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.878+0000 d20010| 2015-11-23T18:44:22.877+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38679 #9 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.878+0000 d20012| 2015-11-23T18:44:22.878+0000 I REPL [SyncSourceFeedback] setting syncSourceFeedback to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.878+0000 d20010| 2015-11-23T18:44:22.878+0000 I NETWORK [conn8] end connection 10.99.163.247:38678 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.879+0000 d20010| 2015-11-23T18:44:22.878+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38680 #10 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.879+0000 d20010| 2015-11-23T18:44:22.878+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38681 #11 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.879+0000 d20012| 2015-11-23T18:44:22.879+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.880+0000 d20010| 2015-11-23T18:44:22.879+0000 I NETWORK [conn9] end connection 10.99.163.247:38679 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.880+0000 d20011| 2015-11-23T18:44:22.879+0000 I REPL [SyncSourceFeedback] setting syncSourceFeedback to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.881+0000 d20010| 2015-11-23T18:44:22.880+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38682 #12 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.881+0000 d20010| 2015-11-23T18:44:22.881+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38683 #13 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:22.882+0000 d20011| 2015-11-23T18:44:22.881+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.618+0000 d20013| 2015-11-23T18:44:23.617+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36267 #4 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.618+0000 d20013| 2015-11-23T18:44:23.618+0000 I NETWORK [conn4] end connection 10.99.163.247:36267 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.619+0000 d20014| 2015-11-23T18:44:23.618+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37206 #4 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.619+0000 d20013| 2015-11-23T18:44:23.618+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36269 #5 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.619+0000 d20014| 2015-11-23T18:44:23.619+0000 I NETWORK [conn4] end connection 10.99.163.247:37206 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.622+0000 d20013| 2015-11-23T18:44:23.620+0000 I NETWORK [conn5] end connection 10.99.163.247:36269 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.622+0000 d20015| 2015-11-23T18:44:23.621+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56540 #4 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.622+0000 d20015| 2015-11-23T18:44:23.621+0000 I NETWORK [conn4] end connection 10.99.163.247:56540 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.637+0000 d20015| 2015-11-23T18:44:23.636+0000 I REPL [replExecDBWorker-0] Starting replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.637+0000 d20015| 2015-11-23T18:44:23.637+0000 W REPL [rsSync] did not receive a valid config yet [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.639+0000 d20015| 2015-11-23T18:44:23.637+0000 I REPL [ReplicationExecutor] New replica set config in use: { _id: "test-rs1", version: 1, protocolVersion: 1, members: [ { _id: 0, host: "ip-10-99-163-247:20013", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 1, host: "ip-10-99-163-247:20014", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 2, host: "ip-10-99-163-247:20015", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 } ], settings: { chainingAllowed: true, heartbeatIntervalMillis: 2000, heartbeatTimeoutSecs: 10, electionTimeoutMillis: 10000, getLastErrorModes: {}, getLastErrorDefaults: { w: 1, wtimeout: 0 } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.639+0000 d20015| 2015-11-23T18:44:23.637+0000 I REPL [ReplicationExecutor] This node is ip-10-99-163-247:20015 in the config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.640+0000 d20015| 2015-11-23T18:44:23.637+0000 I REPL [ReplicationExecutor] transition to STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.641+0000 d20015| 2015-11-23T18:44:23.637+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20013 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.641+0000 d20014| 2015-11-23T18:44:23.638+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37209 #5 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.641+0000 d20015| 2015-11-23T18:44:23.638+0000 I ASIO [NetworkInterfaceASIO-Replication-0] Successfully connected to ip-10-99-163-247:20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.642+0000 d20015| 2015-11-23T18:44:23.638+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20014 is now in state STARTUP [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.642+0000 d20015| 2015-11-23T18:44:23.639+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56542 #5 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.642+0000 d20014| 2015-11-23T18:44:23.639+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.642+0000 d20014| 2015-11-23T18:44:23.640+0000 I REPL [replExecDBWorker-0] Starting replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.643+0000 d20014| 2015-11-23T18:44:23.640+0000 W REPL [rsSync] did not receive a valid config yet [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.644+0000 d20014| 2015-11-23T18:44:23.640+0000 I REPL [ReplicationExecutor] New replica set config in use: { _id: "test-rs1", version: 1, protocolVersion: 1, members: [ { _id: 0, host: "ip-10-99-163-247:20013", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 1, host: "ip-10-99-163-247:20014", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 }, { _id: 2, host: "ip-10-99-163-247:20015", arbiterOnly: false, buildIndexes: true, hidden: false, priority: 1.0, tags: {}, slaveDelay: 0, votes: 1 } ], settings: { chainingAllowed: true, heartbeatIntervalMillis: 2000, heartbeatTimeoutSecs: 10, electionTimeoutMillis: 10000, getLastErrorModes: {}, getLastErrorDefaults: { w: 1, wtimeout: 0 } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.644+0000 d20014| 2015-11-23T18:44:23.640+0000 I REPL [ReplicationExecutor] This node is ip-10-99-163-247:20014 in the config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.644+0000 d20014| 2015-11-23T18:44:23.640+0000 I REPL [ReplicationExecutor] transition to STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.645+0000 d20014| 2015-11-23T18:44:23.640+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20013 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.645+0000 d20014| 2015-11-23T18:44:23.641+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20015 is now in state STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.647+0000 d20013| 2015-11-23T18:44:23.646+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20015 is now in state STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:23.647+0000 d20013| 2015-11-23T18:44:23.646+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20014 is now in state STARTUP2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.638+0000 d20015| 2015-11-23T18:44:24.637+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.638+0000 d20015| 2015-11-23T18:44:24.637+0000 I REPL [rsSync] creating replication oplog of size: 1024MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.641+0000 d20014| 2015-11-23T18:44:24.640+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.641+0000 d20014| 2015-11-23T18:44:24.640+0000 I REPL [rsSync] creating replication oplog of size: 1024MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.855+0000 d20014| 2015-11-23T18:44:24.855+0000 I STORAGE [rsSync] Starting WiredTigerRecordStoreThread local.oplog.rs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.856+0000 d20015| 2015-11-23T18:44:24.855+0000 I STORAGE [rsSync] Starting WiredTigerRecordStoreThread local.oplog.rs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.856+0000 d20015| 2015-11-23T18:44:24.855+0000 I STORAGE [rsSync] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:24.856+0000 d20014| 2015-11-23T18:44:24.855+0000 I STORAGE [rsSync] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.674+0000 d20014| 2015-11-23T18:44:25.673+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.674+0000 d20014| 2015-11-23T18:44:25.673+0000 I REPL [rsSync] initial sync pending [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.674+0000 d20015| 2015-11-23T18:44:25.674+0000 I REPL [rsSync] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.676+0000 d20015| 2015-11-23T18:44:25.675+0000 I REPL [rsSync] initial sync pending [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.685+0000 d20014| 2015-11-23T18:44:25.685+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.686+0000 d20013| 2015-11-23T18:44:25.685+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36273 #6 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.686+0000 d20015| 2015-11-23T18:44:25.686+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.687+0000 d20013| 2015-11-23T18:44:25.686+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36274 #7 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.689+0000 d20014| 2015-11-23T18:44:25.689+0000 I REPL [rsSync] initial sync drop all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.689+0000 d20014| 2015-11-23T18:44:25.689+0000 I STORAGE [rsSync] dropAllDatabasesExceptLocal 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.690+0000 d20014| 2015-11-23T18:44:25.689+0000 I REPL [rsSync] initial sync clone all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.690+0000 d20014| 2015-11-23T18:44:25.689+0000 I REPL [rsSync] initial sync data copy, starting syncup [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.690+0000 d20014| 2015-11-23T18:44:25.689+0000 I REPL [rsSync] oplog sync 1 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.691+0000 d20015| 2015-11-23T18:44:25.689+0000 I REPL [rsSync] initial sync drop all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.691+0000 d20015| 2015-11-23T18:44:25.690+0000 I STORAGE [rsSync] dropAllDatabasesExceptLocal 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.692+0000 d20015| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] initial sync clone all databases [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.692+0000 d20014| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] oplog sync 2 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.692+0000 d20015| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] initial sync data copy, starting syncup [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.692+0000 d20015| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] oplog sync 1 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.693+0000 d20014| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] initial sync building indexes [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.693+0000 d20014| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] oplog sync 3 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.693+0000 d20015| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] oplog sync 2 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.694+0000 d20015| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] initial sync building indexes [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.694+0000 d20015| 2015-11-23T18:44:25.690+0000 I REPL [rsSync] oplog sync 3 of 3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.695+0000 d20014| 2015-11-23T18:44:25.692+0000 I REPL [rsSync] initial sync finishing up [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.695+0000 d20014| 2015-11-23T18:44:25.692+0000 I REPL [rsSync] set minValid=(term: -1, timestamp: Nov 23 18:44:21:1) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.696+0000 d20014| 2015-11-23T18:44:25.693+0000 I REPL [rsSync] initial sync done [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.697+0000 d20015| 2015-11-23T18:44:25.693+0000 I REPL [rsSync] initial sync finishing up [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.698+0000 d20015| 2015-11-23T18:44:25.693+0000 I REPL [rsSync] set minValid=(term: -1, timestamp: Nov 23 18:44:21:1) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.698+0000 d20015| 2015-11-23T18:44:25.694+0000 I REPL [rsSync] initial sync done [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.698+0000 d20014| 2015-11-23T18:44:25.696+0000 I REPL [ReplicationExecutor] transition to RECOVERING [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.699+0000 d20013| 2015-11-23T18:44:25.696+0000 I NETWORK [conn6] end connection 10.99.163.247:36273 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.699+0000 d20013| 2015-11-23T18:44:25.697+0000 I NETWORK [conn7] end connection 10.99.163.247:36274 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.699+0000 d20015| 2015-11-23T18:44:25.697+0000 I REPL [ReplicationExecutor] transition to RECOVERING [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.699+0000 d20014| 2015-11-23T18:44:25.698+0000 I REPL [ReplicationExecutor] transition to SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:25.700+0000 d20015| 2015-11-23T18:44:25.698+0000 I REPL [ReplicationExecutor] transition to SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:26.638+0000 d20015| 2015-11-23T18:44:26.638+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:26.639+0000 d20015| 2015-11-23T18:44:26.638+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20014 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:26.641+0000 d20014| 2015-11-23T18:44:26.641+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:26.642+0000 d20014| 2015-11-23T18:44:26.641+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20015 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:27.648+0000 d20013| 2015-11-23T18:44:27.647+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20014 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:27.648+0000 d20013| 2015-11-23T18:44:27.647+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20015 is now in state SECONDARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.124+0000 d20013| 2015-11-23T18:44:33.124+0000 I REPL [ReplicationExecutor] Starting an election, since we've seen no PRIMARY in the past 10000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.125+0000 d20013| 2015-11-23T18:44:33.124+0000 I REPL [ReplicationExecutor] conducting a dry run election to see if we could be elected [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.130+0000 d20013| 2015-11-23T18:44:33.130+0000 I REPL [ReplicationExecutor] dry election run succeeded, running for election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.131+0000 d20014| 2015-11-23T18:44:33.131+0000 I NETWORK [conn3] end connection 10.99.163.247:37194 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.135+0000 d20013| 2015-11-23T18:44:33.135+0000 I REPL [ReplicationExecutor] election succeeded, assuming primary role in term 1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.135+0000 d20013| 2015-11-23T18:44:33.135+0000 I REPL [ReplicationExecutor] transition to PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.136+0000 d20014| 2015-11-23T18:44:33.135+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37214 #6 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.136+0000 d20013| 2015-11-23T18:44:33.135+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.137+0000 d20014| 2015-11-23T18:44:33.137+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37215 #7 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.138+0000 d20013| 2015-11-23T18:44:33.137+0000 I ASIO [NetworkInterfaceASIO-Replication-1] Successfully connected to ip-10-99-163-247:20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.648+0000 d20013| 2015-11-23T18:44:33.648+0000 I REPL [rsSync] transition to primary complete; database writes are now permitted [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.724+0000 2015-11-23T18:44:33.723+0000 I NETWORK [thread1] Starting new replica set monitor for test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.724+0000 2015-11-23T18:44:33.724+0000 I NETWORK [ReplicaSetMonitorWatcher] starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.726+0000 d20012| 2015-11-23T18:44:33.726+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:45897 #6 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.728+0000 d20010| 2015-11-23T18:44:33.727+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38698 #14 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.736+0000 2015-11-23T18:44:33.735+0000 I NETWORK [thread1] Starting new replica set monitor for test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.737+0000 d20015| 2015-11-23T18:44:33.736+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56551 #6 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.738+0000 d20014| 2015-11-23T18:44:33.737+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37220 #8 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.739+0000 d20013| 2015-11-23T18:44:33.738+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36283 #8 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.740+0000 Resetting db path '/data/db/job0/mongorunner/test-config0' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.746+0000 2015-11-23T18:44:33.745+0000 I - [thread1] shell: started program (sh13641): /data/mci/src/mongod --dbpath /data/db/job0/mongorunner/test-config0 --journal --configsvr --port 20016 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.747+0000 2015-11-23T18:44:33.747+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20016, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.785+0000 c20016| 2015-11-23T18:44:33.784+0000 I CONTROL [initandlisten] MongoDB starting : pid=13641 port=20016 dbpath=/data/db/job0/mongorunner/test-config0 master=1 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.785+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.786+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.786+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.786+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.787+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.787+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.787+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.787+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.788+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.788+0000 c20016| 2015-11-23T18:44:33.785+0000 I CONTROL [initandlisten] options: { net: { port: 20016 }, nopreallocj: true, setParameter: { enableTestCommands: "1" }, sharding: { clusterRole: "configsvr" }, storage: { dbPath: "/data/db/job0/mongorunner/test-config0", engine: "wiredTiger", journal: { enabled: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.806+0000 c20016| 2015-11-23T18:44:33.805+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.880+0000 c20016| 2015-11-23T18:44:33.880+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.881+0000 c20016| 2015-11-23T18:44:33.880+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.881+0000 c20016| 2015-11-23T18:44:33.880+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.882+0000 c20016| 2015-11-23T18:44:33.880+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.894+0000 c20016| 2015-11-23T18:44:33.894+0000 I REPL [initandlisten] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.895+0000 c20016| 2015-11-23T18:44:33.894+0000 I REPL [initandlisten] creating replication oplog of size: 5MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.900+0000 c20016| 2015-11-23T18:44:33.899+0000 I STORAGE [initandlisten] Starting WiredTigerRecordStoreThread local.oplog.$main [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.900+0000 c20016| 2015-11-23T18:44:33.900+0000 I STORAGE [initandlisten] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.936+0000 c20016| 2015-11-23T18:44:33.936+0000 I REPL [initandlisten] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.937+0000 c20016| 2015-11-23T18:44:33.936+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-config0/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.937+0000 c20016| 2015-11-23T18:44:33.936+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.948+0000 2015-11-23T18:44:33.948+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20016, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:33.958+0000 c20016| 2015-11-23T18:44:33.958+0000 I NETWORK [initandlisten] waiting for connections on port 20016 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.149+0000 c20016| 2015-11-23T18:44:34.149+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:49084 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.151+0000 Resetting db path '/data/db/job0/mongorunner/test-config1' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.156+0000 2015-11-23T18:44:34.155+0000 I - [thread1] shell: started program (sh13665): /data/mci/src/mongod --dbpath /data/db/job0/mongorunner/test-config1 --journal --configsvr --port 20017 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.157+0000 2015-11-23T18:44:34.157+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20017, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.195+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] MongoDB starting : pid=13665 port=20017 dbpath=/data/db/job0/mongorunner/test-config1 master=1 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.195+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.195+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.195+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.196+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.196+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.196+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.196+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.197+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.197+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.197+0000 c20017| 2015-11-23T18:44:34.194+0000 I CONTROL [initandlisten] options: { net: { port: 20017 }, nopreallocj: true, setParameter: { enableTestCommands: "1" }, sharding: { clusterRole: "configsvr" }, storage: { dbPath: "/data/db/job0/mongorunner/test-config1", engine: "wiredTiger", journal: { enabled: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.214+0000 c20017| 2015-11-23T18:44:34.214+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.299+0000 c20017| 2015-11-23T18:44:34.298+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.299+0000 c20017| 2015-11-23T18:44:34.298+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.299+0000 c20017| 2015-11-23T18:44:34.298+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.300+0000 c20017| 2015-11-23T18:44:34.298+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.310+0000 c20017| 2015-11-23T18:44:34.310+0000 I REPL [initandlisten] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.311+0000 c20017| 2015-11-23T18:44:34.310+0000 I REPL [initandlisten] creating replication oplog of size: 5MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.313+0000 c20017| 2015-11-23T18:44:34.313+0000 I STORAGE [initandlisten] Starting WiredTigerRecordStoreThread local.oplog.$main [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.314+0000 c20017| 2015-11-23T18:44:34.313+0000 I STORAGE [initandlisten] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.336+0000 c20017| 2015-11-23T18:44:34.336+0000 I REPL [initandlisten] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.337+0000 c20017| 2015-11-23T18:44:34.336+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-config1/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.337+0000 c20017| 2015-11-23T18:44:34.336+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.348+0000 c20017| 2015-11-23T18:44:34.347+0000 I NETWORK [initandlisten] waiting for connections on port 20017 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.358+0000 c20017| 2015-11-23T18:44:34.358+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:35479 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.360+0000 Resetting db path '/data/db/job0/mongorunner/test-config2' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.366+0000 2015-11-23T18:44:34.365+0000 I - [thread1] shell: started program (sh13688): /data/mci/src/mongod --dbpath /data/db/job0/mongorunner/test-config2 --journal --configsvr --port 20018 --nopreallocj --setParameter enableTestCommands=1 --storageEngine wiredTiger [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.367+0000 2015-11-23T18:44:34.366+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20018, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.404+0000 c20018| 2015-11-23T18:44:34.404+0000 I CONTROL [initandlisten] MongoDB starting : pid=13688 port=20018 dbpath=/data/db/job0/mongorunner/test-config2 master=1 64-bit host=ip-10-99-163-247 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.405+0000 c20018| 2015-11-23T18:44:34.404+0000 I CONTROL [initandlisten] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.405+0000 c20018| 2015-11-23T18:44:34.404+0000 I CONTROL [initandlisten] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.405+0000 c20018| 2015-11-23T18:44:34.404+0000 I CONTROL [initandlisten] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.406+0000 c20018| 2015-11-23T18:44:34.405+0000 I CONTROL [initandlisten] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.406+0000 c20018| 2015-11-23T18:44:34.405+0000 I CONTROL [initandlisten] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.406+0000 c20018| 2015-11-23T18:44:34.405+0000 I CONTROL [initandlisten] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.407+0000 c20018| 2015-11-23T18:44:34.405+0000 I CONTROL [initandlisten] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.407+0000 c20018| 2015-11-23T18:44:34.405+0000 I CONTROL [initandlisten] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.407+0000 c20018| 2015-11-23T18:44:34.405+0000 I CONTROL [initandlisten] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.408+0000 c20018| 2015-11-23T18:44:34.405+0000 I CONTROL [initandlisten] options: { net: { port: 20018 }, nopreallocj: true, setParameter: { enableTestCommands: "1" }, sharding: { clusterRole: "configsvr" }, storage: { dbPath: "/data/db/job0/mongorunner/test-config2", engine: "wiredTiger", journal: { enabled: true } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.426+0000 c20018| 2015-11-23T18:44:34.425+0000 I STORAGE [initandlisten] wiredtiger_open config: create,cache_size=3G,session_max=20000,eviction=(threads_max=4),config_base=false,statistics=(fast),log=(enabled=true,archive=true,path=journal,compressor=snappy),file_manager=(close_idle_time=100000),checkpoint=(wait=60,log_size=2GB),statistics_log=(wait=0), [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.500+0000 c20018| 2015-11-23T18:44:34.500+0000 I CONTROL [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.501+0000 c20018| 2015-11-23T18:44:34.500+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.501+0000 c20018| 2015-11-23T18:44:34.500+0000 I CONTROL [initandlisten] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.501+0000 c20018| 2015-11-23T18:44:34.500+0000 I CONTROL [initandlisten] ** WARNING: soft rlimits too low. rlimits set to 1024 processes, 64000 files. Number of processes should be at least 32000 : 0.5 times number of files. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.506+0000 c20018| 2015-11-23T18:44:34.505+0000 I REPL [initandlisten] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.506+0000 c20018| 2015-11-23T18:44:34.505+0000 I REPL [initandlisten] creating replication oplog of size: 5MB... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.507+0000 c20018| 2015-11-23T18:44:34.507+0000 I STORAGE [initandlisten] Starting WiredTigerRecordStoreThread local.oplog.$main [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.508+0000 c20018| 2015-11-23T18:44:34.507+0000 I STORAGE [initandlisten] Scanning the oplog to determine where to place markers for when to truncate [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.536+0000 c20018| 2015-11-23T18:44:34.536+0000 I REPL [initandlisten] ****** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.537+0000 c20018| 2015-11-23T18:44:34.537+0000 I FTDC [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/job0/mongorunner/test-config2/diagnostic.data' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.538+0000 c20018| 2015-11-23T18:44:34.537+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.555+0000 c20018| 2015-11-23T18:44:34.555+0000 I NETWORK [initandlisten] waiting for connections on port 20018 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.568+0000 c20018| 2015-11-23T18:44:34.568+0000 I NETWORK [initandlisten] connection accepted from 127.0.0.1:49909 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.569+0000 "config servers: ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018" [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.570+0000 2015-11-23T18:44:34.569+0000 I NETWORK [thread1] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.571+0000 c20016| 2015-11-23T18:44:34.570+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48277 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.571+0000 2015-11-23T18:44:34.570+0000 I NETWORK [thread1] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.572+0000 c20017| 2015-11-23T18:44:34.571+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55592 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.572+0000 2015-11-23T18:44:34.571+0000 I NETWORK [thread1] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.572+0000 c20018| 2015-11-23T18:44:34.572+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36713 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.574+0000 ShardingTest test : [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.574+0000 { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.574+0000 "config" : "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.574+0000 "shards" : [ [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.574+0000 connection to test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.575+0000 connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.575+0000 ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.575+0000 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.579+0000 2015-11-23T18:44:34.578+0000 I - [thread1] shell: started program (sh13717): /data/mci/src/mongos --configdb ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 --chunkSize 50 --port 20019 --setParameter enableTestCommands=1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.580+0000 2015-11-23T18:44:34.579+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20019, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.603+0000 s20019| 2015-11-23T18:44:34.603+0000 I CONTROL [main] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.603+0000 s20019| 2015-11-23T18:44:34.603+0000 I CONTROL [main] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.604+0000 s20019| 2015-11-23T18:44:34.603+0000 I SHARDING [mongosMain] MongoS version 3.2.0-rc3-103-ge3cd63f starting: pid=13717 port=20019 64-bit host=ip-10-99-163-247 (--help for usage) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.604+0000 s20019| 2015-11-23T18:44:34.603+0000 I CONTROL [mongosMain] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.605+0000 s20019| 2015-11-23T18:44:34.603+0000 I CONTROL [mongosMain] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.605+0000 s20019| 2015-11-23T18:44:34.603+0000 I CONTROL [mongosMain] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.605+0000 s20019| 2015-11-23T18:44:34.604+0000 I CONTROL [mongosMain] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.605+0000 s20019| 2015-11-23T18:44:34.604+0000 I CONTROL [mongosMain] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.606+0000 s20019| 2015-11-23T18:44:34.604+0000 I CONTROL [mongosMain] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.606+0000 s20019| 2015-11-23T18:44:34.604+0000 I CONTROL [mongosMain] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.606+0000 s20019| 2015-11-23T18:44:34.604+0000 I CONTROL [mongosMain] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.606+0000 s20019| 2015-11-23T18:44:34.604+0000 I CONTROL [mongosMain] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.607+0000 s20019| 2015-11-23T18:44:34.604+0000 I CONTROL [mongosMain] options: { net: { port: 20019 }, setParameter: { enableTestCommands: "1" }, sharding: { chunkSize: 50, configDB: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.607+0000 s20019| 2015-11-23T18:44:34.604+0000 I SHARDING [mongosMain] Updating config server connection string to: ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.608+0000 c20016| 2015-11-23T18:44:34.608+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48281 #3 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.610+0000 c20017| 2015-11-23T18:44:34.609+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55596 #3 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.611+0000 c20018| 2015-11-23T18:44:34.611+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36717 #3 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.611+0000 s20019| 2015-11-23T18:44:34.611+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.612+0000 c20016| 2015-11-23T18:44:34.612+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48284 #4 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.612+0000 s20019| 2015-11-23T18:44:34.612+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.613+0000 c20017| 2015-11-23T18:44:34.612+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55599 #4 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.613+0000 s20019| 2015-11-23T18:44:34.613+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.613+0000 c20018| 2015-11-23T18:44:34.613+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36720 #4 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.616+0000 c20016| 2015-11-23T18:44:34.615+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48287 #5 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.617+0000 c20017| 2015-11-23T18:44:34.617+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55602 #5 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.619+0000 c20018| 2015-11-23T18:44:34.618+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36723 #5 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.623+0000 s20019| 2015-11-23T18:44:34.623+0000 I SHARDING [LockPinger] creating distributed lock ping thread for ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 and process ip-10-99-163-247:20019:1448304274:1804289383 (sleeping for 30000ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.624+0000 s20019| 2015-11-23T18:44:34.623+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.624+0000 c20016| 2015-11-23T18:44:34.624+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48290 #6 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.625+0000 s20019| 2015-11-23T18:44:34.624+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.626+0000 c20017| 2015-11-23T18:44:34.626+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55605 #6 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.626+0000 s20019| 2015-11-23T18:44:34.626+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.627+0000 c20018| 2015-11-23T18:44:34.626+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36726 #6 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.641+0000 d20015| 2015-11-23T18:44:34.640+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20013 is now in state PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.643+0000 d20014| 2015-11-23T18:44:34.643+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20013 is now in state PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.691+0000 s20019| 2015-11-23T18:44:34.690+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:44:34.627+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20019:1448304274:1804289383', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.699+0000 s20019| 2015-11-23T18:44:34.698+0000 I SHARDING [mongosMain] distributed lock 'configUpgrade/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'initializing config database to new format v6', ts : 56535e926e3a936d394e6c8f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.699+0000 s20019| 2015-11-23T18:44:34.699+0000 I SHARDING [mongosMain] initializing config server version to 6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.699+0000 s20019| 2015-11-23T18:44:34.699+0000 I SHARDING [mongosMain] writing initial config version at v6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.700+0000 c20016| 2015-11-23T18:44:34.699+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48293 #7 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.701+0000 c20017| 2015-11-23T18:44:34.701+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55608 #7 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.702+0000 c20018| 2015-11-23T18:44:34.702+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36729 #7 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.703+0000 c20018| 2015-11-23T18:44:34.702+0000 I SHARDING [conn7] remote client 10.99.163.247:36729 initialized this host as shard config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.703+0000 c20016| 2015-11-23T18:44:34.702+0000 I SHARDING [conn7] remote client 10.99.163.247:48293 initialized this host as shard config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.703+0000 c20017| 2015-11-23T18:44:34.702+0000 I SHARDING [conn7] remote client 10.99.163.247:55608 initialized this host as shard config [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.721+0000 s20019| 2015-11-23T18:44:34.721+0000 I SHARDING [mongosMain] initialization of config server to v6 successful [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.722+0000 s20019| 2015-11-23T18:44:34.721+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.722+0000 c20016| 2015-11-23T18:44:34.722+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48296 #8 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.723+0000 s20019| 2015-11-23T18:44:34.722+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.723+0000 c20017| 2015-11-23T18:44:34.723+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55611 #8 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.724+0000 s20019| 2015-11-23T18:44:34.723+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.724+0000 c20018| 2015-11-23T18:44:34.724+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36732 #8 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.735+0000 s20019| 2015-11-23T18:44:34.734+0000 I SHARDING [mongosMain] distributed lock 'configUpgrade/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.763+0000 c20016| 2015-11-23T18:44:34.763+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, min: 1 }, name: "ns_1_min_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.764+0000 c20016| 2015-11-23T18:44:34.763+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.764+0000 c20018| 2015-11-23T18:44:34.763+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, min: 1 }, name: "ns_1_min_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.764+0000 c20018| 2015-11-23T18:44:34.763+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.765+0000 c20017| 2015-11-23T18:44:34.763+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, min: 1 }, name: "ns_1_min_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.765+0000 c20017| 2015-11-23T18:44:34.764+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.765+0000 c20016| 2015-11-23T18:44:34.764+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.765+0000 c20018| 2015-11-23T18:44:34.764+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.766+0000 c20017| 2015-11-23T18:44:34.765+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.775+0000 c20016| 2015-11-23T18:44:34.775+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, shard: 1, min: 1 }, name: "ns_1_shard_1_min_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.776+0000 c20016| 2015-11-23T18:44:34.775+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.776+0000 c20018| 2015-11-23T18:44:34.775+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, shard: 1, min: 1 }, name: "ns_1_shard_1_min_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.777+0000 c20017| 2015-11-23T18:44:34.775+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, shard: 1, min: 1 }, name: "ns_1_shard_1_min_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.777+0000 c20018| 2015-11-23T18:44:34.775+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.777+0000 c20017| 2015-11-23T18:44:34.775+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.777+0000 c20016| 2015-11-23T18:44:34.776+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.778+0000 c20017| 2015-11-23T18:44:34.776+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.778+0000 c20018| 2015-11-23T18:44:34.776+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.781+0000 2015-11-23T18:44:34.780+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20019, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.783+0000 c20016| 2015-11-23T18:44:34.783+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, lastmod: 1 }, name: "ns_1_lastmod_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.783+0000 c20016| 2015-11-23T18:44:34.783+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.784+0000 c20018| 2015-11-23T18:44:34.783+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, lastmod: 1 }, name: "ns_1_lastmod_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.784+0000 c20018| 2015-11-23T18:44:34.783+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.785+0000 c20017| 2015-11-23T18:44:34.784+0000 I INDEX [conn7] build index on: config.chunks properties: { v: 1, unique: true, key: { ns: 1, lastmod: 1 }, name: "ns_1_lastmod_1", ns: "config.chunks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.785+0000 c20017| 2015-11-23T18:44:34.784+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.785+0000 c20018| 2015-11-23T18:44:34.785+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.785+0000 c20016| 2015-11-23T18:44:34.785+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.786+0000 c20017| 2015-11-23T18:44:34.786+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.798+0000 c20017| 2015-11-23T18:44:34.797+0000 I INDEX [conn7] build index on: config.shards properties: { v: 1, unique: true, key: { host: 1 }, name: "host_1", ns: "config.shards" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.798+0000 c20017| 2015-11-23T18:44:34.797+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.799+0000 c20017| 2015-11-23T18:44:34.799+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.800+0000 c20016| 2015-11-23T18:44:34.799+0000 I INDEX [conn7] build index on: config.shards properties: { v: 1, unique: true, key: { host: 1 }, name: "host_1", ns: "config.shards" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.800+0000 c20016| 2015-11-23T18:44:34.799+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.802+0000 c20016| 2015-11-23T18:44:34.802+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.806+0000 c20018| 2015-11-23T18:44:34.806+0000 I INDEX [conn7] build index on: config.shards properties: { v: 1, unique: true, key: { host: 1 }, name: "host_1", ns: "config.shards" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.807+0000 c20018| 2015-11-23T18:44:34.806+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.808+0000 c20018| 2015-11-23T18:44:34.808+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.825+0000 c20016| 2015-11-23T18:44:34.824+0000 I INDEX [conn7] build index on: config.locks properties: { v: 1, key: { ts: 1 }, name: "ts_1", ns: "config.locks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.825+0000 c20016| 2015-11-23T18:44:34.825+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.825+0000 c20017| 2015-11-23T18:44:34.824+0000 I INDEX [conn7] build index on: config.locks properties: { v: 1, key: { ts: 1 }, name: "ts_1", ns: "config.locks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.826+0000 c20017| 2015-11-23T18:44:34.824+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.833+0000 c20018| 2015-11-23T18:44:34.832+0000 I INDEX [conn7] build index on: config.locks properties: { v: 1, key: { ts: 1 }, name: "ts_1", ns: "config.locks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.833+0000 c20018| 2015-11-23T18:44:34.832+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.833+0000 c20017| 2015-11-23T18:44:34.833+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.834+0000 c20016| 2015-11-23T18:44:34.833+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.834+0000 c20018| 2015-11-23T18:44:34.834+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.847+0000 c20017| 2015-11-23T18:44:34.847+0000 I INDEX [conn7] build index on: config.locks properties: { v: 1, key: { state: 1, process: 1 }, name: "state_1_process_1", ns: "config.locks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.848+0000 c20017| 2015-11-23T18:44:34.847+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.848+0000 c20018| 2015-11-23T18:44:34.847+0000 I INDEX [conn7] build index on: config.locks properties: { v: 1, key: { state: 1, process: 1 }, name: "state_1_process_1", ns: "config.locks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.848+0000 c20018| 2015-11-23T18:44:34.847+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.849+0000 c20016| 2015-11-23T18:44:34.848+0000 I INDEX [conn7] build index on: config.locks properties: { v: 1, key: { state: 1, process: 1 }, name: "state_1_process_1", ns: "config.locks" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.849+0000 c20016| 2015-11-23T18:44:34.848+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.852+0000 c20018| 2015-11-23T18:44:34.852+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.855+0000 c20017| 2015-11-23T18:44:34.855+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.856+0000 c20016| 2015-11-23T18:44:34.856+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.875+0000 c20017| 2015-11-23T18:44:34.875+0000 I INDEX [conn7] build index on: config.lockpings properties: { v: 1, key: { ping: 1 }, name: "ping_1", ns: "config.lockpings" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.875+0000 c20017| 2015-11-23T18:44:34.875+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.876+0000 c20016| 2015-11-23T18:44:34.875+0000 I INDEX [conn7] build index on: config.lockpings properties: { v: 1, key: { ping: 1 }, name: "ping_1", ns: "config.lockpings" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.876+0000 c20016| 2015-11-23T18:44:34.875+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.876+0000 c20018| 2015-11-23T18:44:34.875+0000 I INDEX [conn7] build index on: config.lockpings properties: { v: 1, key: { ping: 1 }, name: "ping_1", ns: "config.lockpings" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.876+0000 c20018| 2015-11-23T18:44:34.875+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.881+0000 c20017| 2015-11-23T18:44:34.881+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.882+0000 c20016| 2015-11-23T18:44:34.881+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.887+0000 c20018| 2015-11-23T18:44:34.887+0000 I INDEX [conn7] build index done. scanned 1 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.922+0000 c20018| 2015-11-23T18:44:34.922+0000 I INDEX [conn7] build index on: config.tags properties: { v: 1, unique: true, key: { ns: 1, min: 1 }, name: "ns_1_min_1", ns: "config.tags" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.923+0000 c20018| 2015-11-23T18:44:34.922+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.923+0000 c20016| 2015-11-23T18:44:34.923+0000 I INDEX [conn7] build index on: config.tags properties: { v: 1, unique: true, key: { ns: 1, min: 1 }, name: "ns_1_min_1", ns: "config.tags" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.924+0000 c20017| 2015-11-23T18:44:34.923+0000 I INDEX [conn7] build index on: config.tags properties: { v: 1, unique: true, key: { ns: 1, min: 1 }, name: "ns_1_min_1", ns: "config.tags" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.924+0000 c20016| 2015-11-23T18:44:34.923+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.924+0000 c20017| 2015-11-23T18:44:34.923+0000 I INDEX [conn7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.924+0000 c20018| 2015-11-23T18:44:34.923+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.925+0000 c20016| 2015-11-23T18:44:34.925+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.926+0000 c20017| 2015-11-23T18:44:34.925+0000 I INDEX [conn7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.926+0000 s20019| 2015-11-23T18:44:34.925+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.926+0000 s20019| 2015-11-23T18:44:34.926+0000 I SHARDING [Balancer] about to contact config servers and shards [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.927+0000 s20019| 2015-11-23T18:44:34.926+0000 I SHARDING [Balancer] config servers and shards contacted successfully [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.927+0000 s20019| 2015-11-23T18:44:34.926+0000 I SHARDING [Balancer] balancer id: ip-10-99-163-247:20019 started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.927+0000 c20016| 2015-11-23T18:44:34.927+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48300 #9 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.928+0000 c20017| 2015-11-23T18:44:34.928+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55615 #9 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.929+0000 c20018| 2015-11-23T18:44:34.929+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36736 #9 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.947+0000 s20019| 2015-11-23T18:44:34.947+0000 I NETWORK [mongosMain] waiting for connections on port 20019 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.977+0000 s20019| 2015-11-23T18:44:34.977+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'doing balance round', ts : 56535e926e3a936d394e6c92 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.981+0000 s20019| 2015-11-23T18:44:34.981+0000 I NETWORK [mongosMain] connection accepted from 127.0.0.1:51291 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.990+0000 2015-11-23T18:44:34.989+0000 I - [thread1] shell: started program (sh13779): /data/mci/src/mongos --configdb ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 --chunkSize 50 --port 20020 --setParameter enableTestCommands=1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.991+0000 2015-11-23T18:44:34.990+0000 W NETWORK [thread1] Failed to connect to 127.0.0.1:20020, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:34.993+0000 s20019| 2015-11-23T18:44:34.992+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:44:34.992+0000-56535e926e3a936d394e6c93", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304274992), what: "balancer.round", ns: "", details: { executionTimeMillis: 51, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.006+0000 s20019| 2015-11-23T18:44:35.000+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.022+0000 s20020| 2015-11-23T18:44:35.021+0000 I CONTROL [main] ** WARNING: You are running this process as the root user, which is not recommended. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.022+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [main] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.022+0000 s20020| 2015-11-23T18:44:35.022+0000 I SHARDING [mongosMain] MongoS version 3.2.0-rc3-103-ge3cd63f starting: pid=13779 port=20020 64-bit host=ip-10-99-163-247 (--help for usage) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.023+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] db version v3.2.0-rc3-103-ge3cd63f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.023+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] git version: e3cd63fcae3deb1140941a51c85564f098062a23 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.023+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] OpenSSL version: OpenSSL 1.0.1e-fips 11 Feb 2013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.023+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] allocator: tcmalloc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.024+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] modules: enterprise [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.024+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] build environment: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.024+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] distmod: rhel62 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.024+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] distarch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.025+0000 s20020| 2015-11-23T18:44:35.022+0000 I CONTROL [mongosMain] target_arch: x86_64 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.025+0000 s20020| 2015-11-23T18:44:35.023+0000 I CONTROL [mongosMain] options: { net: { port: 20020 }, setParameter: { enableTestCommands: "1" }, sharding: { chunkSize: 50, configDB: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.026+0000 s20020| 2015-11-23T18:44:35.023+0000 I SHARDING [mongosMain] Updating config server connection string to: ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.027+0000 c20016| 2015-11-23T18:44:35.026+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48305 #10 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.029+0000 c20017| 2015-11-23T18:44:35.028+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55620 #10 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.030+0000 c20018| 2015-11-23T18:44:35.029+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36741 #10 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.030+0000 s20020| 2015-11-23T18:44:35.030+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.031+0000 c20016| 2015-11-23T18:44:35.031+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48308 #11 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.032+0000 s20020| 2015-11-23T18:44:35.031+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.032+0000 c20017| 2015-11-23T18:44:35.032+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55623 #11 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.032+0000 s20020| 2015-11-23T18:44:35.032+0000 I NETWORK [mongosMain] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.033+0000 c20018| 2015-11-23T18:44:35.033+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36744 #11 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.035+0000 c20016| 2015-11-23T18:44:35.035+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48311 #12 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.036+0000 c20017| 2015-11-23T18:44:35.036+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55626 #12 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.037+0000 c20018| 2015-11-23T18:44:35.037+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36747 #12 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.045+0000 s20020| 2015-11-23T18:44:35.045+0000 I NETWORK [HostnameCanonicalizationWorker] Starting hostname canonicalization worker [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.046+0000 s20020| 2015-11-23T18:44:35.045+0000 I SHARDING [Balancer] about to contact config servers and shards [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.046+0000 s20020| 2015-11-23T18:44:35.046+0000 I SHARDING [Balancer] config servers and shards contacted successfully [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.046+0000 s20020| 2015-11-23T18:44:35.046+0000 I SHARDING [Balancer] balancer id: ip-10-99-163-247:20020 started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.048+0000 c20016| 2015-11-23T18:44:35.047+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48314 #13 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.048+0000 c20017| 2015-11-23T18:44:35.048+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55629 #13 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.049+0000 c20018| 2015-11-23T18:44:35.049+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36750 #13 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.051+0000 c20016| 2015-11-23T18:44:35.051+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48317 #14 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.053+0000 c20017| 2015-11-23T18:44:35.053+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55632 #14 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.054+0000 c20018| 2015-11-23T18:44:35.054+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36753 #14 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.060+0000 s20020| 2015-11-23T18:44:35.059+0000 I SHARDING [LockPinger] creating distributed lock ping thread for ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 and process ip-10-99-163-247:20020:1448304275:1804289383 (sleeping for 30000ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.060+0000 s20020| 2015-11-23T18:44:35.059+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.061+0000 c20016| 2015-11-23T18:44:35.060+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48320 #15 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.061+0000 s20020| 2015-11-23T18:44:35.061+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.062+0000 c20017| 2015-11-23T18:44:35.061+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55635 #15 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.062+0000 s20020| 2015-11-23T18:44:35.062+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.062+0000 c20018| 2015-11-23T18:44:35.062+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36756 #15 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.068+0000 s20020| 2015-11-23T18:44:35.068+0000 I NETWORK [mongosMain] waiting for connections on port 20020 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.075+0000 s20020| 2015-11-23T18:44:35.074+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:44:35.063+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20020:1448304275:1804289383', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.076+0000 s20020| 2015-11-23T18:44:35.074+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' acquired for 'doing balance round', ts : 56535e9313c05a55d8fa6f8e [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.077+0000 s20020| 2015-11-23T18:44:35.076+0000 I NETWORK [Balancer] scoped connection to ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.078+0000 s20020| 2015-11-23T18:44:35.076+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:44:35.076+0000-56535e9313c05a55d8fa6f8f", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304275076), what: "balancer.round", ns: "", details: { executionTimeMillis: 29, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.079+0000 c20018| 2015-11-23T18:44:35.077+0000 I NETWORK [conn11] end connection 10.99.163.247:36744 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.080+0000 c20017| 2015-11-23T18:44:35.077+0000 I NETWORK [conn11] end connection 10.99.163.247:55623 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.080+0000 c20016| 2015-11-23T18:44:35.077+0000 I NETWORK [conn11] end connection 10.99.163.247:48308 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.081+0000 s20020| 2015-11-23T18:44:35.078+0000 I NETWORK [Balancer] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.081+0000 c20016| 2015-11-23T18:44:35.078+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48323 #16 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.081+0000 s20020| 2015-11-23T18:44:35.079+0000 I NETWORK [Balancer] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.082+0000 c20017| 2015-11-23T18:44:35.079+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55638 #16 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.082+0000 s20020| 2015-11-23T18:44:35.079+0000 I NETWORK [Balancer] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.082+0000 c20018| 2015-11-23T18:44:35.080+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36759 #16 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.086+0000 s20020| 2015-11-23T18:44:35.086+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.192+0000 s20020| 2015-11-23T18:44:35.191+0000 I NETWORK [mongosMain] connection accepted from 127.0.0.1:56238 #1 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.193+0000 ShardingTest test going to add shard : test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.193+0000 s20019| 2015-11-23T18:44:35.193+0000 I NETWORK [conn1] Starting new replica set monitor for test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.194+0000 s20019| 2015-11-23T18:44:35.193+0000 I NETWORK [ReplicaSetMonitorWatcher] starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.194+0000 d20012| 2015-11-23T18:44:35.194+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:45959 #7 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.195+0000 d20010| 2015-11-23T18:44:35.195+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38760 #15 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.197+0000 d20010| 2015-11-23T18:44:35.197+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38761 #16 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.198+0000 s20019| 2015-11-23T18:44:35.197+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-TaskExecutor-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.199+0000 s20019| 2015-11-23T18:44:35.198+0000 I SHARDING [conn1] going to add shard: { _id: "test-rs0", host: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.200+0000 c20016| 2015-11-23T18:44:35.199+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48330 #17 (16 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.201+0000 c20017| 2015-11-23T18:44:35.200+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55645 #17 (16 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.202+0000 c20018| 2015-11-23T18:44:35.201+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36766 #17 (16 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.228+0000 s20019| 2015-11-23T18:44:35.226+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:35.226+0000-56535e936e3a936d394e6c94", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304275226), what: "addShard", ns: "", details: { name: "test-rs0", host: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.228+0000 ShardingTest test going to add shard : test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.229+0000 s20019| 2015-11-23T18:44:35.228+0000 I NETWORK [conn1] Starting new replica set monitor for test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.230+0000 d20015| 2015-11-23T18:44:35.228+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56617 #7 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.230+0000 d20013| 2015-11-23T18:44:35.230+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36348 #9 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.232+0000 d20013| 2015-11-23T18:44:35.231+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36349 #10 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.232+0000 s20019| 2015-11-23T18:44:35.231+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-TaskExecutor-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.233+0000 s20019| 2015-11-23T18:44:35.232+0000 I SHARDING [conn1] going to add shard: { _id: "test-rs1", host: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.240+0000 s20019| 2015-11-23T18:44:35.239+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:35.239+0000-56535e936e3a936d394e6c95", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304275239), what: "addShard", ns: "", details: { name: "test-rs1", host: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.250+0000 d20010| 2015-11-23T18:44:35.249+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38768 #17 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.250+0000 d20010| 2015-11-23T18:44:35.249+0000 I SHARDING [conn17] remote client 10.99.163.247:38768 initialized this host as shard test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.250+0000 d20010| 2015-11-23T18:44:35.250+0000 I SHARDING [ShardingState initialization] first cluster operation detected, adding sharding hook to enable versioning and authentication to remote servers [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.251+0000 d20010| 2015-11-23T18:44:35.250+0000 I SHARDING [ShardingState initialization] Updating config server connection string to: ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.252+0000 d20010| 2015-11-23T18:44:35.252+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.253+0000 c20016| 2015-11-23T18:44:35.252+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48337 #18 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.253+0000 d20010| 2015-11-23T18:44:35.253+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.253+0000 c20017| 2015-11-23T18:44:35.253+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55652 #18 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.254+0000 d20010| 2015-11-23T18:44:35.253+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.254+0000 c20018| 2015-11-23T18:44:35.254+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36773 #18 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.255+0000 d20010| 2015-11-23T18:44:35.255+0000 I NETWORK [ShardingState initialization] Starting new replica set monitor for test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.256+0000 d20010| 2015-11-23T18:44:35.255+0000 I NETWORK [ShardingState initialization] Starting new replica set monitor for test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.257+0000 d20010| 2015-11-23T18:44:35.255+0000 I NETWORK [ReplicaSetMonitorWatcher] starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.257+0000 s20019| 2015-11-23T18:44:35.256+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.257+0000 d20013| 2015-11-23T18:44:35.257+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36354 #11 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.258+0000 d20013| 2015-11-23T18:44:35.257+0000 I SHARDING [conn11] remote client 10.99.163.247:36354 initialized this host as shard test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.258+0000 d20013| 2015-11-23T18:44:35.257+0000 I SHARDING [ShardingState initialization] first cluster operation detected, adding sharding hook to enable versioning and authentication to remote servers [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.258+0000 d20013| 2015-11-23T18:44:35.257+0000 I SHARDING [ShardingState initialization] Updating config server connection string to: ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.260+0000 d20013| 2015-11-23T18:44:35.260+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.261+0000 c20016| 2015-11-23T18:44:35.261+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48341 #19 (18 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.262+0000 d20013| 2015-11-23T18:44:35.261+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.263+0000 c20017| 2015-11-23T18:44:35.262+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55656 #19 (18 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.263+0000 d20013| 2015-11-23T18:44:35.263+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.263+0000 c20018| 2015-11-23T18:44:35.263+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36777 #19 (18 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.264+0000 d20013| 2015-11-23T18:44:35.264+0000 I NETWORK [ShardingState initialization] Starting new replica set monitor for test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.265+0000 d20013| 2015-11-23T18:44:35.264+0000 I NETWORK [ShardingState initialization] Starting new replica set monitor for test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.265+0000 d20013| 2015-11-23T18:44:35.264+0000 I NETWORK [ReplicaSetMonitorWatcher] starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.265+0000 s20019| 2015-11-23T18:44:35.264+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.267+0000 setting random seed: 538399895 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.267+0000 Using 0 threads (requested 0) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.267+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.267+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.267+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.267+0000 The entire schedule of FSM workloads: [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.268+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.268+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.268+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 [ [ "jstests/concurrency/fsm_workloads/explain_group.js" ], [ "jstests/concurrency/fsm_workloads/create_capped_collection.js" ], [ "jstests/concurrency/fsm_workloads/yield_sort_merge.js" ], [ "jstests/concurrency/fsm_workloads/yield_text.js" ], [ "jstests/concurrency/fsm_workloads/findAndModify_remove.js" ], [ "jstests/concurrency/fsm_workloads/findAndModify_inc.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_2d.js" ], [ "jstests/concurrency/fsm_workloads/yield_id_hack.js" ], [ "jstests/concurrency/fsm_workloads/explain_distinct.js" ], [ "jstests/concurrency/fsm_workloads/update_ordered_bulk_inc.js" ], [ "jstests/concurrency/fsm_workloads/findAndModify_update.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_heterogeneous.js" ], [ "jstests/concurrency/fsm_workloads/remove_and_bulk_insert.js" ], [ "jstests/concurrency/fsm_workloads/findAndModify_upsert_collscan.js" ], [ "jstests/concurrency/fsm_workloads/yield_fetch.js" ], [ "jstests/concurrency/fsm_workloads/yield_sort.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_base_noindex.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_text.js" ], [ "jstests/concurrency/fsm_workloads/count_indexed.js" ], [ "jstests/concurrency/fsm_workloads/update_array_noindex.js" ], [ "jstests/concurrency/fsm_workloads/update_multifield_isolated_multiupdate_noindex.js" ], [ "jstests/concurrency/fsm_workloads/update_check_index.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_multikey_noindex.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_large.js" ], [ "jstests/concurrency/fsm_workloads/update_replace.js" ], [ "jstests/concurrency/fsm_workloads/update_inc.js" ], [ "jstests/concurrency/fsm_workloads/update_inc_capped.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_base_capped.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_long_fieldname_noindex.js" ], [ "jstests/concurrency/fsm_workloads/explain_aggregate.js" ], [ "jstests/concurrency/fsm_workloads/update_and_bulk_insert.js" ], [ "jstests/concurrency/fsm_workloads/touch_data.js" ], [ "jstests/concurrency/fsm_workloads/touch_index.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_1char.js" ], [ "jstests/concurrency/fsm_workloads/yield_rooted_or.js" ], [ "jstests/concurrency/fsm_workloads/drop_collection.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_compound.js" ], [ "jstests/concurrency/fsm_workloads/agg_sort.js" ], [ "jstests/concurrency/fsm_workloads/collmod.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_1char_noindex.js" ], [ "jstests/concurrency/fsm_workloads/update_simple.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_2dsphere.js" ], [ "jstests/concurrency/fsm_workloads/collmod_separate_collections.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_unordered_bulk.js" ], [ "jstests/concurrency/fsm_workloads/explain_update.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_ttl.js" ], [ "jstests/concurrency/fsm_workloads/findAndModify_upsert.js" ], [ "jstests/concurrency/fsm_workloads/findAndModify_update_grow.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_multikey.js" ], [ "jstests/concurrency/fsm_workloads/explain_find.js" ], [ "jstests/concurrency/fsm_workloads/explain_count.js" ], [ "jstests/concurrency/fsm_workloads/list_indexes.js" ], [ "jstests/concurrency/fsm_workloads/create_index_background.js" ], [ "jstests/concurrency/fsm_workloads/touch_base.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_text_multikey.js" ], [ "jstests/concurrency/fsm_workloads/update_where.js" ], [ "jstests/concurrency/fsm_workloads/update_multifield_isolated_multiupdate.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_ordered_bulk.js" ], [ "jstests/concurrency/fsm_workloads/update_multifield_noindex.js" ], [ "jstests/concurrency/fsm_workloads/explain_remove.js" ], [ "jstests/concurrency/fsm_workloads/agg_base.js" ], [ "jstests/concurrency/fsm_workloads/update_rename.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_large_noindex.js" ], [ "jstests/concurrency/fsm_workloads/explain.js" ], [ "jstests/concurrency/fsm_workloads/yield_geo_near.js" ], [ "jstests/concurrency/fsm_workloads/create_collection.js" ], [ "jstests/concurrency/fsm_workloads/touch_no_data_no_index.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_upsert.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_where.js" ], [ "jstests/concurrency/fsm_workloads/server_status.js" ], [ "jstests/concurrency/fsm_workloads/update_multifield_multiupdate.js" ], [ "jstests/concurrency/fsm_workloads/update_simple_noindex.js" ], [ "jstests/concurrency/fsm_workloads/yield.js" ], [ "jstests/concurrency/fsm_workloads/reindex.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_base.js" ], [ "jstests/concurrency/fsm_workloads/update_rename_noindex.js" ], [ "jstests/concurrency/fsm_workloads/update_array.js" ], [ "jstests/concurrency/fsm_workloads/update_multifield_multiupdate_noindex.js" ], [ "jstests/concurrency/fsm_workloads/remove_multiple_documents.js" ], [ "jstests/concurrency/fsm_workloads/reindex_background.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_heterogeneous_noindex.js" ], [ "jstests/concurrency/fsm_workloads/update_replace_noindex.js" ], [ "jstests/concurrency/fsm_workloads/create_capped_collection_maxdocs.js" ], [ "jstests/concurrency/fsm_workloads/update_multifield.js" ], [ "jstests/concurrency/fsm_workloads/indexed_insert_long_fieldname.js" ] ] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 End of schedule [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.279+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.280+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.280+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.280+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.280+0000 Workload(s) started: jstests/concurrency/fsm_workloads/explain_group.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.280+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.280+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.280+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.310+0000 s20019| 2015-11-23T18:44:35.309+0000 I SHARDING [conn1] distributed lock 'db0/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535e936e3a936d394e6c96 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.311+0000 s20019| 2015-11-23T18:44:35.310+0000 I SHARDING [conn1] Placing [db0] on: test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.312+0000 s20019| 2015-11-23T18:44:35.310+0000 I SHARDING [conn1] Enabling sharding for database [db0] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.357+0000 s20019| 2015-11-23T18:44:35.356+0000 I SHARDING [conn1] distributed lock 'db0/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.359+0000 d20010| 2015-11-23T18:44:35.358+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38776 #18 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.360+0000 d20010| 2015-11-23T18:44:35.359+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38777 #19 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.373+0000 d20010| 2015-11-23T18:44:35.373+0000 I INDEX [conn19] build index on: db0.coll0 properties: { v: 1, key: { j: 1.0 }, name: "j_1", ns: "db0.coll0" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.374+0000 d20010| 2015-11-23T18:44:35.373+0000 I INDEX [conn19] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.377+0000 d20010| 2015-11-23T18:44:35.376+0000 I INDEX [conn19] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.377+0000 s20019| 2015-11-23T18:44:35.377+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db0.coll0", key: { j: 1.0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.391+0000 d20012| 2015-11-23T18:44:35.391+0000 I INDEX [repl writer worker 4] build index on: db0.coll0 properties: { v: 1, key: { j: 1.0 }, name: "j_1", ns: "db0.coll0" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.391+0000 d20012| 2015-11-23T18:44:35.391+0000 I INDEX [repl writer worker 4] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.394+0000 d20011| 2015-11-23T18:44:35.393+0000 I INDEX [repl writer worker 5] build index on: db0.coll0 properties: { v: 1, key: { j: 1.0 }, name: "j_1", ns: "db0.coll0" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.395+0000 d20012| 2015-11-23T18:44:35.393+0000 I INDEX [repl writer worker 4] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.395+0000 d20011| 2015-11-23T18:44:35.393+0000 I INDEX [repl writer worker 5] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.396+0000 d20011| 2015-11-23T18:44:35.395+0000 I INDEX [repl writer worker 5] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.410+0000 s20019| 2015-11-23T18:44:35.409+0000 I SHARDING [conn1] distributed lock 'db0.coll0/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535e936e3a936d394e6c97 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.411+0000 s20019| 2015-11-23T18:44:35.410+0000 I SHARDING [conn1] enable sharding on: db0.coll0 with shard key: { j: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.412+0000 s20019| 2015-11-23T18:44:35.410+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:35.410+0000-56535e936e3a936d394e6c98", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304275410), what: "shardCollection.start", ns: "db0.coll0", details: { shardKey: { j: 1.0 }, collection: "db0.coll0", primary: "test-rs0:test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", initShards: [], numChunks: 1 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.413+0000 s20019| 2015-11-23T18:44:35.412+0000 I SHARDING [conn1] going to create 1 chunk(s) for: db0.coll0 using new epoch 56535e936e3a936d394e6c99 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.427+0000 s20019| 2015-11-23T18:44:35.427+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db0.coll0: 0ms sequenceNumber: 2 version: 1|0||56535e936e3a936d394e6c99 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.444+0000 s20019| 2015-11-23T18:44:35.443+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db0.coll0: 0ms sequenceNumber: 3 version: 1|0||56535e936e3a936d394e6c99 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.445+0000 d20010| 2015-11-23T18:44:35.444+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38778 #20 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.446+0000 d20010| 2015-11-23T18:44:35.445+0000 I SHARDING [conn20] remotely refreshing metadata for db0.coll0 with requested shard version 1|0||56535e936e3a936d394e6c99, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.446+0000 d20010| 2015-11-23T18:44:35.446+0000 I SHARDING [conn20] collection db0.coll0 was previously unsharded, new metadata loaded with shard version 1|0||56535e936e3a936d394e6c99 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.446+0000 d20010| 2015-11-23T18:44:35.446+0000 I SHARDING [conn20] collection version was loaded at version 1|0||56535e936e3a936d394e6c99, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.447+0000 s20019| 2015-11-23T18:44:35.446+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:35.446+0000-56535e936e3a936d394e6c9a", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304275446), what: "shardCollection.end", ns: "db0.coll0", details: { version: "1|0||56535e936e3a936d394e6c99" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.459+0000 s20019| 2015-11-23T18:44:35.459+0000 I SHARDING [conn1] distributed lock 'db0.coll0/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.459+0000 Using 10 threads (requested 10) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.641+0000 d20015| 2015-11-23T18:44:35.641+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.643+0000 s20020| 2015-11-23T18:44:35.643+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60245 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.643+0000 d20014| 2015-11-23T18:44:35.643+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.653+0000 d20013| 2015-11-23T18:44:35.652+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36362 #12 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.654+0000 d20013| 2015-11-23T18:44:35.652+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36363 #13 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.654+0000 d20013| 2015-11-23T18:44:35.654+0000 I NETWORK [conn13] end connection 10.99.163.247:36363 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.655+0000 d20015| 2015-11-23T18:44:35.653+0000 I REPL [SyncSourceFeedback] setting syncSourceFeedback to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.660+0000 d20013| 2015-11-23T18:44:35.655+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36364 #14 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.660+0000 d20014| 2015-11-23T18:44:35.660+0000 I REPL [SyncSourceFeedback] setting syncSourceFeedback to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.661+0000 d20013| 2015-11-23T18:44:35.660+0000 I NETWORK [conn12] end connection 10.99.163.247:36362 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.663+0000 d20013| 2015-11-23T18:44:35.662+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36366 #15 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.664+0000 s20020| 2015-11-23T18:44:35.663+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60249 #3 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.666+0000 d20014| 2015-11-23T18:44:35.665+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.666+0000 s20019| 2015-11-23T18:44:35.665+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39839 #2 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.671+0000 d20013| 2015-11-23T18:44:35.670+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36368 #16 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.671+0000 d20013| 2015-11-23T18:44:35.670+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36369 #17 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.672+0000 d20015| 2015-11-23T18:44:35.672+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.677+0000 s20019| 2015-11-23T18:44:35.677+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39842 #3 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.683+0000 s20020| 2015-11-23T18:44:35.682+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60255 #4 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.689+0000 s20019| 2015-11-23T18:44:35.689+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39844 #4 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.693+0000 s20019| 2015-11-23T18:44:35.693+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39845 #5 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.699+0000 s20019| 2015-11-23T18:44:35.698+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39847 #6 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.699+0000 s20020| 2015-11-23T18:44:35.698+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60258 #5 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.707+0000 s20020| 2015-11-23T18:44:35.707+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60260 #6 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.715+0000 setting random seed: 1821090012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.715+0000 setting random seed: 3700502729 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.715+0000 setting random seed: 4206440877 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.716+0000 setting random seed: 118859232 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.716+0000 setting random seed: 3824305791 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.716+0000 setting random seed: 1051807095 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.716+0000 setting random seed: 3415901949 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.718+0000 d20010| 2015-11-23T18:44:35.718+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38795 #21 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.719+0000 setting random seed: 2125995622 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.719+0000 d20010| 2015-11-23T18:44:35.718+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38796 #22 (16 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.720+0000 s20020| 2015-11-23T18:44:35.718+0000 I NETWORK [conn3] Starting new replica set monitor for test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.721+0000 s20020| 2015-11-23T18:44:35.718+0000 I NETWORK [conn3] Starting new replica set monitor for test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.721+0000 s20020| 2015-11-23T18:44:35.718+0000 I NETWORK [ReplicaSetMonitorWatcher] starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.722+0000 s20020| 2015-11-23T18:44:35.719+0000 I SHARDING [conn3] ChunkManager: time to load chunks for db0.coll0: 1ms sequenceNumber: 2 version: 1|0||56535e936e3a936d394e6c99 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.722+0000 setting random seed: 1953790632 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.724+0000 d20012| 2015-11-23T18:44:35.723+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:45998 #8 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.725+0000 d20010| 2015-11-23T18:44:35.724+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38797 #23 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.725+0000 d20010| 2015-11-23T18:44:35.725+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38799 #24 (18 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.727+0000 d20011| 2015-11-23T18:44:35.726+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49758 #8 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.727+0000 setting random seed: 281897721 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.731+0000 d20010| 2015-11-23T18:44:35.729+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38801 #25 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.731+0000 d20010| 2015-11-23T18:44:35.729+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38802 #26 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.731+0000 d20010| 2015-11-23T18:44:35.730+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38803 #27 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.732+0000 d20010| 2015-11-23T18:44:35.732+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38804 #28 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.735+0000 d20010| 2015-11-23T18:44:35.735+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38805 #29 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.736+0000 d20010| 2015-11-23T18:44:35.736+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38806 #30 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.738+0000 d20010| 2015-11-23T18:44:35.738+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38807 #31 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.739+0000 d20010| 2015-11-23T18:44:35.738+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38808 #32 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.743+0000 d20010| 2015-11-23T18:44:35.741+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38809 #33 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.743+0000 d20010| 2015-11-23T18:44:35.742+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38810 #34 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.748+0000 s20020| 2015-11-23T18:44:35.744+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.748+0000 d20010| 2015-11-23T18:44:35.748+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38811 #35 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.750+0000 s20020| 2015-11-23T18:44:35.749+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.751+0000 s20020| 2015-11-23T18:44:35.750+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.751+0000 d20010| 2015-11-23T18:44:35.751+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38812 #36 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.752+0000 d20010| 2015-11-23T18:44:35.751+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38813 #37 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.753+0000 d20010| 2015-11-23T18:44:35.752+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38814 #38 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.758+0000 d20010| 2015-11-23T18:44:35.758+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38815 #39 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.764+0000 d20010| 2015-11-23T18:44:35.764+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38816 #40 (34 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.765+0000 d20010| 2015-11-23T18:44:35.764+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38817 #41 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.775+0000 s20020| 2015-11-23T18:44:35.774+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.775+0000 d20010| 2015-11-23T18:44:35.775+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38818 #42 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.778+0000 d20010| 2015-11-23T18:44:35.778+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38819 #43 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.783+0000 d20010| 2015-11-23T18:44:35.783+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.785+0000 d20010| 2015-11-23T18:44:35.784+0000 I SHARDING [conn30] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 6.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.785+0000 c20016| 2015-11-23T18:44:35.785+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48388 #20 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.786+0000 d20010| 2015-11-23T18:44:35.786+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.787+0000 d20010| 2015-11-23T18:44:35.787+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.789+0000 d20010| 2015-11-23T18:44:35.787+0000 I SHARDING [conn20] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 6.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.790+0000 d20010| 2015-11-23T18:44:35.789+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.790+0000 d20010| 2015-11-23T18:44:35.789+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.791+0000 d20010| 2015-11-23T18:44:35.789+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.792+0000 d20010| 2015-11-23T18:44:35.791+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.793+0000 d20010| 2015-11-23T18:44:35.791+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38821 #44 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.796+0000 d20010| 2015-11-23T18:44:35.792+0000 I SHARDING [conn40] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.797+0000 d20010| 2015-11-23T18:44:35.792+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.799+0000 d20010| 2015-11-23T18:44:35.793+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.800+0000 d20010| 2015-11-23T18:44:35.793+0000 I SHARDING [conn38] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 6.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.801+0000 d20010| 2015-11-23T18:44:35.793+0000 I SHARDING [conn24] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.801+0000 c20017| 2015-11-23T18:44:35.793+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55704 #20 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.801+0000 d20010| 2015-11-23T18:44:35.793+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.802+0000 d20010| 2015-11-23T18:44:35.794+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.803+0000 d20010| 2015-11-23T18:44:35.795+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 12.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.803+0000 d20010| 2015-11-23T18:44:35.797+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38823 #45 (39 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.804+0000 d20010| 2015-11-23T18:44:35.798+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.804+0000 c20018| 2015-11-23T18:44:35.798+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36826 #20 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.806+0000 d20010| 2015-11-23T18:44:35.798+0000 I SHARDING [conn45] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 12.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.807+0000 d20010| 2015-11-23T18:44:35.799+0000 I SHARDING [conn36] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 10.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.808+0000 d20010| 2015-11-23T18:44:35.805+0000 I NETWORK [conn20] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.808+0000 d20010| 2015-11-23T18:44:35.805+0000 I NETWORK [conn26] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.809+0000 d20010| 2015-11-23T18:44:35.805+0000 I NETWORK [conn40] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.809+0000 d20010| 2015-11-23T18:44:35.805+0000 I NETWORK [conn39] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.810+0000 d20010| 2015-11-23T18:44:35.805+0000 I NETWORK [conn24] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.810+0000 d20010| 2015-11-23T18:44:35.805+0000 I NETWORK [conn38] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.811+0000 d20010| 2015-11-23T18:44:35.805+0000 I NETWORK [conn37] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.811+0000 d20010| 2015-11-23T18:44:35.806+0000 I NETWORK [conn36] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.811+0000 d20010| 2015-11-23T18:44:35.806+0000 I NETWORK [conn45] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.812+0000 d20010| 2015-11-23T18:44:35.807+0000 I SHARDING [LockPinger] creating distributed lock ping thread for ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 and process ip-10-99-163-247:20010:1448304275:1756418349 (sleeping for 30000ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.812+0000 d20010| 2015-11-23T18:44:35.808+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.813+0000 c20016| 2015-11-23T18:44:35.808+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48393 #21 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.813+0000 d20010| 2015-11-23T18:44:35.811+0000 I NETWORK [conn26] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.814+0000 c20016| 2015-11-23T18:44:35.811+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48394 #22 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.815+0000 c20016| 2015-11-23T18:44:35.811+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48395 #23 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.815+0000 d20010| 2015-11-23T18:44:35.811+0000 I NETWORK [conn24] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.815+0000 d20010| 2015-11-23T18:44:35.812+0000 I NETWORK [conn39] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.816+0000 c20016| 2015-11-23T18:44:35.812+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48396 #24 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.816+0000 c20016| 2015-11-23T18:44:35.812+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48397 #25 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.817+0000 c20017| 2015-11-23T18:44:35.812+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55713 #21 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.817+0000 c20016| 2015-11-23T18:44:35.813+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48398 #26 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.818+0000 c20016| 2015-11-23T18:44:35.813+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48400 #27 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.818+0000 d20010| 2015-11-23T18:44:35.813+0000 I NETWORK [conn36] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.820+0000 c20017| 2015-11-23T18:44:35.813+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55715 #22 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.820+0000 c20017| 2015-11-23T18:44:35.813+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55716 #23 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.821+0000 d20010| 2015-11-23T18:44:35.813+0000 I NETWORK [conn26] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.822+0000 d20010| 2015-11-23T18:44:35.813+0000 I NETWORK [conn37] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.822+0000 c20017| 2015-11-23T18:44:35.814+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55717 #24 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.823+0000 d20010| 2015-11-23T18:44:35.813+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.825+0000 d20010| 2015-11-23T18:44:35.814+0000 I NETWORK [conn24] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.826+0000 c20016| 2015-11-23T18:44:35.814+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48404 #28 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.827+0000 d20010| 2015-11-23T18:44:35.814+0000 I NETWORK [conn40] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.827+0000 c20016| 2015-11-23T18:44:35.814+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48405 #29 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.828+0000 c20018| 2015-11-23T18:44:35.814+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36840 #21 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.828+0000 d20010| 2015-11-23T18:44:35.814+0000 I NETWORK [conn20] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.829+0000 c20017| 2015-11-23T18:44:35.815+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55721 #25 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.829+0000 d20010| 2015-11-23T18:44:35.815+0000 I NETWORK [conn39] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.830+0000 c20017| 2015-11-23T18:44:35.815+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55722 #26 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.830+0000 c20017| 2015-11-23T18:44:35.815+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55723 #27 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.831+0000 d20010| 2015-11-23T18:44:35.815+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.831+0000 d20010| 2015-11-23T18:44:35.816+0000 I NETWORK [conn36] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.832+0000 d20010| 2015-11-23T18:44:35.816+0000 I NETWORK [conn20] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.832+0000 c20018| 2015-11-23T18:44:35.816+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36844 #22 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.832+0000 c20018| 2015-11-23T18:44:35.816+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36845 #23 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.833+0000 d20010| 2015-11-23T18:44:35.816+0000 I NETWORK [conn40] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.834+0000 c20018| 2015-11-23T18:44:35.817+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36846 #24 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.834+0000 c20018| 2015-11-23T18:44:35.817+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36847 #25 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.835+0000 d20010| 2015-11-23T18:44:35.817+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.836+0000 s20019| 2015-11-23T18:44:35.817+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.837+0000 c20018| 2015-11-23T18:44:35.818+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36848 #26 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.838+0000 d20010| 2015-11-23T18:44:35.819+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.840+0000 s20020| 2015-11-23T18:44:35.819+0000 W SHARDING [conn2] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.840+0000 c20016| 2015-11-23T18:44:35.819+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48415 #30 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.841+0000 c20017| 2015-11-23T18:44:35.819+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55730 #28 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.841+0000 d20010| 2015-11-23T18:44:35.819+0000 I NETWORK [conn45] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.842+0000 d20010| 2015-11-23T18:44:35.820+0000 I NETWORK [conn37] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.842+0000 d20010| 2015-11-23T18:44:35.820+0000 W SHARDING [conn24] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.845+0000 s20019| 2015-11-23T18:44:35.820+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.845+0000 c20018| 2015-11-23T18:44:35.821+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36851 #27 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.845+0000 d20010| 2015-11-23T18:44:35.822+0000 I NETWORK [conn38] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.846+0000 d20010| 2015-11-23T18:44:35.823+0000 W SHARDING [conn20] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.848+0000 s20019| 2015-11-23T18:44:35.823+0000 W SHARDING [conn3] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 6.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.848+0000 d20010| 2015-11-23T18:44:35.825+0000 W SHARDING [conn40] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.849+0000 c20017| 2015-11-23T18:44:35.825+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55732 #29 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.855+0000 s20020| 2015-11-23T18:44:35.826+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.865+0000 c20017| 2015-11-23T18:44:35.826+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55733 #30 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.866+0000 c20016| 2015-11-23T18:44:35.829+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48420 #31 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.866+0000 d20010| 2015-11-23T18:44:35.833+0000 I NETWORK [conn38] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.867+0000 c20018| 2015-11-23T18:44:35.839+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36855 #28 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.871+0000 d20010| 2015-11-23T18:44:35.840+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.872+0000 d20010| 2015-11-23T18:44:35.841+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 8.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.872+0000 d20010| 2015-11-23T18:44:35.842+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.873+0000 d20010| 2015-11-23T18:44:35.843+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.886+0000 s20019| 2015-11-23T18:44:35.843+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 8.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.890+0000 d20010| 2015-11-23T18:44:35.844+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 8.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.891+0000 c20018| 2015-11-23T18:44:35.844+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36856 #29 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.892+0000 d20010| 2015-11-23T18:44:35.844+0000 W SHARDING [conn36] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.893+0000 s20020| 2015-11-23T18:44:35.844+0000 W SHARDING [conn3] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 10.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.894+0000 d20010| 2015-11-23T18:44:35.852+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.906+0000 s20019| 2015-11-23T18:44:35.852+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 8.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.907+0000 d20010| 2015-11-23T18:44:35.854+0000 I NETWORK [conn45] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.907+0000 d20010| 2015-11-23T18:44:35.860+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.923+0000 d20010| 2015-11-23T18:44:35.861+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.924+0000 d20010| 2015-11-23T18:44:35.862+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.927+0000 d20010| 2015-11-23T18:44:35.862+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.929+0000 d20010| 2015-11-23T18:44:35.863+0000 I SHARDING [conn20] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.933+0000 s20020| 2015-11-23T18:44:35.863+0000 W SHARDING [conn3] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.937+0000 d20010| 2015-11-23T18:44:35.864+0000 W SHARDING [conn20] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.943+0000 s20019| 2015-11-23T18:44:35.864+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.947+0000 d20010| 2015-11-23T18:44:35.869+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.953+0000 s20020| 2015-11-23T18:44:35.869+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 12.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.955+0000 c20018| 2015-11-23T18:44:35.875+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36857 #30 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.955+0000 d20010| 2015-11-23T18:44:35.880+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.956+0000 d20010| 2015-11-23T18:44:35.880+0000 W SHARDING [conn38] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.965+0000 s20020| 2015-11-23T18:44:35.881+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 6.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:35.969+0000 d20010| 2015-11-23T18:44:35.882+0000 W SHARDING [conn45] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.002+0000 s20019| 2015-11-23T18:44:35.882+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 12.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.003+0000 d20010| 2015-11-23T18:44:35.885+0000 I SHARDING [conn38] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 }, { j: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.004+0000 d20010| 2015-11-23T18:44:35.900+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.005+0000 d20010| 2015-11-23T18:44:35.901+0000 I SHARDING [conn45] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 }, { j: 14.0 }, { j: 24.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.005+0000 d20010| 2015-11-23T18:44:35.902+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.028+0000 d20010| 2015-11-23T18:44:35.902+0000 W SHARDING [conn45] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.029+0000 s20019| 2015-11-23T18:44:35.903+0000 W SHARDING [conn3] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 }, { j: 14.0 }, { j: 24.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.031+0000 s20020| 2015-11-23T18:44:35.903+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 }, { j: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.032+0000 d20010| 2015-11-23T18:44:35.903+0000 I SHARDING [conn20] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 }, { j: 14.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.033+0000 d20010| 2015-11-23T18:44:35.903+0000 W SHARDING [conn38] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.037+0000 d20010| 2015-11-23T18:44:35.913+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.043+0000 d20010| 2015-11-23T18:44:35.913+0000 I SHARDING [conn31] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.044+0000 d20010| 2015-11-23T18:44:35.913+0000 I SHARDING [conn38] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.045+0000 d20010| 2015-11-23T18:44:35.914+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 16.0 }, { j: 32.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.046+0000 d20010| 2015-11-23T18:44:35.915+0000 W SHARDING [conn38] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.047+0000 d20010| 2015-11-23T18:44:35.915+0000 W SHARDING [conn20] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.049+0000 s20020| 2015-11-23T18:44:35.915+0000 W SHARDING [conn3] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.053+0000 s20019| 2015-11-23T18:44:35.915+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 10.0 }, { j: 14.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.053+0000 c20016| 2015-11-23T18:44:35.915+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48424 #32 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.057+0000 d20010| 2015-11-23T18:44:35.933+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.062+0000 s20020| 2015-11-23T18:44:35.933+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 16.0 }, { j: 32.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.063+0000 d20010| 2015-11-23T18:44:35.938+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.068+0000 d20010| 2015-11-23T18:44:35.938+0000 I SHARDING [conn20] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 16.0 }, { j: 20.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.071+0000 d20010| 2015-11-23T18:44:35.940+0000 W SHARDING [conn20] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.074+0000 s20019| 2015-11-23T18:44:35.940+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 16.0 }, { j: 20.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.078+0000 d20010| 2015-11-23T18:44:35.950+0000 I SHARDING [conn31] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.083+0000 d20010| 2015-11-23T18:44:35.950+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.088+0000 d20010| 2015-11-23T18:44:35.951+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 24.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.091+0000 d20010| 2015-11-23T18:44:35.952+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.092+0000 d20010| 2015-11-23T18:44:35.952+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.102+0000 d20010| 2015-11-23T18:44:35.953+0000 I SHARDING [conn36] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.103+0000 d20010| 2015-11-23T18:44:35.953+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.105+0000 s20020| 2015-11-23T18:44:35.953+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 24.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.106+0000 d20010| 2015-11-23T18:44:35.954+0000 W SHARDING [conn36] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.107+0000 s20020| 2015-11-23T18:44:35.954+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.109+0000 s20020| 2015-11-23T18:44:35.954+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.110+0000 d20010| 2015-11-23T18:44:35.954+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.125+0000 d20010| 2015-11-23T18:44:35.960+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.137+0000 d20010| 2015-11-23T18:44:35.960+0000 I SHARDING [conn20] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.140+0000 d20010| 2015-11-23T18:44:35.962+0000 W SHARDING [conn20] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.144+0000 s20019| 2015-11-23T18:44:35.963+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 18.0 }, { j: 22.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.146+0000 d20010| 2015-11-23T18:44:35.973+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.149+0000 d20010| 2015-11-23T18:44:35.974+0000 I SHARDING [conn20] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 30.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.151+0000 d20010| 2015-11-23T18:44:35.975+0000 W SHARDING [conn20] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.157+0000 s20019| 2015-11-23T18:44:35.976+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 30.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.158+0000 d20010| 2015-11-23T18:44:35.980+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.158+0000 d20010| 2015-11-23T18:44:35.982+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.164+0000 d20010| 2015-11-23T18:44:35.982+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 26.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.165+0000 d20010| 2015-11-23T18:44:35.983+0000 I SHARDING [conn36] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 26.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.167+0000 d20010| 2015-11-23T18:44:35.984+0000 W SHARDING [conn36] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.168+0000 d20010| 2015-11-23T18:44:35.984+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.169+0000 s20020| 2015-11-23T18:44:35.984+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 26.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.172+0000 s20020| 2015-11-23T18:44:35.984+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 26.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.172+0000 d20010| 2015-11-23T18:44:35.989+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.173+0000 d20010| 2015-11-23T18:44:35.991+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 24.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.174+0000 d20010| 2015-11-23T18:44:35.993+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.176+0000 s20019| 2015-11-23T18:44:35.993+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 24.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.176+0000 d20010| 2015-11-23T18:44:36.013+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.177+0000 d20010| 2015-11-23T18:44:36.014+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 24.0 }, { j: 28.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.178+0000 d20010| 2015-11-23T18:44:36.016+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.179+0000 s20020| 2015-11-23T18:44:36.017+0000 W SHARDING [conn3] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 24.0 }, { j: 28.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.180+0000 d20010| 2015-11-23T18:44:36.022+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.181+0000 d20010| 2015-11-23T18:44:36.024+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 24.0 }, { j: 28.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.182+0000 d20010| 2015-11-23T18:44:36.025+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.183+0000 s20019| 2015-11-23T18:44:36.025+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 24.0 }, { j: 28.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.184+0000 c20016| 2015-11-23T18:44:36.036+0000 I COMMAND [conn27] command admin.$cmd command: getLastError { getlasterror: 1, fsync: 1 } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:168 locks:{} protocol:op_command 115ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.184+0000 c20016| 2015-11-23T18:44:36.036+0000 I COMMAND [conn18] command admin.$cmd command: getLastError { getlasterror: 1, fsync: 1 } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:91 locks:{} protocol:op_command 115ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.185+0000 d20010| 2015-11-23T18:44:36.036+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.185+0000 d20010| 2015-11-23T18:44:36.037+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.187+0000 d20010| 2015-11-23T18:44:36.038+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 32.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.188+0000 d20010| 2015-11-23T18:44:36.039+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 32.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.189+0000 d20010| 2015-11-23T18:44:36.040+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.192+0000 s20020| 2015-11-23T18:44:36.040+0000 W SHARDING [conn2] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 32.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.193+0000 d20010| 2015-11-23T18:44:36.040+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.194+0000 s20020| 2015-11-23T18:44:36.041+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 32.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.194+0000 d20010| 2015-11-23T18:44:36.051+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.195+0000 d20010| 2015-11-23T18:44:36.052+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.196+0000 d20010| 2015-11-23T18:44:36.052+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 30.0 }, { j: 44.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.198+0000 d20010| 2015-11-23T18:44:36.052+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 30.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.198+0000 d20010| 2015-11-23T18:44:36.053+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.200+0000 s20019| 2015-11-23T18:44:36.053+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 30.0 }, { j: 44.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.201+0000 s20020| 2015-11-23T18:44:36.054+0000 W SHARDING [conn2] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 30.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.202+0000 d20010| 2015-11-23T18:44:36.054+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.202+0000 d20010| 2015-11-23T18:44:36.066+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.204+0000 d20010| 2015-11-23T18:44:36.067+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 28.0 }, { j: 34.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.204+0000 d20010| 2015-11-23T18:44:36.068+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.206+0000 s20020| 2015-11-23T18:44:36.069+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 28.0 }, { j: 34.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.206+0000 d20010| 2015-11-23T18:44:36.075+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.207+0000 d20010| 2015-11-23T18:44:36.076+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 28.0 }, { j: 32.0 }, { j: 40.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.208+0000 d20010| 2015-11-23T18:44:36.078+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.209+0000 s20020| 2015-11-23T18:44:36.078+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 26.0 }, { j: 28.0 }, { j: 32.0 }, { j: 40.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.210+0000 d20010| 2015-11-23T18:44:36.085+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.210+0000 d20010| 2015-11-23T18:44:36.085+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.211+0000 d20010| 2015-11-23T18:44:36.088+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38857 #46 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.212+0000 d20010| 2015-11-23T18:44:36.089+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 32.0 }, { j: 36.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.212+0000 d20010| 2015-11-23T18:44:36.089+0000 I SHARDING [conn17] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.213+0000 d20010| 2015-11-23T18:44:36.090+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.214+0000 s20019| 2015-11-23T18:44:36.090+0000 W SHARDING [conn3] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 32.0 }, { j: 36.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.215+0000 d20010| 2015-11-23T18:44:36.091+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 32.0 }, { j: 36.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.216+0000 s20019| 2015-11-23T18:44:36.092+0000 W SHARDING [conn5] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 32.0 }, { j: 36.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.217+0000 d20010| 2015-11-23T18:44:36.092+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.217+0000 d20010| 2015-11-23T18:44:36.098+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.218+0000 d20010| 2015-11-23T18:44:36.099+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 48.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.218+0000 d20010| 2015-11-23T18:44:36.101+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.219+0000 d20010| 2015-11-23T18:44:36.101+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.220+0000 s20020| 2015-11-23T18:44:36.101+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 48.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.221+0000 d20010| 2015-11-23T18:44:36.103+0000 I SHARDING [conn39] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 46.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.221+0000 d20010| 2015-11-23T18:44:36.104+0000 W SHARDING [conn39] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.223+0000 s20020| 2015-11-23T18:44:36.104+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 46.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.223+0000 d20010| 2015-11-23T18:44:36.108+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38858 #47 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.223+0000 s20019| 2015-11-23T18:44:36.109+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.223+0000 d20010| 2015-11-23T18:44:36.116+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.224+0000 d20010| 2015-11-23T18:44:36.117+0000 I SHARDING [conn36] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 38.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.225+0000 d20010| 2015-11-23T18:44:36.118+0000 W SHARDING [conn36] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.226+0000 s20020| 2015-11-23T18:44:36.118+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 38.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.227+0000 d20010| 2015-11-23T18:44:36.121+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 32.0 }, { j: 36.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.227+0000 d20010| 2015-11-23T18:44:36.122+0000 W SHARDING [conn26] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.229+0000 s20019| 2015-11-23T18:44:36.122+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 32.0 }, { j: 36.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.229+0000 d20010| 2015-11-23T18:44:36.126+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.230+0000 d20010| 2015-11-23T18:44:36.127+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 36.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.231+0000 d20010| 2015-11-23T18:44:36.128+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.232+0000 s20020| 2015-11-23T18:44:36.129+0000 W SHARDING [conn4] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 36.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.233+0000 s20020| 2015-11-23T18:44:36.137+0000 I NETWORK [conn3] end connection 10.99.163.247:60249 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.233+0000 d20010| 2015-11-23T18:44:36.138+0000 I SHARDING [conn46] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.234+0000 d20010| 2015-11-23T18:44:36.140+0000 I SHARDING [conn45] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 36.0 }, { j: 40.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.234+0000 d20010| 2015-11-23T18:44:36.141+0000 W SHARDING [conn45] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.236+0000 s20019| 2015-11-23T18:44:36.141+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 34.0 }, { j: 36.0 }, { j: 40.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.236+0000 d20010| 2015-11-23T18:44:36.148+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.237+0000 d20010| 2015-11-23T18:44:36.148+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 32.0 }, { j: 36.0 }, { j: 38.0 }, { j: 46.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.237+0000 s20019| 2015-11-23T18:44:36.149+0000 I NETWORK [conn4] end connection 10.99.163.247:39844 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.238+0000 d20010| 2015-11-23T18:44:36.150+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.239+0000 s20020| 2015-11-23T18:44:36.150+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 32.0 }, { j: 36.0 }, { j: 38.0 }, { j: 46.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.239+0000 s20020| 2015-11-23T18:44:36.161+0000 I NETWORK [conn4] end connection 10.99.163.247:60255 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.240+0000 c20016| 2015-11-23T18:44:36.165+0000 I COMMAND [conn18] command admin.$cmd command: getLastError { getlasterror: 1, fsync: 1 } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:109 locks:{} protocol:op_command 104ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.240+0000 c20016| 2015-11-23T18:44:36.165+0000 I COMMAND [conn27] command admin.$cmd command: getLastError { getlasterror: 1, fsync: 1 } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:91 locks:{} protocol:op_command 104ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.240+0000 s20020| 2015-11-23T18:44:36.171+0000 I NETWORK [conn2] end connection 10.99.163.247:60245 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.241+0000 s20019| 2015-11-23T18:44:36.172+0000 I NETWORK [conn5] end connection 10.99.163.247:39845 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.241+0000 s20019| 2015-11-23T18:44:36.174+0000 I NETWORK [conn3] end connection 10.99.163.247:39842 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.241+0000 s20020| 2015-11-23T18:44:36.182+0000 I NETWORK [conn5] end connection 10.99.163.247:60258 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.241+0000 d20010| 2015-11-23T18:44:36.189+0000 I SHARDING [conn35] request split points lookup for chunk db0.coll0 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.242+0000 d20010| 2015-11-23T18:44:36.190+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 32.0 }, { j: 36.0 }, { j: 38.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.244+0000 s20020| 2015-11-23T18:44:36.191+0000 W SHARDING [conn6] splitChunk failed - cmd: { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 6.0 }, { j: 8.0 }, { j: 12.0 }, { j: 14.0 }, { j: 16.0 }, { j: 20.0 }, { j: 22.0 }, { j: 24.0 }, { j: 28.0 }, { j: 30.0 }, { j: 32.0 }, { j: 36.0 }, { j: 38.0 }, { j: 42.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.244+0000 d20010| 2015-11-23T18:44:36.191+0000 W SHARDING [conn37] could not acquire collection lock for db0.coll0 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.244+0000 s20019| 2015-11-23T18:44:36.194+0000 I NETWORK [conn6] end connection 10.99.163.247:39847 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.245+0000 s20020| 2015-11-23T18:44:36.209+0000 I NETWORK [conn6] end connection 10.99.163.247:60260 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.286+0000 d20010| 2015-11-23T18:44:36.286+0000 I SHARDING [conn30] distributed lock 'db0.coll0/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ j: MinKey }, { j: MaxKey }) in db0.coll0', ts : 56535e93a5e5109e4dbf13a5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.287+0000 d20010| 2015-11-23T18:44:36.286+0000 I SHARDING [conn30] remotely refreshing metadata for db0.coll0 based on current shard version 1|0||56535e936e3a936d394e6c99, current metadata version is 1|0||56535e936e3a936d394e6c99 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.288+0000 d20010| 2015-11-23T18:44:36.286+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:44:35.820+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20010:1448304275:1756418349', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.288+0000 d20010| 2015-11-23T18:44:36.286+0000 I SHARDING [LockPinger] trying to delete 45 old lock entries for process ip-10-99-163-247:20010:1448304275:1756418349 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.289+0000 d20010| 2015-11-23T18:44:36.286+0000 I SHARDING [conn30] metadata of collection db0.coll0 already up to date (shard version : 1|0||56535e936e3a936d394e6c99, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.289+0000 d20010| 2015-11-23T18:44:36.286+0000 I SHARDING [conn30] splitChunk accepted at version 1|0||56535e936e3a936d394e6c99 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.289+0000 d20010| 2015-11-23T18:44:36.287+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13a9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.450+0000 d20010| 2015-11-23T18:44:36.449+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13a7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.451+0000 d20010| 2015-11-23T18:44:36.451+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13ab [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.452+0000 c20016| 2015-11-23T18:44:36.452+0000 I NETWORK [conn18] end connection 10.99.163.247:48337 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.453+0000 d20010| 2015-11-23T18:44:36.451+0000 I NETWORK [conn30] scoped connection to ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.454+0000 d20010| 2015-11-23T18:44:36.452+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:36.451+0000-56535e94a5e5109e4dbf13d3", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304276451), what: "multi-split", ns: "db0.coll0", details: { before: { min: { j: MinKey }, max: { j: MaxKey } }, number: 1, of: 3, chunk: { min: { j: MinKey }, max: { j: 0.0 }, lastmod: Timestamp 1000|1, lastmodEpoch: ObjectId('56535e936e3a936d394e6c99') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.454+0000 c20017| 2015-11-23T18:44:36.452+0000 I NETWORK [conn18] end connection 10.99.163.247:55652 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.454+0000 c20018| 2015-11-23T18:44:36.452+0000 I NETWORK [conn18] end connection 10.99.163.247:36773 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.455+0000 c20016| 2015-11-23T18:44:36.453+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48427 #33 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.455+0000 d20010| 2015-11-23T18:44:36.453+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13a6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.455+0000 c20017| 2015-11-23T18:44:36.454+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55742 #31 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.456+0000 d20010| 2015-11-23T18:44:36.454+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13a8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.456+0000 c20018| 2015-11-23T18:44:36.455+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36863 #31 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.457+0000 d20010| 2015-11-23T18:44:36.456+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13af [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.458+0000 d20010| 2015-11-23T18:44:36.456+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:36.456+0000-56535e94a5e5109e4dbf13d4", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304276456), what: "multi-split", ns: "db0.coll0", details: { before: { min: { j: MinKey }, max: { j: MaxKey } }, number: 2, of: 3, chunk: { min: { j: 0.0 }, max: { j: 6.0 }, lastmod: Timestamp 1000|2, lastmodEpoch: ObjectId('56535e936e3a936d394e6c99') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.539+0000 d20010| 2015-11-23T18:44:36.538+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:36.538+0000-56535e94a5e5109e4dbf13d5", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304276538), what: "multi-split", ns: "db0.coll0", details: { before: { min: { j: MinKey }, max: { j: MaxKey } }, number: 3, of: 3, chunk: { min: { j: 6.0 }, max: { j: MaxKey }, lastmod: Timestamp 1000|3, lastmodEpoch: ObjectId('56535e936e3a936d394e6c99') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.678+0000 c20016| 2015-11-23T18:44:36.678+0000 I COMMAND [conn27] command admin.$cmd command: getLastError { getlasterror: 1, fsync: 1 } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:109 locks:{} protocol:op_command 139ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.679+0000 c20017| 2015-11-23T18:44:36.678+0000 I COMMAND [conn31] command admin.$cmd command: getLastError { getLastError: true, fsync: true } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:91 locks:{} protocol:op_command 139ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.680+0000 c20016| 2015-11-23T18:44:36.678+0000 I COMMAND [conn33] command admin.$cmd command: getLastError { getLastError: true, fsync: true } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:91 locks:{} protocol:op_command 139ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.680+0000 d20010| 2015-11-23T18:44:36.678+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13ad [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.680+0000 d20010| 2015-11-23T18:44:36.679+0000 I NETWORK [conn30] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.681+0000 c20016| 2015-11-23T18:44:36.680+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48430 #34 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.681+0000 d20010| 2015-11-23T18:44:36.680+0000 I NETWORK [conn30] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.681+0000 c20017| 2015-11-23T18:44:36.681+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55745 #32 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.682+0000 d20010| 2015-11-23T18:44:36.681+0000 I NETWORK [conn30] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.682+0000 c20018| 2015-11-23T18:44:36.681+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36866 #32 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.785+0000 d20010| 2015-11-23T18:44:36.784+0000 I SHARDING [conn30] distributed lock 'db0.coll0/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.785+0000 d20010| 2015-11-23T18:44:36.784+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.787+0000 d20010| 2015-11-23T18:44:36.784+0000 I COMMAND [conn30] command admin.$cmd command: splitChunk { splitChunk: "db0.coll0", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs0", splitKeys: [ { j: 0.0 }, { j: 6.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535e936e3a936d394e6c99') ], epoch: ObjectId('56535e936e3a936d394e6c99') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:134 locks:{ Global: { acquireCount: { r: 4, w: 2 } }, Database: { acquireCount: { r: 1, w: 2 } }, Collection: { acquireCount: { r: 1, W: 2 } } } protocol:op_command 1000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.787+0000 d20010| 2015-11-23T18:44:36.785+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.787+0000 s20019| 2015-11-23T18:44:36.785+0000 I SHARDING [conn2] ChunkManager: time to load chunks for db0.coll0: 0ms sequenceNumber: 4 version: 1|3||56535e936e3a936d394e6c99 based on: 1|0||56535e936e3a936d394e6c99 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.788+0000 s20019| 2015-11-23T18:44:36.786+0000 I SHARDING [conn2] autosplitted db0.coll0 shard: ns: db0.coll0, shard: test-rs0, lastmod: 1|0||56535e936e3a936d394e6c99, min: { j: MinKey }, max: { j: MaxKey } into 3 (splitThreshold 921) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.788+0000 d20010| 2015-11-23T18:44:36.786+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.789+0000 d20010| 2015-11-23T18:44:36.788+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13ac [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.791+0000 d20010| 2015-11-23T18:44:36.790+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13aa [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.793+0000 d20010| 2015-11-23T18:44:36.793+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13ae [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.795+0000 d20010| 2015-11-23T18:44:36.795+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.797+0000 d20010| 2015-11-23T18:44:36.797+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.799+0000 d20010| 2015-11-23T18:44:36.798+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.801+0000 d20010| 2015-11-23T18:44:36.800+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.803+0000 d20010| 2015-11-23T18:44:36.802+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.804+0000 d20010| 2015-11-23T18:44:36.803+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.805+0000 d20010| 2015-11-23T18:44:36.804+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13b9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.806+0000 d20010| 2015-11-23T18:44:36.806+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13bb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.808+0000 d20010| 2015-11-23T18:44:36.807+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13ba [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.810+0000 d20010| 2015-11-23T18:44:36.809+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13bc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.811+0000 d20010| 2015-11-23T18:44:36.811+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13bd [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.813+0000 d20010| 2015-11-23T18:44:36.812+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13bf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.814+0000 d20010| 2015-11-23T18:44:36.813+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13be [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.815+0000 d20010| 2015-11-23T18:44:36.814+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e93a5e5109e4dbf13c0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.816+0000 d20010| 2015-11-23T18:44:36.816+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.817+0000 d20010| 2015-11-23T18:44:36.817+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.818+0000 d20010| 2015-11-23T18:44:36.818+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.819+0000 d20010| 2015-11-23T18:44:36.819+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.821+0000 d20010| 2015-11-23T18:44:36.820+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.822+0000 d20010| 2015-11-23T18:44:36.822+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.824+0000 d20010| 2015-11-23T18:44:36.823+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.825+0000 d20010| 2015-11-23T18:44:36.824+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.827+0000 d20010| 2015-11-23T18:44:36.826+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13c9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.828+0000 d20010| 2015-11-23T18:44:36.827+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13ca [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.829+0000 d20010| 2015-11-23T18:44:36.829+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13cb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.831+0000 d20010| 2015-11-23T18:44:36.830+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13cc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.832+0000 d20010| 2015-11-23T18:44:36.832+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13cd [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.834+0000 d20010| 2015-11-23T18:44:36.833+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13ce [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.835+0000 d20010| 2015-11-23T18:44:36.835+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13cf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.837+0000 d20010| 2015-11-23T18:44:36.837+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13d0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.839+0000 d20010| 2015-11-23T18:44:36.838+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13d1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.840+0000 d20010| 2015-11-23T18:44:36.840+0000 I SHARDING [LockPinger] handled late remove of old distributed lock with ts 56535e94a5e5109e4dbf13d2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.865+0000 s20019| 2015-11-23T18:44:36.865+0000 I NETWORK [conn2] end connection 10.99.163.247:39839 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.868+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.869+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.869+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.870+0000 Workload(s) completed in 1409 ms: jstests/concurrency/fsm_workloads/explain_group.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.870+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.870+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.870+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.870+0000 s20019| 2015-11-23T18:44:36.868+0000 I COMMAND [conn1] DROP: db0.coll0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.871+0000 s20019| 2015-11-23T18:44:36.868+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:36.868+0000-56535e946e3a936d394e6c9b", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304276868), what: "dropCollection.start", ns: "db0.coll0", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.894+0000 s20019| 2015-11-23T18:44:36.893+0000 I SHARDING [conn1] distributed lock 'db0.coll0/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535e946e3a936d394e6c9c [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.894+0000 d20010| 2015-11-23T18:44:36.893+0000 I COMMAND [conn46] CMD: drop db0.coll0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.895+0000 d20013| 2015-11-23T18:44:36.895+0000 I COMMAND [conn11] CMD: drop db0.coll0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.895+0000 d20011| 2015-11-23T18:44:36.895+0000 I COMMAND [repl writer worker 0] CMD: drop db0.coll0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.896+0000 d20012| 2015-11-23T18:44:36.895+0000 I COMMAND [repl writer worker 1] CMD: drop db0.coll0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.907+0000 d20010| 2015-11-23T18:44:36.906+0000 I SHARDING [conn46] remotely refreshing metadata for db0.coll0 with requested shard version 0|0||000000000000000000000000, current shard version is 1|3||56535e936e3a936d394e6c99, current metadata version is 1|3||56535e936e3a936d394e6c99 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.908+0000 d20010| 2015-11-23T18:44:36.906+0000 I SHARDING [conn46] dropping metadata for db0.coll0 at shard version 1|3||56535e936e3a936d394e6c99, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.908+0000 s20019| 2015-11-23T18:44:36.907+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:36.907+0000-56535e946e3a936d394e6c9d", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304276907), what: "dropCollection", ns: "db0.coll0", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.937+0000 s20019| 2015-11-23T18:44:36.937+0000 I SHARDING [conn1] distributed lock 'db0.coll0/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.938+0000 s20019| 2015-11-23T18:44:36.938+0000 I COMMAND [conn1] DROP DATABASE: db0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.939+0000 s20019| 2015-11-23T18:44:36.938+0000 I SHARDING [conn1] DBConfig::dropDatabase: db0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.940+0000 s20019| 2015-11-23T18:44:36.938+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:36.938+0000-56535e946e3a936d394e6c9e", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304276938), what: "dropDatabase.start", ns: "db0", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.944+0000 s20019| 2015-11-23T18:44:36.943+0000 I SHARDING [conn1] DBConfig::dropDatabase: db0 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.944+0000 d20010| 2015-11-23T18:44:36.944+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38865 #48 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.945+0000 d20010| 2015-11-23T18:44:36.945+0000 I COMMAND [conn48] dropDatabase db0 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.945+0000 d20010| 2015-11-23T18:44:36.945+0000 I COMMAND [conn48] dropDatabase db0 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.946+0000 s20019| 2015-11-23T18:44:36.945+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:36.945+0000-56535e946e3a936d394e6c9f", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304276945), what: "dropDatabase", ns: "db0", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.947+0000 d20012| 2015-11-23T18:44:36.945+0000 I COMMAND [repl writer worker 15] dropDatabase db0 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.947+0000 d20012| 2015-11-23T18:44:36.946+0000 I COMMAND [repl writer worker 15] dropDatabase db0 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.947+0000 d20011| 2015-11-23T18:44:36.945+0000 I COMMAND [repl writer worker 4] dropDatabase db0 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.947+0000 d20011| 2015-11-23T18:44:36.945+0000 I COMMAND [repl writer worker 4] dropDatabase db0 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.954+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.954+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.954+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.954+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.954+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.954+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.954+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.979+0000 d20010| 2015-11-23T18:44:36.978+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.979+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.980+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.980+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.980+0000 ip-10-99-163-247:20010: awaitReplication completed in 26 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.980+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.980+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.980+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.984+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 d20012| 2015-11-23T18:44:36.979+0000 I COMMAND [repl writer worker 7] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:36.985+0000 d20011| 2015-11-23T18:44:36.979+0000 I COMMAND [repl writer worker 7] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.009+0000 d20013| 2015-11-23T18:44:37.008+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.009+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.009+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.009+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.009+0000 ip-10-99-163-247:20013: awaitReplication completed in 30 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.010+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.010+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.010+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.010+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.010+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.010+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.010+0000 Workload(s) started: jstests/concurrency/fsm_workloads/create_capped_collection.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.011+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.011+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.011+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.011+0000 d20015| 2015-11-23T18:44:37.009+0000 I COMMAND [repl writer worker 4] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.011+0000 d20014| 2015-11-23T18:44:37.010+0000 I COMMAND [repl writer worker 5] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.025+0000 s20019| 2015-11-23T18:44:37.025+0000 I SHARDING [conn1] distributed lock 'db1/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535e956e3a936d394e6ca0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.027+0000 s20019| 2015-11-23T18:44:37.026+0000 I SHARDING [conn1] Placing [db1] on: test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.027+0000 s20019| 2015-11-23T18:44:37.026+0000 I SHARDING [conn1] Enabling sharding for database [db1] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.034+0000 s20019| 2015-11-23T18:44:37.033+0000 I SHARDING [conn1] distributed lock 'db1/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.042+0000 d20010| 2015-11-23T18:44:37.041+0000 I INDEX [conn19] build index on: db1.coll1 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db1.coll1" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.042+0000 d20010| 2015-11-23T18:44:37.042+0000 I INDEX [conn19] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.044+0000 d20010| 2015-11-23T18:44:37.043+0000 I INDEX [conn19] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.044+0000 s20019| 2015-11-23T18:44:37.044+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db1.coll1", key: { _id: "hashed" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.051+0000 d20012| 2015-11-23T18:44:37.050+0000 I INDEX [repl writer worker 3] build index on: db1.coll1 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db1.coll1" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.051+0000 d20012| 2015-11-23T18:44:37.050+0000 I INDEX [repl writer worker 3] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.053+0000 d20012| 2015-11-23T18:44:37.053+0000 I INDEX [repl writer worker 3] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.058+0000 d20011| 2015-11-23T18:44:37.057+0000 I INDEX [repl writer worker 9] build index on: db1.coll1 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db1.coll1" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.058+0000 d20011| 2015-11-23T18:44:37.057+0000 I INDEX [repl writer worker 9] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.061+0000 d20011| 2015-11-23T18:44:37.060+0000 I INDEX [repl writer worker 9] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.065+0000 s20019| 2015-11-23T18:44:37.064+0000 I SHARDING [conn1] distributed lock 'db1.coll1/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535e956e3a936d394e6ca1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.065+0000 s20019| 2015-11-23T18:44:37.065+0000 I SHARDING [conn1] enable sharding on: db1.coll1 with shard key: { _id: "hashed" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.066+0000 s20019| 2015-11-23T18:44:37.065+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.065+0000-56535e956e3a936d394e6ca2", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304277065), what: "shardCollection.start", ns: "db1.coll1", details: { shardKey: { _id: "hashed" }, collection: "db1.coll1", primary: "test-rs0:test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", initShards: [], numChunks: 2 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.067+0000 s20019| 2015-11-23T18:44:37.066+0000 I SHARDING [conn1] going to create 2 chunk(s) for: db1.coll1 using new epoch 56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.075+0000 s20019| 2015-11-23T18:44:37.075+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db1.coll1: 0ms sequenceNumber: 5 version: 1|1||56535e956e3a936d394e6ca3 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.080+0000 s20019| 2015-11-23T18:44:37.079+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db1.coll1: 0ms sequenceNumber: 6 version: 1|1||56535e956e3a936d394e6ca3 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.081+0000 d20010| 2015-11-23T18:44:37.080+0000 I SHARDING [conn30] remotely refreshing metadata for db1.coll1 with requested shard version 1|1||56535e956e3a936d394e6ca3, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.081+0000 d20010| 2015-11-23T18:44:37.081+0000 I SHARDING [conn30] collection db1.coll1 was previously unsharded, new metadata loaded with shard version 1|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.081+0000 d20010| 2015-11-23T18:44:37.081+0000 I SHARDING [conn30] collection version was loaded at version 1|1||56535e956e3a936d394e6ca3, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.082+0000 s20019| 2015-11-23T18:44:37.081+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.081+0000-56535e956e3a936d394e6ca4", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304277081), what: "shardCollection.end", ns: "db1.coll1", details: { version: "1|1||56535e956e3a936d394e6ca3" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.091+0000 s20019| 2015-11-23T18:44:37.091+0000 I SHARDING [conn1] distributed lock 'db1.coll1/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.091+0000 s20019| 2015-11-23T18:44:37.091+0000 I SHARDING [conn1] moving chunk ns: db1.coll1 moving ( ns: db1.coll1, shard: test-rs0, lastmod: 1|1||56535e956e3a936d394e6ca3, min: { _id: 0 }, max: { _id: MaxKey }) test-rs0 -> test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.092+0000 d20010| 2015-11-23T18:44:37.091+0000 I SHARDING [conn30] moveChunk waiting for full cleanup after move [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.093+0000 d20010| 2015-11-23T18:44:37.091+0000 I SHARDING [conn30] received moveChunk request: { moveChunk: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", to: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", fromShard: "test-rs0", toShard: "test-rs1", min: { _id: 0 }, max: { _id: MaxKey }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535e956e3a936d394e6ca3') ], epoch: ObjectId('56535e956e3a936d394e6ca3') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.108+0000 d20010| 2015-11-23T18:44:37.107+0000 I SHARDING [conn30] distributed lock 'db1.coll1/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'migrating chunk [{ _id: 0 }, { _id: MaxKey }) in db1.coll1', ts : 56535e95a5e5109e4dbf13d7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.110+0000 d20010| 2015-11-23T18:44:37.107+0000 I SHARDING [conn30] remotely refreshing metadata for db1.coll1 based on current shard version 1|1||56535e956e3a936d394e6ca3, current metadata version is 1|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.110+0000 d20010| 2015-11-23T18:44:37.108+0000 I SHARDING [conn30] metadata of collection db1.coll1 already up to date (shard version : 1|1||56535e956e3a936d394e6ca3, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.112+0000 d20010| 2015-11-23T18:44:37.108+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.108+0000-56535e95a5e5109e4dbf13d8", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304277108), what: "moveChunk.start", ns: "db1.coll1", details: { min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs0", to: "test-rs1" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.113+0000 d20010| 2015-11-23T18:44:37.109+0000 I SHARDING [conn30] moveChunk request accepted at version 1|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.113+0000 d20010| 2015-11-23T18:44:37.109+0000 I SHARDING [conn30] moveChunk number of documents: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.113+0000 d20015| 2015-11-23T18:44:37.110+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56718 #8 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.114+0000 d20013| 2015-11-23T18:44:37.111+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36449 #18 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.114+0000 d20013| 2015-11-23T18:44:37.112+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36450 #19 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.115+0000 d20013| 2015-11-23T18:44:37.113+0000 I SHARDING [conn19] remotely refreshing metadata for db1.coll1, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.116+0000 d20013| 2015-11-23T18:44:37.113+0000 I SHARDING [conn19] collection db1.coll1 was previously unsharded, new metadata loaded with shard version 0|0||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.116+0000 d20013| 2015-11-23T18:44:37.114+0000 I SHARDING [conn19] collection version was loaded at version 1|1||56535e956e3a936d394e6ca3, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.117+0000 d20013| 2015-11-23T18:44:37.114+0000 I SHARDING [migrateThread] starting receiving-end of migration of chunk { _id: 0 } -> { _id: MaxKey } for collection db1.coll1 from test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 at epoch 56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.117+0000 d20013| 2015-11-23T18:44:37.115+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36451 #20 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.117+0000 d20012| 2015-11-23T18:44:37.115+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:46070 #9 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.118+0000 d20010| 2015-11-23T18:44:37.116+0000 I SHARDING [conn30] moveChunk data transfer progress: { active: true, ns: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", min: { _id: 0 }, max: { _id: MaxKey }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.119+0000 d20010| 2015-11-23T18:44:37.118+0000 I SHARDING [conn30] moveChunk data transfer progress: { active: true, ns: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", min: { _id: 0 }, max: { _id: MaxKey }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.119+0000 d20010| 2015-11-23T18:44:37.118+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38871 #49 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.120+0000 d20010| 2015-11-23T18:44:37.120+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38872 #50 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.123+0000 d20010| 2015-11-23T18:44:37.122+0000 I SHARDING [conn30] moveChunk data transfer progress: { active: true, ns: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", min: { _id: 0 }, max: { _id: MaxKey }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.131+0000 d20010| 2015-11-23T18:44:37.131+0000 I SHARDING [conn30] moveChunk data transfer progress: { active: true, ns: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", min: { _id: 0 }, max: { _id: MaxKey }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.132+0000 d20013| 2015-11-23T18:44:37.131+0000 I INDEX [migrateThread] build index on: db1.coll1 properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "db1.coll1" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.132+0000 d20013| 2015-11-23T18:44:37.131+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.143+0000 d20013| 2015-11-23T18:44:37.143+0000 I INDEX [migrateThread] build index on: db1.coll1 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db1.coll1" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.143+0000 d20013| 2015-11-23T18:44:37.143+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.148+0000 d20010| 2015-11-23T18:44:37.147+0000 I SHARDING [conn30] moveChunk data transfer progress: { active: true, ns: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", min: { _id: 0 }, max: { _id: MaxKey }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.157+0000 d20013| 2015-11-23T18:44:37.156+0000 I INDEX [migrateThread] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.159+0000 d20013| 2015-11-23T18:44:37.156+0000 I SHARDING [migrateThread] Deleter starting delete for: db1.coll1 from { _id: 0 } -> { _id: MaxKey }, with opId: 364 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.160+0000 d20013| 2015-11-23T18:44:37.157+0000 I SHARDING [migrateThread] rangeDeleter deleted 0 documents for db1.coll1 from { _id: 0 } -> { _id: MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.165+0000 d20015| 2015-11-23T18:44:37.165+0000 I INDEX [repl writer worker 8] build index on: db1.coll1 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db1.coll1" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.165+0000 d20015| 2015-11-23T18:44:37.165+0000 I INDEX [repl writer worker 8] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.166+0000 d20014| 2015-11-23T18:44:37.165+0000 I INDEX [repl writer worker 8] build index on: db1.coll1 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db1.coll1" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.166+0000 d20014| 2015-11-23T18:44:37.165+0000 I INDEX [repl writer worker 8] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.167+0000 d20014| 2015-11-23T18:44:37.167+0000 I INDEX [repl writer worker 8] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.167+0000 d20015| 2015-11-23T18:44:37.167+0000 I INDEX [repl writer worker 8] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.170+0000 d20013| 2015-11-23T18:44:37.169+0000 I SHARDING [migrateThread] Waiting for replication to catch up before entering critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.170+0000 d20013| 2015-11-23T18:44:37.169+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db1.coll1' { _id: 0 } -> { _id: MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.180+0000 d20010| 2015-11-23T18:44:37.179+0000 I SHARDING [conn30] moveChunk data transfer progress: { active: true, ns: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", min: { _id: 0 }, max: { _id: MaxKey }, shardKeyPattern: { _id: "hashed" }, state: "steady", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.180+0000 d20010| 2015-11-23T18:44:37.179+0000 I SHARDING [conn30] About to check if it is safe to enter critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.180+0000 d20010| 2015-11-23T18:44:37.179+0000 I SHARDING [conn30] About to enter migrate critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.181+0000 d20010| 2015-11-23T18:44:37.179+0000 I SHARDING [conn30] moveChunk setting version to: 2|0||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.181+0000 d20013| 2015-11-23T18:44:37.180+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36455 #21 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.191+0000 d20013| 2015-11-23T18:44:37.190+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db1.coll1' { _id: 0 } -> { _id: MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.192+0000 d20013| 2015-11-23T18:44:37.192+0000 I NETWORK [migrateThread] scoped connection to ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.192+0000 c20016| 2015-11-23T18:44:37.192+0000 I NETWORK [conn19] end connection 10.99.163.247:48341 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.193+0000 d20013| 2015-11-23T18:44:37.192+0000 I SHARDING [migrateThread] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.192+0000-56535e956202d0bae25462fb", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304277192), what: "moveChunk.to", ns: "db1.coll1", details: { min: { _id: 0 }, max: { _id: MaxKey }, step 1 of 5: 42, step 2 of 5: 12, step 3 of 5: 0, step 4 of 5: 0, step 5 of 5: 20, note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.194+0000 c20017| 2015-11-23T18:44:37.192+0000 I NETWORK [conn19] end connection 10.99.163.247:55656 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.194+0000 c20018| 2015-11-23T18:44:37.192+0000 I NETWORK [conn19] end connection 10.99.163.247:36777 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.194+0000 c20016| 2015-11-23T18:44:37.193+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48442 #35 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.195+0000 c20017| 2015-11-23T18:44:37.193+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55757 #33 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.195+0000 c20018| 2015-11-23T18:44:37.194+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36878 #33 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.196+0000 d20010| 2015-11-23T18:44:37.195+0000 I SHARDING [conn30] moveChunk migrate commit accepted by TO-shard: { active: false, ns: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", min: { _id: 0 }, max: { _id: MaxKey }, shardKeyPattern: { _id: "hashed" }, state: "done", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.197+0000 d20010| 2015-11-23T18:44:37.195+0000 I SHARDING [conn30] moveChunk updating self version to: 2|1||56535e956e3a936d394e6ca3 through { _id: MinKey } -> { _id: 0 } for collection 'db1.coll1' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.209+0000 d20010| 2015-11-23T18:44:37.207+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.207+0000-56535e95a5e5109e4dbf13d9", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304277207), what: "moveChunk.commit", ns: "db1.coll1", details: { min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs0", to: "test-rs1", cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.209+0000 d20010| 2015-11-23T18:44:37.208+0000 I SHARDING [conn30] MigrateFromStatus::done About to acquire global lock to exit critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.210+0000 d20010| 2015-11-23T18:44:37.208+0000 I SHARDING [conn30] doing delete inline for cleanup of chunk data [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.210+0000 d20010| 2015-11-23T18:44:37.208+0000 I SHARDING [conn30] Deleter starting delete for: db1.coll1 from { _id: 0 } -> { _id: MaxKey }, with opId: 2131 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.210+0000 d20010| 2015-11-23T18:44:37.208+0000 I SHARDING [conn30] rangeDeleter deleted 0 documents for db1.coll1 from { _id: 0 } -> { _id: MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.211+0000 d20010| 2015-11-23T18:44:37.208+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.208+0000-56535e95a5e5109e4dbf13da", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304277208), what: "moveChunk.from", ns: "db1.coll1", details: { min: { _id: 0 }, max: { _id: MaxKey }, step 1 of 6: 0, step 2 of 6: 17, step 3 of 6: 5, step 4 of 6: 65, step 5 of 6: 29, step 6 of 6: 0, to: "test-rs1", from: "test-rs0", note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.223+0000 d20010| 2015-11-23T18:44:37.222+0000 I SHARDING [conn30] distributed lock 'db1.coll1/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.225+0000 d20010| 2015-11-23T18:44:37.222+0000 I COMMAND [conn30] command db1.coll1 command: moveChunk { moveChunk: "db1.coll1", from: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", to: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", fromShard: "test-rs0", toShard: "test-rs1", min: { _id: 0 }, max: { _id: MaxKey }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535e956e3a936d394e6ca3') ], epoch: ObjectId('56535e956e3a936d394e6ca3') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:134 locks:{ Global: { acquireCount: { r: 11, w: 3, R: 2 } }, Database: { acquireCount: { r: 3, w: 3 } }, Collection: { acquireCount: { r: 3, w: 1, W: 2 } } } protocol:op_command 131ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.226+0000 s20019| 2015-11-23T18:44:37.223+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db1.coll1: 0ms sequenceNumber: 7 version: 2|1||56535e956e3a936d394e6ca3 based on: 1|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.227+0000 d20010| 2015-11-23T18:44:37.223+0000 I SHARDING [conn30] received splitChunk request: { splitChunk: "db1.coll1", keyPattern: { _id: "hashed" }, min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs0", splitKeys: [ { _id: -4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535e956e3a936d394e6ca3') ], epoch: ObjectId('56535e956e3a936d394e6ca3') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.234+0000 d20010| 2015-11-23T18:44:37.232+0000 I SHARDING [conn30] distributed lock 'db1.coll1/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ _id: MinKey }, { _id: 0 }) in db1.coll1', ts : 56535e95a5e5109e4dbf13db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.234+0000 d20010| 2015-11-23T18:44:37.232+0000 I SHARDING [conn30] remotely refreshing metadata for db1.coll1 based on current shard version 2|0||56535e956e3a936d394e6ca3, current metadata version is 2|0||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.235+0000 d20010| 2015-11-23T18:44:37.232+0000 I SHARDING [conn30] updating metadata for db1.coll1 from shard version 2|0||56535e956e3a936d394e6ca3 to shard version 2|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.235+0000 d20010| 2015-11-23T18:44:37.232+0000 I SHARDING [conn30] collection version was loaded at version 2|1||56535e956e3a936d394e6ca3, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.235+0000 d20010| 2015-11-23T18:44:37.232+0000 I SHARDING [conn30] splitChunk accepted at version 2|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.241+0000 d20010| 2015-11-23T18:44:37.239+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.239+0000-56535e95a5e5109e4dbf13dc", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304277239), what: "split", ns: "db1.coll1", details: { before: { min: { _id: MinKey }, max: { _id: 0 } }, left: { min: { _id: MinKey }, max: { _id: -4611686018427387902 }, lastmod: Timestamp 2000|2, lastmodEpoch: ObjectId('56535e956e3a936d394e6ca3') }, right: { min: { _id: -4611686018427387902 }, max: { _id: 0 }, lastmod: Timestamp 2000|3, lastmodEpoch: ObjectId('56535e956e3a936d394e6ca3') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.248+0000 d20010| 2015-11-23T18:44:37.247+0000 I SHARDING [conn30] distributed lock 'db1.coll1/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.249+0000 s20019| 2015-11-23T18:44:37.248+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db1.coll1: 0ms sequenceNumber: 8 version: 2|3||56535e956e3a936d394e6ca3 based on: 2|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.249+0000 d20013| 2015-11-23T18:44:37.249+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36459 #22 (16 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.250+0000 d20013| 2015-11-23T18:44:37.249+0000 I SHARDING [conn22] received splitChunk request: { splitChunk: "db1.coll1", keyPattern: { _id: "hashed" }, min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs1", splitKeys: [ { _id: 4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535e956e3a936d394e6ca3') ], epoch: ObjectId('56535e956e3a936d394e6ca3') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.250+0000 c20016| 2015-11-23T18:44:37.250+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48446 #36 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.252+0000 c20017| 2015-11-23T18:44:37.252+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55761 #34 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.254+0000 c20018| 2015-11-23T18:44:37.253+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36882 #34 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.259+0000 d20013| 2015-11-23T18:44:37.258+0000 I SHARDING [LockPinger] creating distributed lock ping thread for ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 and process ip-10-99-163-247:20013:1448304277:68810772 (sleeping for 30000ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.259+0000 d20013| 2015-11-23T18:44:37.258+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.259+0000 c20016| 2015-11-23T18:44:37.259+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48449 #37 (34 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.260+0000 d20013| 2015-11-23T18:44:37.258+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.260+0000 d20013| 2015-11-23T18:44:37.259+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.260+0000 c20017| 2015-11-23T18:44:37.260+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55764 #35 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.261+0000 d20013| 2015-11-23T18:44:37.260+0000 I NETWORK [LockPinger] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.261+0000 c20016| 2015-11-23T18:44:37.260+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48451 #38 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.262+0000 d20013| 2015-11-23T18:44:37.261+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.262+0000 c20017| 2015-11-23T18:44:37.261+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55766 #36 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.262+0000 d20013| 2015-11-23T18:44:37.261+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.263+0000 c20018| 2015-11-23T18:44:37.261+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36887 #35 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.263+0000 c20018| 2015-11-23T18:44:37.263+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36888 #36 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.271+0000 d20013| 2015-11-23T18:44:37.270+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:44:37.262+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20013:1448304277:68810772', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.274+0000 d20013| 2015-11-23T18:44:37.273+0000 I SHARDING [conn22] distributed lock 'db1.coll1/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ _id: 0 }, { _id: MaxKey }) in db1.coll1', ts : 56535e956202d0bae25462fc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.274+0000 d20013| 2015-11-23T18:44:37.273+0000 I SHARDING [conn22] remotely refreshing metadata for db1.coll1 based on current shard version 0|0||56535e956e3a936d394e6ca3, current metadata version is 1|1||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.275+0000 d20013| 2015-11-23T18:44:37.274+0000 I SHARDING [conn22] updating metadata for db1.coll1 from shard version 0|0||56535e956e3a936d394e6ca3 to shard version 2|0||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.275+0000 d20013| 2015-11-23T18:44:37.274+0000 I SHARDING [conn22] collection version was loaded at version 2|3||56535e956e3a936d394e6ca3, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.276+0000 d20013| 2015-11-23T18:44:37.274+0000 I SHARDING [conn22] splitChunk accepted at version 2|0||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.281+0000 d20013| 2015-11-23T18:44:37.280+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:37.280+0000-56535e956202d0bae25462fd", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304277280), what: "split", ns: "db1.coll1", details: { before: { min: { _id: 0 }, max: { _id: MaxKey } }, left: { min: { _id: 0 }, max: { _id: 4611686018427387902 }, lastmod: Timestamp 2000|4, lastmodEpoch: ObjectId('56535e956e3a936d394e6ca3') }, right: { min: { _id: 4611686018427387902 }, max: { _id: MaxKey }, lastmod: Timestamp 2000|5, lastmodEpoch: ObjectId('56535e956e3a936d394e6ca3') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.281+0000 d20013| 2015-11-23T18:44:37.281+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.282+0000 c20016| 2015-11-23T18:44:37.281+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48455 #39 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.282+0000 d20013| 2015-11-23T18:44:37.281+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.283+0000 c20017| 2015-11-23T18:44:37.282+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55770 #37 (34 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.283+0000 d20013| 2015-11-23T18:44:37.282+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.283+0000 c20018| 2015-11-23T18:44:37.283+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36891 #37 (34 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.291+0000 d20013| 2015-11-23T18:44:37.291+0000 I SHARDING [conn22] distributed lock 'db1.coll1/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.292+0000 s20019| 2015-11-23T18:44:37.291+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db1.coll1: 0ms sequenceNumber: 9 version: 2|5||56535e956e3a936d394e6ca3 based on: 2|3||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.292+0000 Using 5 threads (requested 5) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.371+0000 s20020| 2015-11-23T18:44:37.370+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60356 #7 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.381+0000 s20020| 2015-11-23T18:44:37.381+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60357 #8 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.387+0000 s20019| 2015-11-23T18:44:37.387+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39946 #7 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.395+0000 s20019| 2015-11-23T18:44:37.394+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39947 #8 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.401+0000 s20019| 2015-11-23T18:44:37.401+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39948 #9 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.404+0000 setting random seed: 3063817189 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.405+0000 setting random seed: 1400558261 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.405+0000 setting random seed: 2098848865 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.406+0000 setting random seed: 1362322558 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.406+0000 setting random seed: 1432003928 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:37.407+0000 s20020| 2015-11-23T18:44:37.406+0000 I SHARDING [conn8] ChunkManager: time to load chunks for db1.coll1: 0ms sequenceNumber: 3 version: 2|5||56535e956e3a936d394e6ca3 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.248+0000 d20010| 2015-11-23T18:44:40.248+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38895 #51 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.257+0000 s20020| 2015-11-23T18:44:40.256+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-1-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.266+0000 d20010| 2015-11-23T18:44:40.266+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38896 #52 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.269+0000 s20020| 2015-11-23T18:44:40.269+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.363+0000 d20010| 2015-11-23T18:44:40.363+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38897 #53 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.366+0000 s20020| 2015-11-23T18:44:40.365+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-3-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.371+0000 d20010| 2015-11-23T18:44:40.370+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38898 #54 (48 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.374+0000 s20020| 2015-11-23T18:44:40.373+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.484+0000 d20010| 2015-11-23T18:44:40.483+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38899 #55 (49 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.491+0000 s20019| 2015-11-23T18:44:40.490+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.494+0000 d20010| 2015-11-23T18:44:40.494+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38900 #56 (50 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.500+0000 s20019| 2015-11-23T18:44:40.499+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-1-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.612+0000 d20010| 2015-11-23T18:44:40.611+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38901 #57 (51 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.613+0000 s20019| 2015-11-23T18:44:40.612+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.619+0000 d20010| 2015-11-23T18:44:40.618+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38902 #58 (52 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:40.622+0000 s20019| 2015-11-23T18:44:40.621+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-3-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:43.735+0000 d20011| 2015-11-23T18:44:43.734+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49862 #9 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.049+0000 s20019| 2015-11-23T18:44:45.049+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'doing balance round', ts : 56535e9d6e3a936d394e6ca5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.057+0000 s20019| 2015-11-23T18:44:45.056+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:44:45.056+0000-56535e9d6e3a936d394e6ca6", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304285056), what: "balancer.round", ns: "", details: { executionTimeMillis: 47, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.110+0000 d20015| 2015-11-23T18:44:45.110+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56757 #9 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.122+0000 d20013| 2015-11-23T18:44:45.122+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36488 #23 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.125+0000 s20019| 2015-11-23T18:44:45.124+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.135+0000 d20013| 2015-11-23T18:44:45.135+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36489 #24 (18 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.139+0000 s20020| 2015-11-23T18:44:45.138+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.200+0000 d20014| 2015-11-23T18:44:45.200+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37428 #9 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.212+0000 d20011| 2015-11-23T18:44:45.211+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49867 #10 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.224+0000 s20020| 2015-11-23T18:44:45.223+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' acquired for 'doing balance round', ts : 56535e9d13c05a55d8fa6f90 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.231+0000 s20020| 2015-11-23T18:44:45.230+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:44:45.230+0000-56535e9d13c05a55d8fa6f91", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304285230), what: "balancer.round", ns: "", details: { executionTimeMillis: 142, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.262+0000 d20014| 2015-11-23T18:44:45.261+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37430 #10 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.263+0000 s20020| 2015-11-23T18:44:45.262+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.267+0000 d20015| 2015-11-23T18:44:45.267+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56763 #10 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.275+0000 d20012| 2015-11-23T18:44:45.275+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:46112 #10 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.284+0000 d20013| 2015-11-23T18:44:45.283+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36495 #25 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.284+0000 d20010| 2015-11-23T18:44:45.284+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38914 #59 (53 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.287+0000 d20011| 2015-11-23T18:44:45.286+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49873 #11 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.288+0000 d20014| 2015-11-23T18:44:45.288+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37436 #11 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.293+0000 d20011| 2015-11-23T18:44:45.293+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:49875 #12 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:45.726+0000 d20014| 2015-11-23T18:44:45.726+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37438 #12 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:48.781+0000 s20020| 2015-11-23T18:44:48.781+0000 I NETWORK [conn7] end connection 10.99.163.247:60356 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:48.788+0000 s20020| 2015-11-23T18:44:48.788+0000 I NETWORK [conn8] end connection 10.99.163.247:60357 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.284+0000 d20010| 2015-11-23T18:44:49.283+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38919 #60 (54 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.284+0000 d20012| 2015-11-23T18:44:49.284+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.392+0000 s20019| 2015-11-23T18:44:49.391+0000 I NETWORK [conn7] end connection 10.99.163.247:39946 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.485+0000 s20019| 2015-11-23T18:44:49.485+0000 I NETWORK [conn9] end connection 10.99.163.247:39948 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.553+0000 s20019| 2015-11-23T18:44:49.552+0000 I NETWORK [conn8] end connection 10.99.163.247:39947 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.558+0000 s20019| 2015-11-23T18:44:49.557+0000 I COMMAND [conn1] DROP: db1.create_capped_collection0_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.558+0000 s20019| 2015-11-23T18:44:49.557+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.558+0000 d20010| 2015-11-23T18:44:49.558+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection0_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.560+0000 s20019| 2015-11-23T18:44:49.559+0000 I COMMAND [conn1] DROP: db1.create_capped_collection0_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.560+0000 s20019| 2015-11-23T18:44:49.559+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.560+0000 d20010| 2015-11-23T18:44:49.559+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection0_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.560+0000 d20012| 2015-11-23T18:44:49.560+0000 I COMMAND [repl writer worker 4] CMD: drop db1.create_capped_collection0_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.561+0000 s20019| 2015-11-23T18:44:49.561+0000 I COMMAND [conn1] DROP: db1.create_capped_collection0_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.561+0000 s20019| 2015-11-23T18:44:49.561+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.562+0000 d20011| 2015-11-23T18:44:49.560+0000 I COMMAND [repl writer worker 13] CMD: drop db1.create_capped_collection0_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.562+0000 d20010| 2015-11-23T18:44:49.561+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection0_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.564+0000 s20019| 2015-11-23T18:44:49.564+0000 I COMMAND [conn1] DROP: db1.create_capped_collection0_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.564+0000 s20019| 2015-11-23T18:44:49.564+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.565+0000 d20010| 2015-11-23T18:44:49.564+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection0_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.565+0000 d20012| 2015-11-23T18:44:49.564+0000 I COMMAND [repl writer worker 1] CMD: drop db1.create_capped_collection0_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.565+0000 d20011| 2015-11-23T18:44:49.564+0000 I COMMAND [repl writer worker 1] CMD: drop db1.create_capped_collection0_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.570+0000 s20019| 2015-11-23T18:44:49.570+0000 I COMMAND [conn1] DROP: db1.create_capped_collection1_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.570+0000 s20019| 2015-11-23T18:44:49.570+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.570+0000 d20012| 2015-11-23T18:44:49.570+0000 I COMMAND [repl writer worker 10] CMD: drop db1.create_capped_collection0_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.571+0000 d20010| 2015-11-23T18:44:49.570+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection1_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.572+0000 d20011| 2015-11-23T18:44:49.571+0000 I COMMAND [repl writer worker 3] CMD: drop db1.create_capped_collection0_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.574+0000 d20010| 2015-11-23T18:44:49.573+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38920 #61 (55 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.574+0000 d20011| 2015-11-23T18:44:49.574+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.576+0000 d20012| 2015-11-23T18:44:49.576+0000 I COMMAND [repl writer worker 7] CMD: drop db1.create_capped_collection0_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.577+0000 d20011| 2015-11-23T18:44:49.577+0000 I COMMAND [repl writer worker 15] CMD: drop db1.create_capped_collection0_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.579+0000 s20019| 2015-11-23T18:44:49.579+0000 I COMMAND [conn1] DROP: db1.create_capped_collection1_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.579+0000 s20019| 2015-11-23T18:44:49.579+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.580+0000 d20010| 2015-11-23T18:44:49.579+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection1_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.582+0000 d20012| 2015-11-23T18:44:49.581+0000 I COMMAND [repl writer worker 3] CMD: drop db1.create_capped_collection1_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.585+0000 s20019| 2015-11-23T18:44:49.585+0000 I COMMAND [conn1] DROP: db1.create_capped_collection1_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.585+0000 s20019| 2015-11-23T18:44:49.585+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.586+0000 d20010| 2015-11-23T18:44:49.585+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection1_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.586+0000 d20011| 2015-11-23T18:44:49.585+0000 I COMMAND [repl writer worker 8] CMD: drop db1.create_capped_collection1_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.591+0000 d20012| 2015-11-23T18:44:49.591+0000 I COMMAND [repl writer worker 15] CMD: drop db1.create_capped_collection1_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.597+0000 s20019| 2015-11-23T18:44:49.596+0000 I COMMAND [conn1] DROP: db1.create_capped_collection1_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.597+0000 s20019| 2015-11-23T18:44:49.597+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.598+0000 d20010| 2015-11-23T18:44:49.597+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection1_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.598+0000 d20011| 2015-11-23T18:44:49.597+0000 I COMMAND [repl writer worker 14] CMD: drop db1.create_capped_collection1_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.608+0000 d20012| 2015-11-23T18:44:49.608+0000 I COMMAND [repl writer worker 0] CMD: drop db1.create_capped_collection1_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.617+0000 s20019| 2015-11-23T18:44:49.617+0000 I COMMAND [conn1] DROP: db1.create_capped_collection2_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.617+0000 s20019| 2015-11-23T18:44:49.617+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.617+0000 d20010| 2015-11-23T18:44:49.617+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection2_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.618+0000 d20011| 2015-11-23T18:44:49.617+0000 I COMMAND [repl writer worker 0] CMD: drop db1.create_capped_collection1_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.628+0000 s20019| 2015-11-23T18:44:49.627+0000 I COMMAND [conn1] DROP: db1.create_capped_collection2_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.628+0000 s20019| 2015-11-23T18:44:49.627+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.628+0000 d20012| 2015-11-23T18:44:49.627+0000 I COMMAND [repl writer worker 2] CMD: drop db1.create_capped_collection1_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.628+0000 d20010| 2015-11-23T18:44:49.627+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection2_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.630+0000 d20011| 2015-11-23T18:44:49.630+0000 I COMMAND [repl writer worker 4] CMD: drop db1.create_capped_collection1_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.645+0000 d20012| 2015-11-23T18:44:49.644+0000 I COMMAND [repl writer worker 9] CMD: drop db1.create_capped_collection2_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.650+0000 s20019| 2015-11-23T18:44:49.650+0000 I COMMAND [conn1] DROP: db1.create_capped_collection2_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.651+0000 s20019| 2015-11-23T18:44:49.650+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.651+0000 d20010| 2015-11-23T18:44:49.650+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection2_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.651+0000 d20011| 2015-11-23T18:44:49.650+0000 I COMMAND [repl writer worker 2] CMD: drop db1.create_capped_collection2_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.669+0000 d20012| 2015-11-23T18:44:49.669+0000 I COMMAND [repl writer worker 13] CMD: drop db1.create_capped_collection2_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.675+0000 s20019| 2015-11-23T18:44:49.675+0000 I COMMAND [conn1] DROP: db1.create_capped_collection2_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.675+0000 s20019| 2015-11-23T18:44:49.675+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.676+0000 d20010| 2015-11-23T18:44:49.675+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection2_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.676+0000 d20011| 2015-11-23T18:44:49.675+0000 I COMMAND [repl writer worker 9] CMD: drop db1.create_capped_collection2_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.679+0000 d20012| 2015-11-23T18:44:49.679+0000 I COMMAND [repl writer worker 6] CMD: drop db1.create_capped_collection2_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.688+0000 d20011| 2015-11-23T18:44:49.687+0000 I COMMAND [repl writer worker 12] CMD: drop db1.create_capped_collection2_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.692+0000 s20019| 2015-11-23T18:44:49.691+0000 I COMMAND [conn1] DROP: db1.create_capped_collection3_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.692+0000 s20019| 2015-11-23T18:44:49.691+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.692+0000 d20010| 2015-11-23T18:44:49.692+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection3_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.696+0000 d20012| 2015-11-23T18:44:49.696+0000 I COMMAND [repl writer worker 8] CMD: drop db1.create_capped_collection2_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.711+0000 s20019| 2015-11-23T18:44:49.710+0000 I COMMAND [conn1] DROP: db1.create_capped_collection3_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.711+0000 s20019| 2015-11-23T18:44:49.711+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.711+0000 d20010| 2015-11-23T18:44:49.711+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection3_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.712+0000 d20011| 2015-11-23T18:44:49.711+0000 I COMMAND [repl writer worker 5] CMD: drop db1.create_capped_collection2_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.713+0000 d20012| 2015-11-23T18:44:49.713+0000 I COMMAND [repl writer worker 12] CMD: drop db1.create_capped_collection3_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.718+0000 s20019| 2015-11-23T18:44:49.718+0000 I COMMAND [conn1] DROP: db1.create_capped_collection3_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.718+0000 s20019| 2015-11-23T18:44:49.718+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.718+0000 d20010| 2015-11-23T18:44:49.718+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection3_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.722+0000 d20011| 2015-11-23T18:44:49.721+0000 I COMMAND [repl writer worker 10] CMD: drop db1.create_capped_collection3_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.726+0000 d20012| 2015-11-23T18:44:49.725+0000 I COMMAND [repl writer worker 5] CMD: drop db1.create_capped_collection3_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.726+0000 s20019| 2015-11-23T18:44:49.726+0000 I COMMAND [conn1] DROP: db1.create_capped_collection3_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.726+0000 s20019| 2015-11-23T18:44:49.726+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.727+0000 d20010| 2015-11-23T18:44:49.726+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection3_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.732+0000 d20011| 2015-11-23T18:44:49.732+0000 I COMMAND [repl writer worker 7] CMD: drop db1.create_capped_collection3_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.736+0000 s20019| 2015-11-23T18:44:49.736+0000 I COMMAND [conn1] DROP: db1.create_capped_collection4_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.737+0000 s20019| 2015-11-23T18:44:49.736+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.737+0000 d20010| 2015-11-23T18:44:49.736+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection4_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.738+0000 d20012| 2015-11-23T18:44:49.737+0000 I COMMAND [repl writer worker 11] CMD: drop db1.create_capped_collection3_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.740+0000 s20019| 2015-11-23T18:44:49.739+0000 I COMMAND [conn1] DROP: db1.create_capped_collection4_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.740+0000 s20019| 2015-11-23T18:44:49.739+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.741+0000 d20011| 2015-11-23T18:44:49.740+0000 I COMMAND [repl writer worker 11] CMD: drop db1.create_capped_collection3_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.741+0000 d20010| 2015-11-23T18:44:49.740+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection4_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.744+0000 d20012| 2015-11-23T18:44:49.743+0000 I COMMAND [repl writer worker 14] CMD: drop db1.create_capped_collection3_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.744+0000 s20019| 2015-11-23T18:44:49.744+0000 I COMMAND [conn1] DROP: db1.create_capped_collection4_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.745+0000 s20019| 2015-11-23T18:44:49.744+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.745+0000 d20010| 2015-11-23T18:44:49.744+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection4_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.747+0000 d20011| 2015-11-23T18:44:49.746+0000 I COMMAND [repl writer worker 6] CMD: drop db1.create_capped_collection3_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.754+0000 s20019| 2015-11-23T18:44:49.754+0000 I COMMAND [conn1] DROP: db1.create_capped_collection4_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.755+0000 s20019| 2015-11-23T18:44:49.754+0000 I COMMAND [conn1] drop going to do passthrough [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.755+0000 d20010| 2015-11-23T18:44:49.754+0000 I COMMAND [conn26] CMD: drop db1.create_capped_collection4_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.756+0000 d20012| 2015-11-23T18:44:49.755+0000 I COMMAND [repl writer worker 4] CMD: drop db1.create_capped_collection4_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.757+0000 d20011| 2015-11-23T18:44:49.757+0000 I COMMAND [repl writer worker 13] CMD: drop db1.create_capped_collection4_0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.759+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.759+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.759+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.760+0000 Workload(s) completed in 12467 ms: jstests/concurrency/fsm_workloads/create_capped_collection.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.760+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.761+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.761+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.761+0000 s20019| 2015-11-23T18:44:49.759+0000 I COMMAND [conn1] DROP: db1.coll1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.762+0000 s20019| 2015-11-23T18:44:49.759+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:49.759+0000-56535ea16e3a936d394e6ca7", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304289759), what: "dropCollection.start", ns: "db1.coll1", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.762+0000 d20012| 2015-11-23T18:44:49.761+0000 I COMMAND [repl writer worker 1] CMD: drop db1.create_capped_collection4_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.763+0000 d20011| 2015-11-23T18:44:49.763+0000 I COMMAND [repl writer worker 1] CMD: drop db1.create_capped_collection4_1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.769+0000 d20012| 2015-11-23T18:44:49.768+0000 I COMMAND [repl writer worker 10] CMD: drop db1.create_capped_collection4_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.772+0000 d20011| 2015-11-23T18:44:49.772+0000 I COMMAND [repl writer worker 3] CMD: drop db1.create_capped_collection4_2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.773+0000 d20012| 2015-11-23T18:44:49.773+0000 I COMMAND [repl writer worker 7] CMD: drop db1.create_capped_collection4_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.777+0000 d20011| 2015-11-23T18:44:49.777+0000 I COMMAND [repl writer worker 15] CMD: drop db1.create_capped_collection4_3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.785+0000 s20019| 2015-11-23T18:44:49.784+0000 I SHARDING [conn1] distributed lock 'db1.coll1/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535ea16e3a936d394e6ca8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.785+0000 d20010| 2015-11-23T18:44:49.784+0000 I COMMAND [conn46] CMD: drop db1.coll1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.787+0000 d20011| 2015-11-23T18:44:49.787+0000 I COMMAND [repl writer worker 8] CMD: drop db1.coll1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.787+0000 d20013| 2015-11-23T18:44:49.786+0000 I COMMAND [conn11] CMD: drop db1.coll1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.788+0000 d20012| 2015-11-23T18:44:49.787+0000 I COMMAND [repl writer worker 3] CMD: drop db1.coll1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.789+0000 d20014| 2015-11-23T18:44:49.788+0000 I COMMAND [repl writer worker 9] CMD: drop db1.coll1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.790+0000 d20015| 2015-11-23T18:44:49.788+0000 I COMMAND [repl writer worker 9] CMD: drop db1.coll1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.808+0000 d20010| 2015-11-23T18:44:49.807+0000 I SHARDING [conn46] remotely refreshing metadata for db1.coll1 with requested shard version 0|0||000000000000000000000000, current shard version is 2|3||56535e956e3a936d394e6ca3, current metadata version is 2|3||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.808+0000 d20010| 2015-11-23T18:44:49.808+0000 I SHARDING [conn46] dropping metadata for db1.coll1 at shard version 2|3||56535e956e3a936d394e6ca3, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.809+0000 d20013| 2015-11-23T18:44:49.808+0000 I SHARDING [conn11] remotely refreshing metadata for db1.coll1 with requested shard version 0|0||000000000000000000000000, current shard version is 2|5||56535e956e3a936d394e6ca3, current metadata version is 2|5||56535e956e3a936d394e6ca3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.809+0000 d20013| 2015-11-23T18:44:49.808+0000 I SHARDING [conn11] dropping metadata for db1.coll1 at shard version 2|5||56535e956e3a936d394e6ca3, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.810+0000 s20019| 2015-11-23T18:44:49.809+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:49.809+0000-56535ea16e3a936d394e6ca9", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304289809), what: "dropCollection", ns: "db1.coll1", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.836+0000 s20019| 2015-11-23T18:44:49.835+0000 I SHARDING [conn1] distributed lock 'db1.coll1/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.836+0000 s20019| 2015-11-23T18:44:49.836+0000 I COMMAND [conn1] DROP DATABASE: db1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.837+0000 s20019| 2015-11-23T18:44:49.836+0000 I SHARDING [conn1] DBConfig::dropDatabase: db1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.837+0000 s20019| 2015-11-23T18:44:49.836+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:49.836+0000-56535ea16e3a936d394e6caa", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304289836), what: "dropDatabase.start", ns: "db1", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.845+0000 s20019| 2015-11-23T18:44:49.845+0000 I SHARDING [conn1] DBConfig::dropDatabase: db1 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.846+0000 d20010| 2015-11-23T18:44:49.845+0000 I COMMAND [conn48] dropDatabase db1 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.846+0000 d20010| 2015-11-23T18:44:49.845+0000 I COMMAND [conn48] dropDatabase db1 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.846+0000 d20012| 2015-11-23T18:44:49.845+0000 I COMMAND [repl writer worker 15] dropDatabase db1 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.847+0000 d20012| 2015-11-23T18:44:49.845+0000 I COMMAND [repl writer worker 15] dropDatabase db1 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.847+0000 s20019| 2015-11-23T18:44:49.845+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:49.845+0000-56535ea16e3a936d394e6cab", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304289845), what: "dropDatabase", ns: "db1", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.848+0000 d20011| 2015-11-23T18:44:49.845+0000 I COMMAND [repl writer worker 14] dropDatabase db1 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.848+0000 d20011| 2015-11-23T18:44:49.846+0000 I COMMAND [repl writer worker 14] dropDatabase db1 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.855+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.855+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.855+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.855+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.856+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.856+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.856+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.888+0000 d20010| 2015-11-23T18:44:49.888+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.888+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.888+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.888+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.889+0000 ip-10-99-163-247:20010: awaitReplication completed in 34 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.889+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.889+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.889+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.889+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.889+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.889+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.890+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.890+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.890+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.890+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.890+0000 d20012| 2015-11-23T18:44:49.888+0000 I COMMAND [repl writer worker 9] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.890+0000 d20011| 2015-11-23T18:44:49.889+0000 I COMMAND [repl writer worker 2] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.921+0000 d20013| 2015-11-23T18:44:49.921+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.921+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.921+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.922+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.923+0000 ip-10-99-163-247:20013: awaitReplication completed in 33 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.923+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.923+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.923+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.923+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.923+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.924+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.924+0000 Workload(s) started: jstests/concurrency/fsm_workloads/yield_sort_merge.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.924+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.924+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.924+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.925+0000 d20015| 2015-11-23T18:44:49.922+0000 I COMMAND [repl writer worker 11] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.925+0000 d20014| 2015-11-23T18:44:49.922+0000 I COMMAND [repl writer worker 12] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.958+0000 s20019| 2015-11-23T18:44:49.957+0000 I SHARDING [conn1] distributed lock 'db2/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535ea16e3a936d394e6cac [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.959+0000 s20019| 2015-11-23T18:44:49.959+0000 I SHARDING [conn1] Placing [db2] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.959+0000 s20019| 2015-11-23T18:44:49.959+0000 I SHARDING [conn1] Enabling sharding for database [db2] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.965+0000 s20019| 2015-11-23T18:44:49.965+0000 I SHARDING [conn1] distributed lock 'db2/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.967+0000 d20013| 2015-11-23T18:44:49.966+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36503 #26 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.968+0000 d20013| 2015-11-23T18:44:49.968+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36504 #27 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.979+0000 d20013| 2015-11-23T18:44:49.978+0000 I INDEX [conn27] build index on: db2.coll2 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.979+0000 d20013| 2015-11-23T18:44:49.978+0000 I INDEX [conn27] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.980+0000 d20013| 2015-11-23T18:44:49.979+0000 I INDEX [conn27] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.980+0000 s20019| 2015-11-23T18:44:49.980+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db2.coll2", key: { _id: "hashed" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.985+0000 d20014| 2015-11-23T18:44:49.985+0000 I INDEX [repl writer worker 14] build index on: db2.coll2 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.986+0000 d20014| 2015-11-23T18:44:49.985+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.987+0000 d20015| 2015-11-23T18:44:49.987+0000 I INDEX [repl writer worker 13] build index on: db2.coll2 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.988+0000 d20014| 2015-11-23T18:44:49.987+0000 I INDEX [repl writer worker 14] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.988+0000 d20015| 2015-11-23T18:44:49.987+0000 I INDEX [repl writer worker 13] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:49.989+0000 d20015| 2015-11-23T18:44:49.989+0000 I INDEX [repl writer worker 13] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.003+0000 s20019| 2015-11-23T18:44:49.999+0000 I SHARDING [conn1] distributed lock 'db2.coll2/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535ea16e3a936d394e6cad [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.003+0000 s20019| 2015-11-23T18:44:49.999+0000 I SHARDING [conn1] enable sharding on: db2.coll2 with shard key: { _id: "hashed" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.005+0000 s20019| 2015-11-23T18:44:49.999+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:49.999+0000-56535ea16e3a936d394e6cae", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304289999), what: "shardCollection.start", ns: "db2.coll2", details: { shardKey: { _id: "hashed" }, collection: "db2.coll2", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 2 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.006+0000 s20019| 2015-11-23T18:44:50.005+0000 I SHARDING [conn1] going to create 2 chunk(s) for: db2.coll2 using new epoch 56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.012+0000 s20019| 2015-11-23T18:44:50.011+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db2.coll2: 0ms sequenceNumber: 10 version: 1|1||56535ea26e3a936d394e6caf based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.017+0000 s20019| 2015-11-23T18:44:50.016+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db2.coll2: 0ms sequenceNumber: 11 version: 1|1||56535ea26e3a936d394e6caf based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.018+0000 d20013| 2015-11-23T18:44:50.017+0000 I SHARDING [conn22] remotely refreshing metadata for db2.coll2 with requested shard version 1|1||56535ea26e3a936d394e6caf, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.019+0000 d20013| 2015-11-23T18:44:50.018+0000 I SHARDING [conn22] collection db2.coll2 was previously unsharded, new metadata loaded with shard version 1|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.019+0000 d20013| 2015-11-23T18:44:50.018+0000 I SHARDING [conn22] collection version was loaded at version 1|1||56535ea26e3a936d394e6caf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.020+0000 s20019| 2015-11-23T18:44:50.018+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:50.018+0000-56535ea26e3a936d394e6cb0", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304290018), what: "shardCollection.end", ns: "db2.coll2", details: { version: "1|1||56535ea26e3a936d394e6caf" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.028+0000 s20019| 2015-11-23T18:44:50.028+0000 I SHARDING [conn1] distributed lock 'db2.coll2/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.029+0000 s20019| 2015-11-23T18:44:50.028+0000 I SHARDING [conn1] moving chunk ns: db2.coll2 moving ( ns: db2.coll2, shard: test-rs1, lastmod: 1|0||56535ea26e3a936d394e6caf, min: { _id: MinKey }, max: { _id: 0 }) test-rs1 -> test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.030+0000 d20013| 2015-11-23T18:44:50.028+0000 I SHARDING [conn22] moveChunk waiting for full cleanup after move [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.031+0000 d20013| 2015-11-23T18:44:50.028+0000 I SHARDING [conn22] received moveChunk request: { moveChunk: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535ea26e3a936d394e6caf') ], epoch: ObjectId('56535ea26e3a936d394e6caf') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.035+0000 d20013| 2015-11-23T18:44:50.034+0000 I SHARDING [conn22] distributed lock 'db2.coll2/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'migrating chunk [{ _id: MinKey }, { _id: 0 }) in db2.coll2', ts : 56535ea26202d0bae25462ff [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.036+0000 d20013| 2015-11-23T18:44:50.034+0000 I SHARDING [conn22] remotely refreshing metadata for db2.coll2 based on current shard version 1|1||56535ea26e3a936d394e6caf, current metadata version is 1|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.036+0000 d20013| 2015-11-23T18:44:50.035+0000 I SHARDING [conn22] metadata of collection db2.coll2 already up to date (shard version : 1|1||56535ea26e3a936d394e6caf, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.037+0000 d20013| 2015-11-23T18:44:50.035+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:50.035+0000-56535ea26202d0bae2546300", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304290035), what: "moveChunk.start", ns: "db2.coll2", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.037+0000 d20013| 2015-11-23T18:44:50.035+0000 I SHARDING [conn22] moveChunk request accepted at version 1|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.038+0000 d20013| 2015-11-23T18:44:50.036+0000 I SHARDING [conn22] moveChunk number of documents: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.038+0000 d20010| 2015-11-23T18:44:50.037+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38923 #62 (56 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.038+0000 d20010| 2015-11-23T18:44:50.037+0000 I SHARDING [conn62] remotely refreshing metadata for db2.coll2, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.039+0000 d20010| 2015-11-23T18:44:50.038+0000 I SHARDING [conn62] collection db2.coll2 was previously unsharded, new metadata loaded with shard version 0|0||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.039+0000 d20010| 2015-11-23T18:44:50.038+0000 I SHARDING [conn62] collection version was loaded at version 1|1||56535ea26e3a936d394e6caf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.040+0000 d20010| 2015-11-23T18:44:50.038+0000 I SHARDING [migrateThread] starting receiving-end of migration of chunk { _id: MinKey } -> { _id: 0 } for collection db2.coll2 from test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 at epoch 56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.040+0000 d20013| 2015-11-23T18:44:50.039+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.042+0000 d20013| 2015-11-23T18:44:50.042+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.045+0000 d20010| 2015-11-23T18:44:50.045+0000 I INDEX [migrateThread] build index on: db2.coll2 properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.045+0000 d20010| 2015-11-23T18:44:50.045+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.047+0000 d20013| 2015-11-23T18:44:50.046+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.051+0000 d20010| 2015-11-23T18:44:50.051+0000 I INDEX [migrateThread] build index on: db2.coll2 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.051+0000 d20010| 2015-11-23T18:44:50.051+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.054+0000 d20010| 2015-11-23T18:44:50.053+0000 I INDEX [migrateThread] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.054+0000 d20010| 2015-11-23T18:44:50.053+0000 I SHARDING [migrateThread] Deleter starting delete for: db2.coll2 from { _id: MinKey } -> { _id: 0 }, with opId: 34543 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.055+0000 d20010| 2015-11-23T18:44:50.054+0000 I SHARDING [migrateThread] rangeDeleter deleted 0 documents for db2.coll2 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.056+0000 d20013| 2015-11-23T18:44:50.054+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.062+0000 d20011| 2015-11-23T18:44:50.062+0000 I INDEX [repl writer worker 5] build index on: db2.coll2 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.063+0000 d20011| 2015-11-23T18:44:50.062+0000 I INDEX [repl writer worker 5] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.066+0000 d20012| 2015-11-23T18:44:50.065+0000 I INDEX [repl writer worker 8] build index on: db2.coll2 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.066+0000 d20012| 2015-11-23T18:44:50.065+0000 I INDEX [repl writer worker 8] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.066+0000 d20011| 2015-11-23T18:44:50.065+0000 I INDEX [repl writer worker 5] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.068+0000 d20012| 2015-11-23T18:44:50.068+0000 I INDEX [repl writer worker 8] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.069+0000 d20010| 2015-11-23T18:44:50.068+0000 I SHARDING [migrateThread] Waiting for replication to catch up before entering critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.069+0000 d20010| 2015-11-23T18:44:50.069+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db2.coll2' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.072+0000 d20013| 2015-11-23T18:44:50.071+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "steady", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.072+0000 d20013| 2015-11-23T18:44:50.071+0000 I SHARDING [conn22] About to check if it is safe to enter critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.073+0000 d20013| 2015-11-23T18:44:50.071+0000 I SHARDING [conn22] About to enter migrate critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.073+0000 d20013| 2015-11-23T18:44:50.071+0000 I SHARDING [conn22] moveChunk setting version to: 2|0||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.073+0000 d20010| 2015-11-23T18:44:50.072+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38924 #63 (57 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.079+0000 d20010| 2015-11-23T18:44:50.079+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db2.coll2' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.080+0000 d20010| 2015-11-23T18:44:50.079+0000 I SHARDING [migrateThread] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:50.079+0000-56535ea2a5e5109e4dbf13dd", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304290079), what: "moveChunk.to", ns: "db2.coll2", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 5: 14, step 2 of 5: 15, step 3 of 5: 0, step 4 of 5: 0, step 5 of 5: 10, note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.081+0000 c20016| 2015-11-23T18:44:50.080+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48493 #40 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.081+0000 c20017| 2015-11-23T18:44:50.081+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55808 #38 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.082+0000 c20018| 2015-11-23T18:44:50.082+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36929 #38 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.084+0000 d20013| 2015-11-23T18:44:50.083+0000 I SHARDING [conn22] moveChunk migrate commit accepted by TO-shard: { active: false, ns: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "done", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.085+0000 d20013| 2015-11-23T18:44:50.084+0000 I SHARDING [conn22] moveChunk updating self version to: 2|1||56535ea26e3a936d394e6caf through { _id: 0 } -> { _id: MaxKey } for collection 'db2.coll2' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.110+0000 d20013| 2015-11-23T18:44:50.109+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:50.109+0000-56535ea26202d0bae2546301", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304290109), what: "moveChunk.commit", ns: "db2.coll2", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0", cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.110+0000 d20013| 2015-11-23T18:44:50.110+0000 I SHARDING [conn22] MigrateFromStatus::done About to acquire global lock to exit critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.110+0000 d20013| 2015-11-23T18:44:50.110+0000 I SHARDING [conn22] doing delete inline for cleanup of chunk data [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.111+0000 d20013| 2015-11-23T18:44:50.110+0000 I SHARDING [conn22] Deleter starting delete for: db2.coll2 from { _id: MinKey } -> { _id: 0 }, with opId: 522 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.111+0000 d20013| 2015-11-23T18:44:50.110+0000 I SHARDING [conn22] rangeDeleter deleted 0 documents for db2.coll2 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.112+0000 d20013| 2015-11-23T18:44:50.110+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:50.110+0000-56535ea26202d0bae2546302", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304290110), what: "moveChunk.from", ns: "db2.coll2", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 6: 0, step 2 of 6: 7, step 3 of 6: 2, step 4 of 6: 32, step 5 of 6: 39, step 6 of 6: 0, to: "test-rs0", from: "test-rs1", note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.142+0000 d20013| 2015-11-23T18:44:50.141+0000 I SHARDING [conn22] distributed lock 'db2.coll2/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.145+0000 d20013| 2015-11-23T18:44:50.141+0000 I COMMAND [conn22] command db2.coll2 command: moveChunk { moveChunk: "db2.coll2", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535ea26e3a936d394e6caf') ], epoch: ObjectId('56535ea26e3a936d394e6caf') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:134 locks:{ Global: { acquireCount: { r: 11, w: 3, R: 2 } }, Database: { acquireCount: { r: 3, w: 3 } }, Collection: { acquireCount: { r: 3, w: 1, W: 2 } } } protocol:op_command 113ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.146+0000 s20019| 2015-11-23T18:44:50.142+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db2.coll2: 0ms sequenceNumber: 12 version: 2|1||56535ea26e3a936d394e6caf based on: 1|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.146+0000 d20010| 2015-11-23T18:44:50.143+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db2.coll2", keyPattern: { _id: "hashed" }, min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs0", splitKeys: [ { _id: -4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535ea26e3a936d394e6caf') ], epoch: ObjectId('56535ea26e3a936d394e6caf') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.162+0000 d20010| 2015-11-23T18:44:50.162+0000 I SHARDING [conn26] distributed lock 'db2.coll2/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ _id: MinKey }, { _id: 0 }) in db2.coll2', ts : 56535ea2a5e5109e4dbf13de [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.163+0000 d20010| 2015-11-23T18:44:50.162+0000 I SHARDING [conn26] remotely refreshing metadata for db2.coll2 based on current shard version 0|0||56535ea26e3a936d394e6caf, current metadata version is 1|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.163+0000 d20010| 2015-11-23T18:44:50.162+0000 I SHARDING [conn26] updating metadata for db2.coll2 from shard version 0|0||56535ea26e3a936d394e6caf to shard version 2|0||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.164+0000 d20010| 2015-11-23T18:44:50.162+0000 I SHARDING [conn26] collection version was loaded at version 2|1||56535ea26e3a936d394e6caf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.164+0000 d20010| 2015-11-23T18:44:50.162+0000 I SHARDING [conn26] splitChunk accepted at version 2|0||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.176+0000 d20010| 2015-11-23T18:44:50.174+0000 I SHARDING [conn26] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:50.174+0000-56535ea2a5e5109e4dbf13df", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38802", time: new Date(1448304290174), what: "split", ns: "db2.coll2", details: { before: { min: { _id: MinKey }, max: { _id: 0 } }, left: { min: { _id: MinKey }, max: { _id: -4611686018427387902 }, lastmod: Timestamp 2000|2, lastmodEpoch: ObjectId('56535ea26e3a936d394e6caf') }, right: { min: { _id: -4611686018427387902 }, max: { _id: 0 }, lastmod: Timestamp 2000|3, lastmodEpoch: ObjectId('56535ea26e3a936d394e6caf') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.194+0000 d20010| 2015-11-23T18:44:50.194+0000 I SHARDING [conn26] distributed lock 'db2.coll2/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.195+0000 s20019| 2015-11-23T18:44:50.195+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db2.coll2: 0ms sequenceNumber: 13 version: 2|3||56535ea26e3a936d394e6caf based on: 2|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.196+0000 d20013| 2015-11-23T18:44:50.195+0000 I SHARDING [conn22] received splitChunk request: { splitChunk: "db2.coll2", keyPattern: { _id: "hashed" }, min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs1", splitKeys: [ { _id: 4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535ea26e3a936d394e6caf') ], epoch: ObjectId('56535ea26e3a936d394e6caf') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.207+0000 d20013| 2015-11-23T18:44:50.206+0000 I SHARDING [conn22] distributed lock 'db2.coll2/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ _id: 0 }, { _id: MaxKey }) in db2.coll2', ts : 56535ea26202d0bae2546303 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.208+0000 d20013| 2015-11-23T18:44:50.206+0000 I SHARDING [conn22] remotely refreshing metadata for db2.coll2 based on current shard version 2|0||56535ea26e3a936d394e6caf, current metadata version is 2|0||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.208+0000 d20013| 2015-11-23T18:44:50.207+0000 I SHARDING [conn22] updating metadata for db2.coll2 from shard version 2|0||56535ea26e3a936d394e6caf to shard version 2|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.208+0000 d20013| 2015-11-23T18:44:50.207+0000 I SHARDING [conn22] collection version was loaded at version 2|3||56535ea26e3a936d394e6caf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.209+0000 d20013| 2015-11-23T18:44:50.207+0000 I SHARDING [conn22] splitChunk accepted at version 2|1||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.217+0000 d20013| 2015-11-23T18:44:50.216+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:50.216+0000-56535ea26202d0bae2546304", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304290216), what: "split", ns: "db2.coll2", details: { before: { min: { _id: 0 }, max: { _id: MaxKey } }, left: { min: { _id: 0 }, max: { _id: 4611686018427387902 }, lastmod: Timestamp 2000|4, lastmodEpoch: ObjectId('56535ea26e3a936d394e6caf') }, right: { min: { _id: 4611686018427387902 }, max: { _id: MaxKey }, lastmod: Timestamp 2000|5, lastmodEpoch: ObjectId('56535ea26e3a936d394e6caf') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.233+0000 d20013| 2015-11-23T18:44:50.232+0000 I SHARDING [conn22] distributed lock 'db2.coll2/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.234+0000 s20019| 2015-11-23T18:44:50.233+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db2.coll2: 0ms sequenceNumber: 14 version: 2|5||56535ea26e3a936d394e6caf based on: 2|3||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.235+0000 Using 5 threads (requested 5) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.276+0000 d20010| 2015-11-23T18:44:50.276+0000 I INDEX [conn26] build index on: db2.coll2 properties: { v: 1, key: { a: 1.0, b: 1.0 }, name: "a_1_b_1", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.277+0000 d20010| 2015-11-23T18:44:50.276+0000 I INDEX [conn26] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.277+0000 d20013| 2015-11-23T18:44:50.277+0000 I INDEX [conn22] build index on: db2.coll2 properties: { v: 1, key: { a: 1.0, b: 1.0 }, name: "a_1_b_1", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.277+0000 d20013| 2015-11-23T18:44:50.277+0000 I INDEX [conn22] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.281+0000 d20010| 2015-11-23T18:44:50.280+0000 I INDEX [conn26] build index done. scanned 91 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.281+0000 d20013| 2015-11-23T18:44:50.281+0000 I INDEX [conn22] build index done. scanned 109 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.286+0000 d20011| 2015-11-23T18:44:50.285+0000 I INDEX [repl writer worker 7] build index on: db2.coll2 properties: { v: 1, key: { a: 1.0, b: 1.0 }, name: "a_1_b_1", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.287+0000 d20011| 2015-11-23T18:44:50.285+0000 I INDEX [repl writer worker 7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.287+0000 d20012| 2015-11-23T18:44:50.285+0000 I INDEX [repl writer worker 10] build index on: db2.coll2 properties: { v: 1, key: { a: 1.0, b: 1.0 }, name: "a_1_b_1", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.287+0000 d20012| 2015-11-23T18:44:50.285+0000 I INDEX [repl writer worker 10] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.288+0000 d20015| 2015-11-23T18:44:50.285+0000 I INDEX [repl writer worker 14] build index on: db2.coll2 properties: { v: 1, key: { a: 1.0, b: 1.0 }, name: "a_1_b_1", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.288+0000 d20015| 2015-11-23T18:44:50.285+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.289+0000 d20014| 2015-11-23T18:44:50.287+0000 I INDEX [repl writer worker 10] build index on: db2.coll2 properties: { v: 1, key: { a: 1.0, b: 1.0 }, name: "a_1_b_1", ns: "db2.coll2" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.289+0000 d20014| 2015-11-23T18:44:50.287+0000 I INDEX [repl writer worker 10] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.290+0000 d20012| 2015-11-23T18:44:50.287+0000 I INDEX [repl writer worker 10] build index done. scanned 91 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.290+0000 d20015| 2015-11-23T18:44:50.287+0000 I INDEX [repl writer worker 14] build index done. scanned 109 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.291+0000 d20011| 2015-11-23T18:44:50.288+0000 I INDEX [repl writer worker 7] build index done. scanned 91 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.292+0000 d20014| 2015-11-23T18:44:50.289+0000 I INDEX [repl writer worker 10] build index done. scanned 109 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.378+0000 s20019| 2015-11-23T18:44:50.377+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39982 #10 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.378+0000 s20020| 2015-11-23T18:44:50.377+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60395 #9 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.388+0000 s20020| 2015-11-23T18:44:50.388+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60396 #10 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.395+0000 s20019| 2015-11-23T18:44:50.394+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:39985 #11 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.395+0000 s20020| 2015-11-23T18:44:50.395+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60398 #11 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.404+0000 setting random seed: 403609303 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.405+0000 setting random seed: 1085143284 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.405+0000 setting random seed: 2955546455 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.405+0000 setting random seed: 2772526173 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.405+0000 setting random seed: 3332767617 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.406+0000 d20013| 2015-11-23T18:44:50.405+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36515 #28 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.407+0000 s20020| 2015-11-23T18:44:50.406+0000 I SHARDING [conn9] ChunkManager: time to load chunks for db2.coll2: 0ms sequenceNumber: 4 version: 2|5||56535ea26e3a936d394e6caf based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.407+0000 s20019| 2015-11-23T18:44:50.407+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.408+0000 d20013| 2015-11-23T18:44:50.407+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36516 #29 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.408+0000 d20013| 2015-11-23T18:44:50.407+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36517 #30 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.408+0000 d20013| 2015-11-23T18:44:50.407+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36518 #31 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.409+0000 s20020| 2015-11-23T18:44:50.408+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-1-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.409+0000 s20020| 2015-11-23T18:44:50.409+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.410+0000 s20019| 2015-11-23T18:44:50.409+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-3-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.411+0000 d20013| 2015-11-23T18:44:50.411+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36519 #32 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.414+0000 d20013| 2015-11-23T18:44:50.412+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36520 #33 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.415+0000 d20013| 2015-11-23T18:44:50.414+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36521 #34 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.417+0000 s20019| 2015-11-23T18:44:50.414+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.418+0000 d20013| 2015-11-23T18:44:50.417+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36522 #35 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.430+0000 s20020| 2015-11-23T18:44:50.424+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.435+0000 d20013| 2015-11-23T18:44:50.430+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36523 #36 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.435+0000 d20013| 2015-11-23T18:44:50.431+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36524 #37 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.436+0000 d20013| 2015-11-23T18:44:50.434+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36525 #38 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.437+0000 d20013| 2015-11-23T18:44:50.435+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36526 #39 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.438+0000 s20019| 2015-11-23T18:44:50.437+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-1-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.438+0000 s20020| 2015-11-23T18:44:50.438+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.439+0000 d20013| 2015-11-23T18:44:50.438+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36527 #40 (34 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.441+0000 s20020| 2015-11-23T18:44:50.440+0000 I NETWORK [conn11] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.442+0000 c20016| 2015-11-23T18:44:50.442+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48514 #41 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.455+0000 s20020| 2015-11-23T18:44:50.454+0000 I NETWORK [conn11] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.459+0000 c20017| 2015-11-23T18:44:50.459+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55829 #39 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.465+0000 s20020| 2015-11-23T18:44:50.464+0000 I NETWORK [conn11] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.468+0000 c20018| 2015-11-23T18:44:50.468+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36950 #39 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.577+0000 d20013| 2015-11-23T18:44:50.576+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36531 #41 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.771+0000 d20013| 2015-11-23T18:44:50.770+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36532 #42 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.772+0000 s20020| 2015-11-23T18:44:50.771+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-3-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.904+0000 d20013| 2015-11-23T18:44:50.903+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36533 #43 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:50.916+0000 s20020| 2015-11-23T18:44:50.915+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-3-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:52.441+0000 d20013| 2015-11-23T18:44:52.440+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36534 #44 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:52.443+0000 s20020| 2015-11-23T18:44:52.442+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:55.149+0000 s20019| 2015-11-23T18:44:55.148+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'doing balance round', ts : 56535ea76e3a936d394e6cb1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:55.156+0000 s20019| 2015-11-23T18:44:55.155+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:44:55.155+0000-56535ea76e3a936d394e6cb2", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304295155), what: "balancer.round", ns: "", details: { executionTimeMillis: 23, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:55.166+0000 s20019| 2015-11-23T18:44:55.165+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:55.279+0000 s20020| 2015-11-23T18:44:55.278+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' acquired for 'doing balance round', ts : 56535ea713c05a55d8fa6f92 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:55.287+0000 s20020| 2015-11-23T18:44:55.286+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:44:55.286+0000-56535ea713c05a55d8fa6f93", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304295286), what: "balancer.round", ns: "", details: { executionTimeMillis: 22, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:55.298+0000 s20020| 2015-11-23T18:44:55.298+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:55.481+0000 s20020| 2015-11-23T18:44:55.480+0000 I NETWORK [conn10] end connection 10.99.163.247:60396 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:56.143+0000 s20019| 2015-11-23T18:44:56.143+0000 I NETWORK [conn10] end connection 10.99.163.247:39982 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:56.558+0000 s20019| 2015-11-23T18:44:56.557+0000 I NETWORK [conn11] end connection 10.99.163.247:39985 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:56.564+0000 s20020| 2015-11-23T18:44:56.564+0000 I NETWORK [conn11] end connection 10.99.163.247:60398 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.093+0000 s20020| 2015-11-23T18:44:57.093+0000 I NETWORK [conn9] end connection 10.99.163.247:60395 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.100+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.101+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.101+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.101+0000 Workload(s) completed in 6866 ms: jstests/concurrency/fsm_workloads/yield_sort_merge.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.101+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.101+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.102+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.102+0000 s20019| 2015-11-23T18:44:57.100+0000 I COMMAND [conn1] DROP: db2.coll2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.103+0000 s20019| 2015-11-23T18:44:57.101+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.101+0000-56535ea96e3a936d394e6cb3", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304297101), what: "dropCollection.start", ns: "db2.coll2", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.114+0000 s20019| 2015-11-23T18:44:57.114+0000 I SHARDING [conn1] distributed lock 'db2.coll2/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535ea96e3a936d394e6cb4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.114+0000 d20010| 2015-11-23T18:44:57.114+0000 I COMMAND [conn46] CMD: drop db2.coll2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.117+0000 d20013| 2015-11-23T18:44:57.115+0000 I COMMAND [conn11] CMD: drop db2.coll2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.118+0000 d20012| 2015-11-23T18:44:57.116+0000 I COMMAND [repl writer worker 12] CMD: drop db2.coll2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.119+0000 d20011| 2015-11-23T18:44:57.116+0000 I COMMAND [repl writer worker 6] CMD: drop db2.coll2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.119+0000 d20014| 2015-11-23T18:44:57.117+0000 I COMMAND [repl writer worker 13] CMD: drop db2.coll2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.120+0000 d20015| 2015-11-23T18:44:57.117+0000 I COMMAND [repl writer worker 2] CMD: drop db2.coll2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.124+0000 d20010| 2015-11-23T18:44:57.123+0000 I SHARDING [conn46] remotely refreshing metadata for db2.coll2 with requested shard version 0|0||000000000000000000000000, current shard version is 2|3||56535ea26e3a936d394e6caf, current metadata version is 2|3||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.124+0000 d20010| 2015-11-23T18:44:57.123+0000 I SHARDING [conn46] dropping metadata for db2.coll2 at shard version 2|3||56535ea26e3a936d394e6caf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.125+0000 d20013| 2015-11-23T18:44:57.124+0000 I SHARDING [conn11] remotely refreshing metadata for db2.coll2 with requested shard version 0|0||000000000000000000000000, current shard version is 2|5||56535ea26e3a936d394e6caf, current metadata version is 2|5||56535ea26e3a936d394e6caf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.126+0000 d20013| 2015-11-23T18:44:57.124+0000 I SHARDING [conn11] dropping metadata for db2.coll2 at shard version 2|5||56535ea26e3a936d394e6caf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.126+0000 s20019| 2015-11-23T18:44:57.124+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.124+0000-56535ea96e3a936d394e6cb5", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304297124), what: "dropCollection", ns: "db2.coll2", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.135+0000 s20019| 2015-11-23T18:44:57.135+0000 I SHARDING [conn1] distributed lock 'db2.coll2/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.136+0000 s20019| 2015-11-23T18:44:57.136+0000 I COMMAND [conn1] DROP DATABASE: db2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.136+0000 s20019| 2015-11-23T18:44:57.136+0000 I SHARDING [conn1] DBConfig::dropDatabase: db2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.137+0000 s20019| 2015-11-23T18:44:57.136+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.136+0000-56535ea96e3a936d394e6cb6", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304297136), what: "dropDatabase.start", ns: "db2", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.142+0000 s20019| 2015-11-23T18:44:57.141+0000 I SHARDING [conn1] DBConfig::dropDatabase: db2 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.143+0000 d20013| 2015-11-23T18:44:57.142+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36536 #45 (39 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.143+0000 d20013| 2015-11-23T18:44:57.143+0000 I COMMAND [conn45] dropDatabase db2 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.143+0000 d20013| 2015-11-23T18:44:57.143+0000 I COMMAND [conn45] dropDatabase db2 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.144+0000 d20015| 2015-11-23T18:44:57.144+0000 I COMMAND [repl writer worker 11] dropDatabase db2 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.144+0000 d20015| 2015-11-23T18:44:57.144+0000 I COMMAND [repl writer worker 11] dropDatabase db2 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.145+0000 d20014| 2015-11-23T18:44:57.144+0000 I COMMAND [repl writer worker 14] dropDatabase db2 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.145+0000 d20014| 2015-11-23T18:44:57.144+0000 I COMMAND [repl writer worker 14] dropDatabase db2 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.145+0000 s20019| 2015-11-23T18:44:57.144+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.143+0000-56535ea96e3a936d394e6cb7", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304297143), what: "dropDatabase", ns: "db2", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.148+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.148+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.148+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.148+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.148+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.149+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.149+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.163+0000 d20010| 2015-11-23T18:44:57.162+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.163+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.163+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.163+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.164+0000 ip-10-99-163-247:20010: awaitReplication completed in 15 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.164+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.164+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.164+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.164+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.165+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.165+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.165+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.165+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.165+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.165+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.165+0000 d20012| 2015-11-23T18:44:57.163+0000 I COMMAND [repl writer worker 14] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.166+0000 d20011| 2015-11-23T18:44:57.164+0000 I COMMAND [repl writer worker 0] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.178+0000 d20013| 2015-11-23T18:44:57.178+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.178+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.178+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.178+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.179+0000 ip-10-99-163-247:20013: awaitReplication completed in 15 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.179+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.179+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.179+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.179+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.180+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.180+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.180+0000 Workload(s) started: jstests/concurrency/fsm_workloads/yield_text.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.180+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.180+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.180+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.181+0000 d20015| 2015-11-23T18:44:57.178+0000 I COMMAND [repl writer worker 9] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.181+0000 d20014| 2015-11-23T18:44:57.179+0000 I COMMAND [repl writer worker 5] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.191+0000 s20019| 2015-11-23T18:44:57.190+0000 I SHARDING [conn1] distributed lock 'db3/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535ea96e3a936d394e6cb8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.192+0000 s20019| 2015-11-23T18:44:57.192+0000 I SHARDING [conn1] Placing [db3] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.192+0000 s20019| 2015-11-23T18:44:57.192+0000 I SHARDING [conn1] Enabling sharding for database [db3] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.200+0000 s20019| 2015-11-23T18:44:57.200+0000 I SHARDING [conn1] distributed lock 'db3/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.216+0000 d20013| 2015-11-23T18:44:57.216+0000 I INDEX [conn27] build index on: db3.coll3 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db3.coll3" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.217+0000 d20013| 2015-11-23T18:44:57.216+0000 I INDEX [conn27] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.219+0000 d20013| 2015-11-23T18:44:57.218+0000 I INDEX [conn27] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.219+0000 s20019| 2015-11-23T18:44:57.219+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db3.coll3", key: { _id: "hashed" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.232+0000 d20014| 2015-11-23T18:44:57.231+0000 I INDEX [repl writer worker 4] build index on: db3.coll3 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db3.coll3" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.232+0000 d20014| 2015-11-23T18:44:57.231+0000 I INDEX [repl writer worker 4] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.236+0000 d20014| 2015-11-23T18:44:57.235+0000 I INDEX [repl writer worker 4] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.240+0000 d20015| 2015-11-23T18:44:57.239+0000 I INDEX [repl writer worker 0] build index on: db3.coll3 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db3.coll3" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.240+0000 d20015| 2015-11-23T18:44:57.239+0000 I INDEX [repl writer worker 0] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.241+0000 d20015| 2015-11-23T18:44:57.240+0000 I INDEX [repl writer worker 0] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.250+0000 s20019| 2015-11-23T18:44:57.250+0000 I SHARDING [conn1] distributed lock 'db3.coll3/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535ea96e3a936d394e6cb9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.251+0000 s20019| 2015-11-23T18:44:57.250+0000 I SHARDING [conn1] enable sharding on: db3.coll3 with shard key: { _id: "hashed" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.252+0000 s20019| 2015-11-23T18:44:57.250+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.250+0000-56535ea96e3a936d394e6cba", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304297250), what: "shardCollection.start", ns: "db3.coll3", details: { shardKey: { _id: "hashed" }, collection: "db3.coll3", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 2 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.252+0000 s20019| 2015-11-23T18:44:57.251+0000 I SHARDING [conn1] going to create 2 chunk(s) for: db3.coll3 using new epoch 56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.270+0000 s20019| 2015-11-23T18:44:57.269+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db3.coll3: 0ms sequenceNumber: 15 version: 1|1||56535ea96e3a936d394e6cbb based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.274+0000 s20019| 2015-11-23T18:44:57.274+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db3.coll3: 0ms sequenceNumber: 16 version: 1|1||56535ea96e3a936d394e6cbb based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.276+0000 d20013| 2015-11-23T18:44:57.275+0000 I SHARDING [conn41] remotely refreshing metadata for db3.coll3 with requested shard version 1|1||56535ea96e3a936d394e6cbb, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.277+0000 d20013| 2015-11-23T18:44:57.276+0000 I SHARDING [conn41] collection db3.coll3 was previously unsharded, new metadata loaded with shard version 1|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.277+0000 d20013| 2015-11-23T18:44:57.276+0000 I SHARDING [conn41] collection version was loaded at version 1|1||56535ea96e3a936d394e6cbb, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.278+0000 s20019| 2015-11-23T18:44:57.276+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.276+0000-56535ea96e3a936d394e6cbc", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304297276), what: "shardCollection.end", ns: "db3.coll3", details: { version: "1|1||56535ea96e3a936d394e6cbb" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.285+0000 s20019| 2015-11-23T18:44:57.284+0000 I SHARDING [conn1] distributed lock 'db3.coll3/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.286+0000 s20019| 2015-11-23T18:44:57.284+0000 I SHARDING [conn1] moving chunk ns: db3.coll3 moving ( ns: db3.coll3, shard: test-rs1, lastmod: 1|0||56535ea96e3a936d394e6cbb, min: { _id: MinKey }, max: { _id: 0 }) test-rs1 -> test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.286+0000 d20013| 2015-11-23T18:44:57.285+0000 I SHARDING [conn41] moveChunk waiting for full cleanup after move [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.288+0000 d20013| 2015-11-23T18:44:57.285+0000 I SHARDING [conn41] received moveChunk request: { moveChunk: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535ea96e3a936d394e6cbb') ], epoch: ObjectId('56535ea96e3a936d394e6cbb') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.297+0000 d20013| 2015-11-23T18:44:57.296+0000 I SHARDING [conn41] distributed lock 'db3.coll3/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'migrating chunk [{ _id: MinKey }, { _id: 0 }) in db3.coll3', ts : 56535ea96202d0bae2546306 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.297+0000 d20013| 2015-11-23T18:44:57.296+0000 I SHARDING [conn41] remotely refreshing metadata for db3.coll3 based on current shard version 1|1||56535ea96e3a936d394e6cbb, current metadata version is 1|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.298+0000 d20013| 2015-11-23T18:44:57.297+0000 I SHARDING [conn41] metadata of collection db3.coll3 already up to date (shard version : 1|1||56535ea96e3a936d394e6cbb, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.299+0000 d20013| 2015-11-23T18:44:57.297+0000 I SHARDING [conn41] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.297+0000-56535ea96202d0bae2546307", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36531", time: new Date(1448304297297), what: "moveChunk.start", ns: "db3.coll3", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.299+0000 c20016| 2015-11-23T18:44:57.298+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48523 #42 (39 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.299+0000 c20017| 2015-11-23T18:44:57.299+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55838 #40 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.301+0000 c20018| 2015-11-23T18:44:57.300+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36959 #40 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.302+0000 d20013| 2015-11-23T18:44:57.301+0000 I SHARDING [conn41] moveChunk request accepted at version 1|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.303+0000 d20013| 2015-11-23T18:44:57.302+0000 I SHARDING [conn41] moveChunk number of documents: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.303+0000 d20010| 2015-11-23T18:44:57.302+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38958 #64 (58 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.304+0000 d20010| 2015-11-23T18:44:57.303+0000 I SHARDING [conn64] remotely refreshing metadata for db3.coll3, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.305+0000 d20010| 2015-11-23T18:44:57.303+0000 I SHARDING [conn64] collection db3.coll3 was previously unsharded, new metadata loaded with shard version 0|0||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.305+0000 d20010| 2015-11-23T18:44:57.304+0000 I SHARDING [conn64] collection version was loaded at version 1|1||56535ea96e3a936d394e6cbb, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.306+0000 d20010| 2015-11-23T18:44:57.304+0000 I SHARDING [migrateThread] starting receiving-end of migration of chunk { _id: MinKey } -> { _id: 0 } for collection db3.coll3 from test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 at epoch 56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.307+0000 d20013| 2015-11-23T18:44:57.305+0000 I SHARDING [conn41] moveChunk data transfer progress: { active: true, ns: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.308+0000 d20013| 2015-11-23T18:44:57.307+0000 I SHARDING [conn41] moveChunk data transfer progress: { active: true, ns: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.312+0000 d20013| 2015-11-23T18:44:57.312+0000 I SHARDING [conn41] moveChunk data transfer progress: { active: true, ns: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.321+0000 d20013| 2015-11-23T18:44:57.320+0000 I SHARDING [conn41] moveChunk data transfer progress: { active: true, ns: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.325+0000 d20010| 2015-11-23T18:44:57.324+0000 I INDEX [migrateThread] build index on: db3.coll3 properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "db3.coll3" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.325+0000 d20010| 2015-11-23T18:44:57.324+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.328+0000 d20010| 2015-11-23T18:44:57.328+0000 I INDEX [migrateThread] build index on: db3.coll3 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db3.coll3" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.328+0000 d20010| 2015-11-23T18:44:57.328+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.329+0000 d20010| 2015-11-23T18:44:57.329+0000 I INDEX [migrateThread] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.330+0000 d20010| 2015-11-23T18:44:57.329+0000 I SHARDING [migrateThread] Deleter starting delete for: db3.coll3 from { _id: MinKey } -> { _id: 0 }, with opId: 40462 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.331+0000 d20010| 2015-11-23T18:44:57.330+0000 I SHARDING [migrateThread] rangeDeleter deleted 0 documents for db3.coll3 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.334+0000 d20012| 2015-11-23T18:44:57.333+0000 I INDEX [repl writer worker 4] build index on: db3.coll3 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db3.coll3" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.335+0000 d20012| 2015-11-23T18:44:57.333+0000 I INDEX [repl writer worker 4] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.335+0000 d20011| 2015-11-23T18:44:57.334+0000 I INDEX [repl writer worker 12] build index on: db3.coll3 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db3.coll3" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.336+0000 d20011| 2015-11-23T18:44:57.334+0000 I INDEX [repl writer worker 12] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.336+0000 d20012| 2015-11-23T18:44:57.335+0000 I INDEX [repl writer worker 4] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.336+0000 d20011| 2015-11-23T18:44:57.335+0000 I INDEX [repl writer worker 12] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.337+0000 d20010| 2015-11-23T18:44:57.336+0000 I SHARDING [migrateThread] Waiting for replication to catch up before entering critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.337+0000 d20013| 2015-11-23T18:44:57.336+0000 I SHARDING [conn41] moveChunk data transfer progress: { active: true, ns: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "catchup", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.338+0000 d20010| 2015-11-23T18:44:57.336+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db3.coll3' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.369+0000 d20013| 2015-11-23T18:44:57.368+0000 I SHARDING [conn41] moveChunk data transfer progress: { active: true, ns: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "steady", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.369+0000 d20013| 2015-11-23T18:44:57.368+0000 I SHARDING [conn41] About to check if it is safe to enter critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.370+0000 d20013| 2015-11-23T18:44:57.369+0000 I SHARDING [conn41] About to enter migrate critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.370+0000 d20013| 2015-11-23T18:44:57.369+0000 I SHARDING [conn41] moveChunk setting version to: 2|0||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.377+0000 d20010| 2015-11-23T18:44:57.377+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db3.coll3' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.378+0000 d20010| 2015-11-23T18:44:57.377+0000 I SHARDING [migrateThread] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.377+0000-56535ea9a5e5109e4dbf13e0", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304297377), what: "moveChunk.to", ns: "db3.coll3", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 5: 25, step 2 of 5: 6, step 3 of 5: 0, step 4 of 5: 0, step 5 of 5: 41, note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.379+0000 c20016| 2015-11-23T18:44:57.378+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48527 #43 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.380+0000 c20017| 2015-11-23T18:44:57.380+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55842 #41 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.381+0000 c20018| 2015-11-23T18:44:57.381+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36963 #41 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.383+0000 d20013| 2015-11-23T18:44:57.382+0000 I SHARDING [conn41] moveChunk migrate commit accepted by TO-shard: { active: false, ns: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "done", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.383+0000 d20013| 2015-11-23T18:44:57.382+0000 I SHARDING [conn41] moveChunk updating self version to: 2|1||56535ea96e3a936d394e6cbb through { _id: 0 } -> { _id: MaxKey } for collection 'db3.coll3' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.392+0000 d20013| 2015-11-23T18:44:57.390+0000 I SHARDING [conn41] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.390+0000-56535ea96202d0bae2546308", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36531", time: new Date(1448304297390), what: "moveChunk.commit", ns: "db3.coll3", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0", cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.392+0000 d20013| 2015-11-23T18:44:57.392+0000 I SHARDING [conn41] MigrateFromStatus::done About to acquire global lock to exit critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.393+0000 d20013| 2015-11-23T18:44:57.392+0000 I SHARDING [conn41] doing delete inline for cleanup of chunk data [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.393+0000 d20013| 2015-11-23T18:44:57.392+0000 I SHARDING [conn41] Deleter starting delete for: db3.coll3 from { _id: MinKey } -> { _id: 0 }, with opId: 6914 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.393+0000 d20013| 2015-11-23T18:44:57.392+0000 I SHARDING [conn41] rangeDeleter deleted 0 documents for db3.coll3 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.394+0000 d20013| 2015-11-23T18:44:57.392+0000 I SHARDING [conn41] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.392+0000-56535ea96202d0bae2546309", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36531", time: new Date(1448304297392), what: "moveChunk.from", ns: "db3.coll3", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 6: 0, step 2 of 6: 16, step 3 of 6: 2, step 4 of 6: 64, step 5 of 6: 23, step 6 of 6: 0, to: "test-rs0", from: "test-rs1", note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.409+0000 d20013| 2015-11-23T18:44:57.409+0000 I SHARDING [conn41] distributed lock 'db3.coll3/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.412+0000 d20013| 2015-11-23T18:44:57.409+0000 I COMMAND [conn41] command db3.coll3 command: moveChunk { moveChunk: "db3.coll3", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535ea96e3a936d394e6cbb') ], epoch: ObjectId('56535ea96e3a936d394e6cbb') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:134 locks:{ Global: { acquireCount: { r: 11, w: 3, R: 2 } }, Database: { acquireCount: { r: 3, w: 3 } }, Collection: { acquireCount: { r: 3, w: 1, W: 2 } } } protocol:op_command 124ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.412+0000 s20019| 2015-11-23T18:44:57.410+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db3.coll3: 0ms sequenceNumber: 17 version: 2|1||56535ea96e3a936d394e6cbb based on: 1|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.413+0000 d20010| 2015-11-23T18:44:57.410+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db3.coll3", keyPattern: { _id: "hashed" }, min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs0", splitKeys: [ { _id: -4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535ea96e3a936d394e6cbb') ], epoch: ObjectId('56535ea96e3a936d394e6cbb') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.427+0000 d20010| 2015-11-23T18:44:57.426+0000 I SHARDING [conn26] distributed lock 'db3.coll3/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ _id: MinKey }, { _id: 0 }) in db3.coll3', ts : 56535ea9a5e5109e4dbf13e1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.428+0000 d20010| 2015-11-23T18:44:57.426+0000 I SHARDING [conn26] remotely refreshing metadata for db3.coll3 based on current shard version 0|0||56535ea96e3a936d394e6cbb, current metadata version is 1|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.429+0000 d20010| 2015-11-23T18:44:57.427+0000 I SHARDING [conn26] updating metadata for db3.coll3 from shard version 0|0||56535ea96e3a936d394e6cbb to shard version 2|0||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.429+0000 d20010| 2015-11-23T18:44:57.427+0000 I SHARDING [conn26] collection version was loaded at version 2|1||56535ea96e3a936d394e6cbb, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.429+0000 d20010| 2015-11-23T18:44:57.427+0000 I SHARDING [conn26] splitChunk accepted at version 2|0||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.447+0000 d20010| 2015-11-23T18:44:57.445+0000 I SHARDING [conn26] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.445+0000-56535ea9a5e5109e4dbf13e2", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38802", time: new Date(1448304297445), what: "split", ns: "db3.coll3", details: { before: { min: { _id: MinKey }, max: { _id: 0 } }, left: { min: { _id: MinKey }, max: { _id: -4611686018427387902 }, lastmod: Timestamp 2000|2, lastmodEpoch: ObjectId('56535ea96e3a936d394e6cbb') }, right: { min: { _id: -4611686018427387902 }, max: { _id: 0 }, lastmod: Timestamp 2000|3, lastmodEpoch: ObjectId('56535ea96e3a936d394e6cbb') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.462+0000 d20010| 2015-11-23T18:44:57.461+0000 I SHARDING [conn26] distributed lock 'db3.coll3/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.463+0000 s20019| 2015-11-23T18:44:57.462+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db3.coll3: 0ms sequenceNumber: 18 version: 2|3||56535ea96e3a936d394e6cbb based on: 2|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.464+0000 d20013| 2015-11-23T18:44:57.462+0000 I SHARDING [conn41] received splitChunk request: { splitChunk: "db3.coll3", keyPattern: { _id: "hashed" }, min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs1", splitKeys: [ { _id: 4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535ea96e3a936d394e6cbb') ], epoch: ObjectId('56535ea96e3a936d394e6cbb') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.482+0000 d20013| 2015-11-23T18:44:57.482+0000 I SHARDING [conn41] distributed lock 'db3.coll3/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ _id: 0 }, { _id: MaxKey }) in db3.coll3', ts : 56535ea96202d0bae254630a [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.483+0000 d20013| 2015-11-23T18:44:57.482+0000 I SHARDING [conn41] remotely refreshing metadata for db3.coll3 based on current shard version 2|0||56535ea96e3a936d394e6cbb, current metadata version is 2|0||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.484+0000 d20013| 2015-11-23T18:44:57.482+0000 I SHARDING [conn41] updating metadata for db3.coll3 from shard version 2|0||56535ea96e3a936d394e6cbb to shard version 2|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.484+0000 d20013| 2015-11-23T18:44:57.482+0000 I SHARDING [conn41] collection version was loaded at version 2|3||56535ea96e3a936d394e6cbb, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.485+0000 d20013| 2015-11-23T18:44:57.483+0000 I SHARDING [conn41] splitChunk accepted at version 2|1||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.501+0000 d20013| 2015-11-23T18:44:57.499+0000 I SHARDING [conn41] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:44:57.499+0000-56535ea96202d0bae254630b", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36531", time: new Date(1448304297499), what: "split", ns: "db3.coll3", details: { before: { min: { _id: 0 }, max: { _id: MaxKey } }, left: { min: { _id: 0 }, max: { _id: 4611686018427387902 }, lastmod: Timestamp 2000|4, lastmodEpoch: ObjectId('56535ea96e3a936d394e6cbb') }, right: { min: { _id: 4611686018427387902 }, max: { _id: MaxKey }, lastmod: Timestamp 2000|5, lastmodEpoch: ObjectId('56535ea96e3a936d394e6cbb') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.519+0000 d20013| 2015-11-23T18:44:57.519+0000 I SHARDING [conn41] distributed lock 'db3.coll3/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.521+0000 s20019| 2015-11-23T18:44:57.520+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db3.coll3: 0ms sequenceNumber: 19 version: 2|5||56535ea96e3a936d394e6cbb based on: 2|3||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.521+0000 Using 5 threads (requested 5) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.572+0000 d20010| 2015-11-23T18:44:57.570+0000 I INDEX [conn26] build index on: db3.coll3 properties: { v: 1, key: { _fts: "text", _ftsx: 1 }, name: "yield_text_text", ns: "db3.coll3", weights: { yield_text: 1 }, default_language: "english", language_override: "language", textIndexVersion: 3 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.572+0000 d20010| 2015-11-23T18:44:57.570+0000 I INDEX [conn26] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.575+0000 d20013| 2015-11-23T18:44:57.574+0000 I INDEX [conn41] build index on: db3.coll3 properties: { v: 1, key: { _fts: "text", _ftsx: 1 }, name: "yield_text_text", ns: "db3.coll3", weights: { yield_text: 1 }, default_language: "english", language_override: "language", textIndexVersion: 3 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.577+0000 d20010| 2015-11-23T18:44:57.574+0000 I INDEX [conn26] build index done. scanned 91 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.577+0000 d20013| 2015-11-23T18:44:57.574+0000 I INDEX [conn41] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.578+0000 d20011| 2015-11-23T18:44:57.576+0000 I INDEX [repl writer worker 7] build index on: db3.coll3 properties: { v: 1, key: { _fts: "text", _ftsx: 1 }, name: "yield_text_text", ns: "db3.coll3", weights: { yield_text: 1 }, default_language: "english", language_override: "language", textIndexVersion: 3 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.578+0000 d20011| 2015-11-23T18:44:57.576+0000 I INDEX [repl writer worker 7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.581+0000 d20012| 2015-11-23T18:44:57.580+0000 I INDEX [repl writer worker 14] build index on: db3.coll3 properties: { v: 1, key: { _fts: "text", _ftsx: 1 }, name: "yield_text_text", ns: "db3.coll3", weights: { yield_text: 1 }, default_language: "english", language_override: "language", textIndexVersion: 3 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.581+0000 d20012| 2015-11-23T18:44:57.580+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.582+0000 d20011| 2015-11-23T18:44:57.580+0000 I INDEX [repl writer worker 7] build index done. scanned 91 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.582+0000 d20013| 2015-11-23T18:44:57.582+0000 I INDEX [conn41] build index done. scanned 109 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.586+0000 d20012| 2015-11-23T18:44:57.584+0000 I INDEX [repl writer worker 14] build index done. scanned 91 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.586+0000 d20014| 2015-11-23T18:44:57.585+0000 I INDEX [repl writer worker 3] build index on: db3.coll3 properties: { v: 1, key: { _fts: "text", _ftsx: 1 }, name: "yield_text_text", ns: "db3.coll3", weights: { yield_text: 1 }, default_language: "english", language_override: "language", textIndexVersion: 3 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.587+0000 d20014| 2015-11-23T18:44:57.585+0000 I INDEX [repl writer worker 3] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.590+0000 d20015| 2015-11-23T18:44:57.587+0000 I INDEX [repl writer worker 11] build index on: db3.coll3 properties: { v: 1, key: { _fts: "text", _ftsx: 1 }, name: "yield_text_text", ns: "db3.coll3", weights: { yield_text: 1 }, default_language: "english", language_override: "language", textIndexVersion: 3 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.591+0000 d20015| 2015-11-23T18:44:57.587+0000 I INDEX [repl writer worker 11] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.592+0000 d20014| 2015-11-23T18:44:57.591+0000 I INDEX [repl writer worker 3] build index done. scanned 109 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.592+0000 d20015| 2015-11-23T18:44:57.591+0000 I INDEX [repl writer worker 11] build index done. scanned 109 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.679+0000 s20019| 2015-11-23T18:44:57.678+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40016 #12 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.689+0000 s20020| 2015-11-23T18:44:57.689+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60429 #12 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.691+0000 s20020| 2015-11-23T18:44:57.691+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60430 #13 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.697+0000 s20019| 2015-11-23T18:44:57.697+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40019 #13 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.703+0000 s20019| 2015-11-23T18:44:57.703+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40020 #14 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.710+0000 setting random seed: 3815883497 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.710+0000 setting random seed: 2792769733 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.710+0000 setting random seed: 2766452617 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.711+0000 setting random seed: 2609677257 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.711+0000 setting random seed: 1189650663 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.714+0000 s20020| 2015-11-23T18:44:57.712+0000 I SHARDING [conn12] ChunkManager: time to load chunks for db3.coll3: 0ms sequenceNumber: 5 version: 2|5||56535ea96e3a936d394e6cbb based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:57.716+0000 d20013| 2015-11-23T18:44:57.716+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36549 #46 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:58.678+0000 d20013| 2015-11-23T18:44:58.678+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36550 #47 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:58.685+0000 s20019| 2015-11-23T18:44:58.684+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-3-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:58.941+0000 s20020| 2015-11-23T18:44:58.940+0000 I NETWORK [conn12] end connection 10.99.163.247:60429 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.243+0000 d20010| 2015-11-23T18:44:59.242+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38969 #65 (59 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.246+0000 s20019| 2015-11-23T18:44:59.245+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-3-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.774+0000 d20010| 2015-11-23T18:44:59.774+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:38970 #66 (60 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.775+0000 d20013| 2015-11-23T18:44:59.774+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36553 #48 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.777+0000 s20019| 2015-11-23T18:44:59.777+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.778+0000 s20019| 2015-11-23T18:44:59.777+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.934+0000 s20019| 2015-11-23T18:44:59.933+0000 I NETWORK [conn13] end connection 10.99.163.247:40019 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:44:59.983+0000 s20020| 2015-11-23T18:44:59.982+0000 I NETWORK [conn13] end connection 10.99.163.247:60430 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.239+0000 s20019| 2015-11-23T18:45:00.239+0000 I NETWORK [conn12] end connection 10.99.163.247:40016 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.827+0000 s20019| 2015-11-23T18:45:00.827+0000 I NETWORK [conn14] end connection 10.99.163.247:40020 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.835+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.835+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.835+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.835+0000 Workload(s) completed in 3314 ms: jstests/concurrency/fsm_workloads/yield_text.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.835+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.836+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.836+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.836+0000 s20019| 2015-11-23T18:45:00.835+0000 I COMMAND [conn1] DROP: db3.coll3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.837+0000 s20019| 2015-11-23T18:45:00.835+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:00.835+0000-56535eac6e3a936d394e6cbd", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304300835), what: "dropCollection.start", ns: "db3.coll3", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.847+0000 s20019| 2015-11-23T18:45:00.847+0000 I SHARDING [conn1] distributed lock 'db3.coll3/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535eac6e3a936d394e6cbe [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.847+0000 d20010| 2015-11-23T18:45:00.847+0000 I COMMAND [conn46] CMD: drop db3.coll3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.848+0000 d20013| 2015-11-23T18:45:00.848+0000 I COMMAND [conn11] CMD: drop db3.coll3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.849+0000 d20011| 2015-11-23T18:45:00.848+0000 I COMMAND [repl writer worker 12] CMD: drop db3.coll3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.849+0000 d20012| 2015-11-23T18:45:00.848+0000 I COMMAND [repl writer worker 11] CMD: drop db3.coll3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.851+0000 d20015| 2015-11-23T18:45:00.849+0000 I COMMAND [repl writer worker 5] CMD: drop db3.coll3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.851+0000 d20014| 2015-11-23T18:45:00.850+0000 I COMMAND [repl writer worker 3] CMD: drop db3.coll3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.857+0000 d20010| 2015-11-23T18:45:00.855+0000 I SHARDING [conn46] remotely refreshing metadata for db3.coll3 with requested shard version 0|0||000000000000000000000000, current shard version is 2|3||56535ea96e3a936d394e6cbb, current metadata version is 2|3||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.857+0000 d20010| 2015-11-23T18:45:00.856+0000 I SHARDING [conn46] dropping metadata for db3.coll3 at shard version 2|3||56535ea96e3a936d394e6cbb, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.858+0000 d20013| 2015-11-23T18:45:00.856+0000 I SHARDING [conn11] remotely refreshing metadata for db3.coll3 with requested shard version 0|0||000000000000000000000000, current shard version is 2|5||56535ea96e3a936d394e6cbb, current metadata version is 2|5||56535ea96e3a936d394e6cbb [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.858+0000 d20013| 2015-11-23T18:45:00.856+0000 I SHARDING [conn11] dropping metadata for db3.coll3 at shard version 2|5||56535ea96e3a936d394e6cbb, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.859+0000 s20019| 2015-11-23T18:45:00.857+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:00.857+0000-56535eac6e3a936d394e6cbf", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304300857), what: "dropCollection", ns: "db3.coll3", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.867+0000 s20019| 2015-11-23T18:45:00.867+0000 I SHARDING [conn1] distributed lock 'db3.coll3/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.868+0000 s20019| 2015-11-23T18:45:00.868+0000 I COMMAND [conn1] DROP DATABASE: db3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.869+0000 s20019| 2015-11-23T18:45:00.868+0000 I SHARDING [conn1] DBConfig::dropDatabase: db3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.869+0000 s20019| 2015-11-23T18:45:00.868+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:00.868+0000-56535eac6e3a936d394e6cc0", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304300868), what: "dropDatabase.start", ns: "db3", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.872+0000 s20019| 2015-11-23T18:45:00.871+0000 I SHARDING [conn1] DBConfig::dropDatabase: db3 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.873+0000 d20013| 2015-11-23T18:45:00.872+0000 I COMMAND [conn45] dropDatabase db3 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.873+0000 d20013| 2015-11-23T18:45:00.872+0000 I COMMAND [conn45] dropDatabase db3 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.873+0000 s20019| 2015-11-23T18:45:00.872+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:00.872+0000-56535eac6e3a936d394e6cc1", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304300872), what: "dropDatabase", ns: "db3", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.874+0000 d20014| 2015-11-23T18:45:00.872+0000 I COMMAND [repl writer worker 4] dropDatabase db3 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.874+0000 d20014| 2015-11-23T18:45:00.872+0000 I COMMAND [repl writer worker 4] dropDatabase db3 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.874+0000 d20015| 2015-11-23T18:45:00.872+0000 I COMMAND [repl writer worker 13] dropDatabase db3 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.875+0000 d20015| 2015-11-23T18:45:00.872+0000 I COMMAND [repl writer worker 13] dropDatabase db3 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.876+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.877+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.877+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.877+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.877+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.877+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.877+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.892+0000 d20010| 2015-11-23T18:45:00.891+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.892+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.892+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.892+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.893+0000 ip-10-99-163-247:20010: awaitReplication completed in 16 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.893+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.893+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.893+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.893+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.893+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.893+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.894+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.894+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.894+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.894+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.894+0000 d20012| 2015-11-23T18:45:00.892+0000 I COMMAND [repl writer worker 2] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.894+0000 d20011| 2015-11-23T18:45:00.893+0000 I COMMAND [repl writer worker 6] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.911+0000 d20013| 2015-11-23T18:45:00.910+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.911+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.911+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 ip-10-99-163-247:20013: awaitReplication completed in 19 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 d20014| 2015-11-23T18:45:00.911+0000 I COMMAND [repl writer worker 15] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 d20015| 2015-11-23T18:45:00.911+0000 I COMMAND [repl writer worker 8] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.912+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.913+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.913+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.913+0000 Workload(s) started: jstests/concurrency/fsm_workloads/findAndModify_remove.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.913+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.913+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.913+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.953+0000 s20019| 2015-11-23T18:45:00.953+0000 I SHARDING [conn1] distributed lock 'db4/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535eac6e3a936d394e6cc2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.955+0000 s20019| 2015-11-23T18:45:00.955+0000 I SHARDING [conn1] Placing [db4] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.955+0000 s20019| 2015-11-23T18:45:00.955+0000 I SHARDING [conn1] Enabling sharding for database [db4] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.981+0000 s20019| 2015-11-23T18:45:00.981+0000 I SHARDING [conn1] distributed lock 'db4/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:00.999+0000 d20013| 2015-11-23T18:45:00.998+0000 I INDEX [conn46] build index on: db4.coll4 properties: { v: 1, key: { tid: 1.0 }, name: "tid_1", ns: "db4.coll4" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.000+0000 d20013| 2015-11-23T18:45:00.999+0000 I INDEX [conn46] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.005+0000 d20013| 2015-11-23T18:45:01.004+0000 I INDEX [conn46] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.005+0000 s20019| 2015-11-23T18:45:01.005+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db4.coll4", key: { tid: 1.0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.016+0000 d20015| 2015-11-23T18:45:01.015+0000 I INDEX [repl writer worker 0] build index on: db4.coll4 properties: { v: 1, key: { tid: 1.0 }, name: "tid_1", ns: "db4.coll4" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.016+0000 d20015| 2015-11-23T18:45:01.015+0000 I INDEX [repl writer worker 0] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.022+0000 d20015| 2015-11-23T18:45:01.021+0000 I INDEX [repl writer worker 0] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.026+0000 d20014| 2015-11-23T18:45:01.026+0000 I INDEX [repl writer worker 0] build index on: db4.coll4 properties: { v: 1, key: { tid: 1.0 }, name: "tid_1", ns: "db4.coll4" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.027+0000 d20014| 2015-11-23T18:45:01.026+0000 I INDEX [repl writer worker 0] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.028+0000 d20014| 2015-11-23T18:45:01.028+0000 I INDEX [repl writer worker 0] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.045+0000 s20019| 2015-11-23T18:45:01.045+0000 I SHARDING [conn1] distributed lock 'db4.coll4/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535ead6e3a936d394e6cc3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.046+0000 s20019| 2015-11-23T18:45:01.045+0000 I SHARDING [conn1] enable sharding on: db4.coll4 with shard key: { tid: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.047+0000 s20019| 2015-11-23T18:45:01.045+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:01.045+0000-56535ead6e3a936d394e6cc4", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304301045), what: "shardCollection.start", ns: "db4.coll4", details: { shardKey: { tid: 1.0 }, collection: "db4.coll4", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 1 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.048+0000 s20019| 2015-11-23T18:45:01.047+0000 I SHARDING [conn1] going to create 1 chunk(s) for: db4.coll4 using new epoch 56535ead6e3a936d394e6cc5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.063+0000 s20019| 2015-11-23T18:45:01.063+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db4.coll4: 0ms sequenceNumber: 20 version: 1|0||56535ead6e3a936d394e6cc5 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.070+0000 s20019| 2015-11-23T18:45:01.069+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db4.coll4: 0ms sequenceNumber: 21 version: 1|0||56535ead6e3a936d394e6cc5 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.071+0000 d20013| 2015-11-23T18:45:01.070+0000 I SHARDING [conn22] remotely refreshing metadata for db4.coll4 with requested shard version 1|0||56535ead6e3a936d394e6cc5, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.071+0000 d20013| 2015-11-23T18:45:01.070+0000 I SHARDING [conn22] collection db4.coll4 was previously unsharded, new metadata loaded with shard version 1|0||56535ead6e3a936d394e6cc5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.071+0000 d20013| 2015-11-23T18:45:01.070+0000 I SHARDING [conn22] collection version was loaded at version 1|0||56535ead6e3a936d394e6cc5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.072+0000 s20019| 2015-11-23T18:45:01.070+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:01.070+0000-56535ead6e3a936d394e6cc6", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304301070), what: "shardCollection.end", ns: "db4.coll4", details: { version: "1|0||56535ead6e3a936d394e6cc5" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.093+0000 s20019| 2015-11-23T18:45:01.093+0000 I SHARDING [conn1] distributed lock 'db4.coll4/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.094+0000 Using 20 threads (requested 20) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.421+0000 s20020| 2015-11-23T18:45:01.420+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60438 #14 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.454+0000 s20020| 2015-11-23T18:45:01.453+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60439 #15 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.462+0000 s20019| 2015-11-23T18:45:01.458+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40028 #15 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.467+0000 s20019| 2015-11-23T18:45:01.466+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40029 #16 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.469+0000 s20019| 2015-11-23T18:45:01.469+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40030 #17 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.475+0000 s20019| 2015-11-23T18:45:01.473+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40031 #18 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.493+0000 s20020| 2015-11-23T18:45:01.493+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60444 #16 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.497+0000 s20020| 2015-11-23T18:45:01.497+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60445 #17 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.512+0000 s20020| 2015-11-23T18:45:01.511+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60446 #18 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.522+0000 s20019| 2015-11-23T18:45:01.521+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40035 #19 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.522+0000 s20019| 2015-11-23T18:45:01.522+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40036 #20 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.524+0000 s20020| 2015-11-23T18:45:01.524+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60449 #19 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.529+0000 s20020| 2015-11-23T18:45:01.528+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60450 #20 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.529+0000 s20019| 2015-11-23T18:45:01.529+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40039 #21 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.530+0000 s20020| 2015-11-23T18:45:01.529+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60452 #21 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.536+0000 s20020| 2015-11-23T18:45:01.536+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60453 #22 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.538+0000 s20019| 2015-11-23T18:45:01.537+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40042 #22 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.539+0000 s20019| 2015-11-23T18:45:01.538+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40043 #23 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.540+0000 s20020| 2015-11-23T18:45:01.539+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60456 #23 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.542+0000 s20019| 2015-11-23T18:45:01.541+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40045 #24 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.556+0000 setting random seed: 467648958 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.557+0000 setting random seed: 3938354208 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.557+0000 setting random seed: 701611605 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.557+0000 setting random seed: 1049246839 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.557+0000 setting random seed: 2214659565 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.557+0000 setting random seed: 1978593421 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.558+0000 setting random seed: 545270601 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.558+0000 setting random seed: 1520153083 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.558+0000 setting random seed: 3862169230 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.558+0000 setting random seed: 3922019564 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.559+0000 d20013| 2015-11-23T18:45:01.550+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36574 #49 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.559+0000 d20013| 2015-11-23T18:45:01.551+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36575 #50 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.561+0000 setting random seed: 2026945511 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.561+0000 setting random seed: 3398344234 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.561+0000 setting random seed: 3340681519 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.562+0000 setting random seed: 2158312172 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.563+0000 s20020| 2015-11-23T18:45:01.555+0000 I SHARDING [conn14] ChunkManager: time to load chunks for db4.coll4: 0ms sequenceNumber: 6 version: 1|0||56535ead6e3a936d394e6cc5 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.563+0000 d20013| 2015-11-23T18:45:01.558+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36576 #51 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.563+0000 s20019| 2015-11-23T18:45:01.559+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.564+0000 setting random seed: 196246786 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.564+0000 setting random seed: 2447961747 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.565+0000 setting random seed: 170997112 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.565+0000 d20013| 2015-11-23T18:45:01.562+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36577 #52 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.566+0000 d20013| 2015-11-23T18:45:01.563+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36578 #53 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.566+0000 setting random seed: 512085848 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.566+0000 setting random seed: 705868431 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.566+0000 setting random seed: 3193851791 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.569+0000 d20013| 2015-11-23T18:45:01.567+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36579 #54 (48 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.570+0000 d20013| 2015-11-23T18:45:01.567+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36580 #55 (49 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.572+0000 d20013| 2015-11-23T18:45:01.571+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36581 #56 (50 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.577+0000 d20013| 2015-11-23T18:45:01.574+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36582 #57 (51 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.578+0000 d20013| 2015-11-23T18:45:01.576+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36583 #58 (52 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.597+0000 s20020| 2015-11-23T18:45:01.589+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.629+0000 d20013| 2015-11-23T18:45:01.599+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36584 #59 (53 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.640+0000 d20013| 2015-11-23T18:45:01.626+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36585 #60 (54 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.642+0000 d20013| 2015-11-23T18:45:01.632+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36586 #61 (55 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.646+0000 d20013| 2015-11-23T18:45:01.643+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36587 #62 (56 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.647+0000 d20013| 2015-11-23T18:45:01.644+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36588 #63 (57 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.654+0000 d20013| 2015-11-23T18:45:01.647+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36589 #64 (58 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.666+0000 d20013| 2015-11-23T18:45:01.662+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36590 #65 (59 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.681+0000 d20013| 2015-11-23T18:45:01.663+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36591 #66 (60 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.738+0000 d20013| 2015-11-23T18:45:01.736+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36592 #67 (61 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.777+0000 s20020| 2015-11-23T18:45:01.776+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.782+0000 d20013| 2015-11-23T18:45:01.781+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36593 #68 (62 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.813+0000 d20015| 2015-11-23T18:45:01.813+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.913+0000 s20019| 2015-11-23T18:45:01.912+0000 I NETWORK [conn17] end connection 10.99.163.247:40030 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.917+0000 s20020| 2015-11-23T18:45:01.916+0000 I NETWORK [conn16] end connection 10.99.163.247:60444 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.917+0000 d20013| 2015-11-23T18:45:01.917+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36594 #69 (63 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.919+0000 s20020| 2015-11-23T18:45:01.919+0000 I NETWORK [conn18] end connection 10.99.163.247:60446 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.926+0000 d20014| 2015-11-23T18:45:01.926+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.927+0000 s20019| 2015-11-23T18:45:01.926+0000 I NETWORK [conn15] end connection 10.99.163.247:40028 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.929+0000 s20020| 2015-11-23T18:45:01.927+0000 I NETWORK [conn14] end connection 10.99.163.247:60438 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.929+0000 s20019| 2015-11-23T18:45:01.928+0000 I NETWORK [conn20] end connection 10.99.163.247:40036 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.930+0000 s20020| 2015-11-23T18:45:01.930+0000 I NETWORK [conn22] end connection 10.99.163.247:60453 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.931+0000 s20019| 2015-11-23T18:45:01.930+0000 I NETWORK [conn19] end connection 10.99.163.247:40035 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.932+0000 s20019| 2015-11-23T18:45:01.931+0000 I NETWORK [conn16] end connection 10.99.163.247:40029 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.932+0000 s20019| 2015-11-23T18:45:01.931+0000 I NETWORK [conn21] end connection 10.99.163.247:40039 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.933+0000 s20020| 2015-11-23T18:45:01.932+0000 I NETWORK [conn17] end connection 10.99.163.247:60445 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.939+0000 s20019| 2015-11-23T18:45:01.938+0000 I NETWORK [conn18] end connection 10.99.163.247:40031 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.942+0000 s20020| 2015-11-23T18:45:01.941+0000 I NETWORK [conn21] end connection 10.99.163.247:60452 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.944+0000 s20019| 2015-11-23T18:45:01.943+0000 I NETWORK [conn22] end connection 10.99.163.247:40042 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.945+0000 s20020| 2015-11-23T18:45:01.944+0000 I NETWORK [conn19] end connection 10.99.163.247:60449 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.949+0000 s20020| 2015-11-23T18:45:01.949+0000 I NETWORK [conn15] end connection 10.99.163.247:60439 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.951+0000 s20019| 2015-11-23T18:45:01.950+0000 I NETWORK [conn24] end connection 10.99.163.247:40045 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.951+0000 s20020| 2015-11-23T18:45:01.950+0000 I NETWORK [conn23] end connection 10.99.163.247:60456 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.952+0000 s20020| 2015-11-23T18:45:01.951+0000 I NETWORK [conn20] end connection 10.99.163.247:60450 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.961+0000 s20019| 2015-11-23T18:45:01.961+0000 I NETWORK [conn23] end connection 10.99.163.247:40043 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.963+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.963+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.963+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.964+0000 Workload(s) completed in 870 ms: jstests/concurrency/fsm_workloads/findAndModify_remove.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.964+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.964+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.964+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.964+0000 s20019| 2015-11-23T18:45:01.963+0000 I COMMAND [conn1] DROP: db4.coll4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.965+0000 s20019| 2015-11-23T18:45:01.964+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:01.963+0000-56535ead6e3a936d394e6cc7", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304301963), what: "dropCollection.start", ns: "db4.coll4", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.975+0000 s20019| 2015-11-23T18:45:01.974+0000 I SHARDING [conn1] distributed lock 'db4.coll4/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535ead6e3a936d394e6cc8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.975+0000 d20010| 2015-11-23T18:45:01.974+0000 I COMMAND [conn46] CMD: drop db4.coll4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.976+0000 d20013| 2015-11-23T18:45:01.975+0000 I COMMAND [conn50] CMD: drop db4.coll4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.978+0000 d20015| 2015-11-23T18:45:01.977+0000 I COMMAND [repl writer worker 6] CMD: drop db4.coll4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.978+0000 d20014| 2015-11-23T18:45:01.978+0000 I COMMAND [repl writer worker 2] CMD: drop db4.coll4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.982+0000 d20013| 2015-11-23T18:45:01.981+0000 I SHARDING [conn50] remotely refreshing metadata for db4.coll4 with requested shard version 0|0||000000000000000000000000, current shard version is 1|0||56535ead6e3a936d394e6cc5, current metadata version is 1|0||56535ead6e3a936d394e6cc5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.982+0000 d20013| 2015-11-23T18:45:01.981+0000 I SHARDING [conn50] dropping metadata for db4.coll4 at shard version 1|0||56535ead6e3a936d394e6cc5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:01.983+0000 s20019| 2015-11-23T18:45:01.981+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:01.981+0000-56535ead6e3a936d394e6cc9", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304301981), what: "dropCollection", ns: "db4.coll4", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.005+0000 s20019| 2015-11-23T18:45:02.004+0000 I SHARDING [conn1] distributed lock 'db4.coll4/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.005+0000 s20019| 2015-11-23T18:45:02.005+0000 I COMMAND [conn1] DROP DATABASE: db4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.006+0000 s20019| 2015-11-23T18:45:02.005+0000 I SHARDING [conn1] DBConfig::dropDatabase: db4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.007+0000 s20019| 2015-11-23T18:45:02.005+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.005+0000-56535eae6e3a936d394e6cca", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304302005), what: "dropDatabase.start", ns: "db4", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.017+0000 s20019| 2015-11-23T18:45:02.016+0000 I SHARDING [conn1] DBConfig::dropDatabase: db4 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.018+0000 d20013| 2015-11-23T18:45:02.017+0000 I COMMAND [conn45] dropDatabase db4 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.018+0000 d20013| 2015-11-23T18:45:02.017+0000 I COMMAND [conn45] dropDatabase db4 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.019+0000 s20019| 2015-11-23T18:45:02.017+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.017+0000-56535eae6e3a936d394e6ccb", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304302017), what: "dropDatabase", ns: "db4", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.019+0000 d20015| 2015-11-23T18:45:02.017+0000 I COMMAND [repl writer worker 14] dropDatabase db4 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.019+0000 d20015| 2015-11-23T18:45:02.017+0000 I COMMAND [repl writer worker 14] dropDatabase db4 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.020+0000 d20014| 2015-11-23T18:45:02.018+0000 I COMMAND [repl writer worker 10] dropDatabase db4 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.020+0000 d20014| 2015-11-23T18:45:02.018+0000 I COMMAND [repl writer worker 10] dropDatabase db4 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.021+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.021+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.021+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.022+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.022+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.022+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.022+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.054+0000 d20010| 2015-11-23T18:45:02.054+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.055+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.055+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.055+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.055+0000 ip-10-99-163-247:20010: awaitReplication completed in 33 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.055+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.055+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.056+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.057+0000 d20011| 2015-11-23T18:45:02.054+0000 I COMMAND [repl writer worker 3] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.057+0000 d20012| 2015-11-23T18:45:02.054+0000 I COMMAND [repl writer worker 10] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.082+0000 d20013| 2015-11-23T18:45:02.082+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.083+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.083+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.083+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.083+0000 ip-10-99-163-247:20013: awaitReplication completed in 28 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.084+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.084+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.085+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.085+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.085+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.085+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.085+0000 Workload(s) started: jstests/concurrency/fsm_workloads/findAndModify_inc.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.085+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.085+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.086+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.086+0000 d20014| 2015-11-23T18:45:02.082+0000 I COMMAND [repl writer worker 8] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.087+0000 d20015| 2015-11-23T18:45:02.083+0000 I COMMAND [repl writer worker 8] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.107+0000 s20019| 2015-11-23T18:45:02.107+0000 I SHARDING [conn1] distributed lock 'db5/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535eae6e3a936d394e6ccc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.109+0000 s20019| 2015-11-23T18:45:02.108+0000 I SHARDING [conn1] Placing [db5] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.109+0000 s20019| 2015-11-23T18:45:02.108+0000 I SHARDING [conn1] Enabling sharding for database [db5] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.119+0000 s20019| 2015-11-23T18:45:02.118+0000 I SHARDING [conn1] distributed lock 'db5/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.129+0000 d20013| 2015-11-23T18:45:02.128+0000 I INDEX [conn53] build index on: db5.coll5 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db5.coll5" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.130+0000 d20013| 2015-11-23T18:45:02.129+0000 I INDEX [conn53] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.131+0000 d20013| 2015-11-23T18:45:02.129+0000 I INDEX [conn53] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.132+0000 s20019| 2015-11-23T18:45:02.130+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db5.coll5", key: { _id: "hashed" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.138+0000 d20015| 2015-11-23T18:45:02.137+0000 I INDEX [repl writer worker 7] build index on: db5.coll5 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db5.coll5" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.138+0000 d20015| 2015-11-23T18:45:02.137+0000 I INDEX [repl writer worker 7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.139+0000 d20014| 2015-11-23T18:45:02.137+0000 I INDEX [repl writer worker 12] build index on: db5.coll5 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db5.coll5" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.139+0000 d20014| 2015-11-23T18:45:02.137+0000 I INDEX [repl writer worker 12] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.139+0000 d20015| 2015-11-23T18:45:02.138+0000 I INDEX [repl writer worker 7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.140+0000 d20014| 2015-11-23T18:45:02.139+0000 I INDEX [repl writer worker 12] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.148+0000 s20019| 2015-11-23T18:45:02.147+0000 I SHARDING [conn1] distributed lock 'db5.coll5/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535eae6e3a936d394e6ccd [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.148+0000 s20019| 2015-11-23T18:45:02.148+0000 I SHARDING [conn1] enable sharding on: db5.coll5 with shard key: { _id: "hashed" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.149+0000 s20019| 2015-11-23T18:45:02.148+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.148+0000-56535eae6e3a936d394e6cce", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304302148), what: "shardCollection.start", ns: "db5.coll5", details: { shardKey: { _id: "hashed" }, collection: "db5.coll5", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 2 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.149+0000 s20019| 2015-11-23T18:45:02.149+0000 I SHARDING [conn1] going to create 2 chunk(s) for: db5.coll5 using new epoch 56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.156+0000 s20019| 2015-11-23T18:45:02.155+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db5.coll5: 0ms sequenceNumber: 22 version: 1|1||56535eae6e3a936d394e6ccf based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.161+0000 s20019| 2015-11-23T18:45:02.161+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db5.coll5: 0ms sequenceNumber: 23 version: 1|1||56535eae6e3a936d394e6ccf based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.162+0000 d20013| 2015-11-23T18:45:02.161+0000 I SHARDING [conn61] remotely refreshing metadata for db5.coll5 with requested shard version 1|1||56535eae6e3a936d394e6ccf, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.162+0000 d20013| 2015-11-23T18:45:02.162+0000 I SHARDING [conn61] collection db5.coll5 was previously unsharded, new metadata loaded with shard version 1|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.163+0000 d20013| 2015-11-23T18:45:02.162+0000 I SHARDING [conn61] collection version was loaded at version 1|1||56535eae6e3a936d394e6ccf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.163+0000 s20019| 2015-11-23T18:45:02.162+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.162+0000-56535eae6e3a936d394e6cd0", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304302162), what: "shardCollection.end", ns: "db5.coll5", details: { version: "1|1||56535eae6e3a936d394e6ccf" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.173+0000 s20019| 2015-11-23T18:45:02.172+0000 I SHARDING [conn1] distributed lock 'db5.coll5/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.173+0000 s20019| 2015-11-23T18:45:02.173+0000 I SHARDING [conn1] moving chunk ns: db5.coll5 moving ( ns: db5.coll5, shard: test-rs1, lastmod: 1|0||56535eae6e3a936d394e6ccf, min: { _id: MinKey }, max: { _id: 0 }) test-rs1 -> test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.173+0000 d20013| 2015-11-23T18:45:02.173+0000 I SHARDING [conn61] moveChunk waiting for full cleanup after move [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.175+0000 d20013| 2015-11-23T18:45:02.173+0000 I SHARDING [conn61] received moveChunk request: { moveChunk: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eae6e3a936d394e6ccf') ], epoch: ObjectId('56535eae6e3a936d394e6ccf') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.182+0000 d20013| 2015-11-23T18:45:02.181+0000 I SHARDING [conn61] distributed lock 'db5.coll5/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'migrating chunk [{ _id: MinKey }, { _id: 0 }) in db5.coll5', ts : 56535eae6202d0bae254630e [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.182+0000 d20013| 2015-11-23T18:45:02.181+0000 I SHARDING [conn61] remotely refreshing metadata for db5.coll5 based on current shard version 1|1||56535eae6e3a936d394e6ccf, current metadata version is 1|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.183+0000 d20013| 2015-11-23T18:45:02.182+0000 I SHARDING [conn61] metadata of collection db5.coll5 already up to date (shard version : 1|1||56535eae6e3a936d394e6ccf, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.184+0000 d20013| 2015-11-23T18:45:02.182+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.182+0000-56535eae6202d0bae254630f", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304302182), what: "moveChunk.start", ns: "db5.coll5", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.184+0000 c20016| 2015-11-23T18:45:02.183+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48582 #44 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.184+0000 c20017| 2015-11-23T18:45:02.184+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55897 #42 (39 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.186+0000 c20018| 2015-11-23T18:45:02.185+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37018 #42 (39 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.186+0000 d20013| 2015-11-23T18:45:02.186+0000 I SHARDING [conn61] moveChunk request accepted at version 1|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.187+0000 d20013| 2015-11-23T18:45:02.186+0000 I SHARDING [conn61] moveChunk number of documents: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.187+0000 d20010| 2015-11-23T18:45:02.187+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39017 #67 (61 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.188+0000 d20010| 2015-11-23T18:45:02.188+0000 I SHARDING [conn67] remotely refreshing metadata for db5.coll5, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.188+0000 d20010| 2015-11-23T18:45:02.188+0000 I SHARDING [conn67] collection db5.coll5 was previously unsharded, new metadata loaded with shard version 0|0||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.189+0000 d20010| 2015-11-23T18:45:02.188+0000 I SHARDING [conn67] collection version was loaded at version 1|1||56535eae6e3a936d394e6ccf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.189+0000 d20010| 2015-11-23T18:45:02.188+0000 I SHARDING [migrateThread] starting receiving-end of migration of chunk { _id: MinKey } -> { _id: 0 } for collection db5.coll5 from test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 at epoch 56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.190+0000 d20013| 2015-11-23T18:45:02.190+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.193+0000 d20013| 2015-11-23T18:45:02.192+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.197+0000 d20013| 2015-11-23T18:45:02.196+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.197+0000 d20010| 2015-11-23T18:45:02.197+0000 I INDEX [migrateThread] build index on: db5.coll5 properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "db5.coll5" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.198+0000 d20010| 2015-11-23T18:45:02.197+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.204+0000 d20010| 2015-11-23T18:45:02.204+0000 I INDEX [migrateThread] build index on: db5.coll5 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db5.coll5" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.204+0000 d20010| 2015-11-23T18:45:02.204+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.205+0000 d20013| 2015-11-23T18:45:02.205+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.206+0000 d20010| 2015-11-23T18:45:02.206+0000 I INDEX [migrateThread] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.206+0000 d20010| 2015-11-23T18:45:02.206+0000 I SHARDING [migrateThread] Deleter starting delete for: db5.coll5 from { _id: MinKey } -> { _id: 0 }, with opId: 43487 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.207+0000 d20010| 2015-11-23T18:45:02.207+0000 I SHARDING [migrateThread] rangeDeleter deleted 0 documents for db5.coll5 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.221+0000 d20011| 2015-11-23T18:45:02.221+0000 I INDEX [repl writer worker 4] build index on: db5.coll5 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db5.coll5" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.222+0000 d20011| 2015-11-23T18:45:02.221+0000 I INDEX [repl writer worker 4] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.222+0000 d20013| 2015-11-23T18:45:02.222+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.223+0000 d20012| 2015-11-23T18:45:02.221+0000 I INDEX [repl writer worker 14] build index on: db5.coll5 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db5.coll5" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.223+0000 d20012| 2015-11-23T18:45:02.221+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.223+0000 d20011| 2015-11-23T18:45:02.222+0000 I INDEX [repl writer worker 4] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.227+0000 d20012| 2015-11-23T18:45:02.227+0000 I INDEX [repl writer worker 14] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.235+0000 d20010| 2015-11-23T18:45:02.234+0000 I SHARDING [migrateThread] Waiting for replication to catch up before entering critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.235+0000 d20010| 2015-11-23T18:45:02.234+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db5.coll5' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.255+0000 d20013| 2015-11-23T18:45:02.254+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "steady", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.255+0000 d20013| 2015-11-23T18:45:02.254+0000 I SHARDING [conn61] About to check if it is safe to enter critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.255+0000 d20013| 2015-11-23T18:45:02.254+0000 I SHARDING [conn61] About to enter migrate critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.256+0000 d20013| 2015-11-23T18:45:02.254+0000 I SHARDING [conn61] moveChunk setting version to: 2|0||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.256+0000 d20010| 2015-11-23T18:45:02.255+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db5.coll5' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.257+0000 d20010| 2015-11-23T18:45:02.255+0000 I SHARDING [migrateThread] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.255+0000-56535eaea5e5109e4dbf13e3", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304302255), what: "moveChunk.to", ns: "db5.coll5", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 5: 17, step 2 of 5: 28, step 3 of 5: 0, step 4 of 5: 0, step 5 of 5: 20, note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.258+0000 d20013| 2015-11-23T18:45:02.257+0000 I SHARDING [conn61] moveChunk migrate commit accepted by TO-shard: { active: false, ns: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "done", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.258+0000 d20013| 2015-11-23T18:45:02.257+0000 I SHARDING [conn61] moveChunk updating self version to: 2|1||56535eae6e3a936d394e6ccf through { _id: 0 } -> { _id: MaxKey } for collection 'db5.coll5' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.281+0000 d20013| 2015-11-23T18:45:02.279+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.279+0000-56535eae6202d0bae2546310", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304302279), what: "moveChunk.commit", ns: "db5.coll5", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0", cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.281+0000 d20013| 2015-11-23T18:45:02.280+0000 I SHARDING [conn61] MigrateFromStatus::done About to acquire global lock to exit critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.281+0000 d20013| 2015-11-23T18:45:02.280+0000 I SHARDING [conn61] doing delete inline for cleanup of chunk data [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.282+0000 d20013| 2015-11-23T18:45:02.280+0000 I SHARDING [conn61] Deleter starting delete for: db5.coll5 from { _id: MinKey } -> { _id: 0 }, with opId: 11612 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.282+0000 d20013| 2015-11-23T18:45:02.280+0000 I SHARDING [conn61] rangeDeleter deleted 0 documents for db5.coll5 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.283+0000 d20013| 2015-11-23T18:45:02.280+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.280+0000-56535eae6202d0bae2546311", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304302280), what: "moveChunk.from", ns: "db5.coll5", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 6: 0, step 2 of 6: 13, step 3 of 6: 2, step 4 of 6: 65, step 5 of 6: 25, step 6 of 6: 0, to: "test-rs0", from: "test-rs1", note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.307+0000 d20013| 2015-11-23T18:45:02.306+0000 I SHARDING [conn61] distributed lock 'db5.coll5/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.309+0000 d20013| 2015-11-23T18:45:02.306+0000 I COMMAND [conn61] command db5.coll5 command: moveChunk { moveChunk: "db5.coll5", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eae6e3a936d394e6ccf') ], epoch: ObjectId('56535eae6e3a936d394e6ccf') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:154 locks:{ Global: { acquireCount: { r: 11, w: 3, R: 2 } }, Database: { acquireCount: { r: 3, w: 3 } }, Collection: { acquireCount: { r: 3, w: 1, W: 2 } } } protocol:op_command 133ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.310+0000 s20019| 2015-11-23T18:45:02.307+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db5.coll5: 0ms sequenceNumber: 24 version: 2|1||56535eae6e3a936d394e6ccf based on: 1|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.310+0000 d20010| 2015-11-23T18:45:02.308+0000 I SHARDING [conn30] received splitChunk request: { splitChunk: "db5.coll5", keyPattern: { _id: "hashed" }, min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs0", splitKeys: [ { _id: -4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eae6e3a936d394e6ccf') ], epoch: ObjectId('56535eae6e3a936d394e6ccf') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.325+0000 d20010| 2015-11-23T18:45:02.325+0000 I SHARDING [conn30] distributed lock 'db5.coll5/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ _id: MinKey }, { _id: 0 }) in db5.coll5', ts : 56535eaea5e5109e4dbf13e4 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.326+0000 d20010| 2015-11-23T18:45:02.325+0000 I SHARDING [conn30] remotely refreshing metadata for db5.coll5 based on current shard version 0|0||56535eae6e3a936d394e6ccf, current metadata version is 1|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.326+0000 d20010| 2015-11-23T18:45:02.325+0000 I SHARDING [conn30] updating metadata for db5.coll5 from shard version 0|0||56535eae6e3a936d394e6ccf to shard version 2|0||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.327+0000 d20010| 2015-11-23T18:45:02.325+0000 I SHARDING [conn30] collection version was loaded at version 2|1||56535eae6e3a936d394e6ccf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.327+0000 d20010| 2015-11-23T18:45:02.325+0000 I SHARDING [conn30] splitChunk accepted at version 2|0||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.336+0000 d20010| 2015-11-23T18:45:02.334+0000 I SHARDING [conn30] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.334+0000-56535eaea5e5109e4dbf13e5", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38806", time: new Date(1448304302334), what: "split", ns: "db5.coll5", details: { before: { min: { _id: MinKey }, max: { _id: 0 } }, left: { min: { _id: MinKey }, max: { _id: -4611686018427387902 }, lastmod: Timestamp 2000|2, lastmodEpoch: ObjectId('56535eae6e3a936d394e6ccf') }, right: { min: { _id: -4611686018427387902 }, max: { _id: 0 }, lastmod: Timestamp 2000|3, lastmodEpoch: ObjectId('56535eae6e3a936d394e6ccf') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.350+0000 d20010| 2015-11-23T18:45:02.349+0000 I SHARDING [conn30] distributed lock 'db5.coll5/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.350+0000 s20019| 2015-11-23T18:45:02.350+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db5.coll5: 0ms sequenceNumber: 25 version: 2|3||56535eae6e3a936d394e6ccf based on: 2|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.351+0000 d20013| 2015-11-23T18:45:02.350+0000 I SHARDING [conn61] received splitChunk request: { splitChunk: "db5.coll5", keyPattern: { _id: "hashed" }, min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs1", splitKeys: [ { _id: 4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eae6e3a936d394e6ccf') ], epoch: ObjectId('56535eae6e3a936d394e6ccf') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.368+0000 d20013| 2015-11-23T18:45:02.367+0000 I SHARDING [conn61] distributed lock 'db5.coll5/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ _id: 0 }, { _id: MaxKey }) in db5.coll5', ts : 56535eae6202d0bae2546312 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.368+0000 d20013| 2015-11-23T18:45:02.367+0000 I SHARDING [conn61] remotely refreshing metadata for db5.coll5 based on current shard version 2|0||56535eae6e3a936d394e6ccf, current metadata version is 2|0||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.369+0000 d20013| 2015-11-23T18:45:02.368+0000 I SHARDING [conn61] updating metadata for db5.coll5 from shard version 2|0||56535eae6e3a936d394e6ccf to shard version 2|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.369+0000 d20013| 2015-11-23T18:45:02.368+0000 I SHARDING [conn61] collection version was loaded at version 2|3||56535eae6e3a936d394e6ccf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.370+0000 d20013| 2015-11-23T18:45:02.368+0000 I SHARDING [conn61] splitChunk accepted at version 2|1||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.377+0000 d20013| 2015-11-23T18:45:02.375+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:02.375+0000-56535eae6202d0bae2546313", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304302375), what: "split", ns: "db5.coll5", details: { before: { min: { _id: 0 }, max: { _id: MaxKey } }, left: { min: { _id: 0 }, max: { _id: 4611686018427387902 }, lastmod: Timestamp 2000|4, lastmodEpoch: ObjectId('56535eae6e3a936d394e6ccf') }, right: { min: { _id: 4611686018427387902 }, max: { _id: MaxKey }, lastmod: Timestamp 2000|5, lastmodEpoch: ObjectId('56535eae6e3a936d394e6ccf') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.388+0000 d20013| 2015-11-23T18:45:02.388+0000 I SHARDING [conn61] distributed lock 'db5.coll5/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.389+0000 s20019| 2015-11-23T18:45:02.388+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db5.coll5: 0ms sequenceNumber: 26 version: 2|5||56535eae6e3a936d394e6ccf based on: 2|3||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.390+0000 Using 20 threads (requested 20) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.740+0000 s20019| 2015-11-23T18:45:02.739+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40072 #25 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.752+0000 s20019| 2015-11-23T18:45:02.752+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40073 #26 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.759+0000 s20020| 2015-11-23T18:45:02.759+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60486 #24 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.772+0000 s20020| 2015-11-23T18:45:02.772+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60487 #25 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.789+0000 s20020| 2015-11-23T18:45:02.788+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60488 #26 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.795+0000 s20020| 2015-11-23T18:45:02.795+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60489 #27 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.800+0000 s20020| 2015-11-23T18:45:02.800+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60490 #28 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.801+0000 s20020| 2015-11-23T18:45:02.801+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60491 #29 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.803+0000 s20019| 2015-11-23T18:45:02.802+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40080 #27 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.808+0000 s20020| 2015-11-23T18:45:02.808+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60493 #30 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.815+0000 s20020| 2015-11-23T18:45:02.815+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60494 #31 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.816+0000 s20020| 2015-11-23T18:45:02.816+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60495 #32 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.821+0000 s20019| 2015-11-23T18:45:02.820+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40084 #28 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.823+0000 s20019| 2015-11-23T18:45:02.822+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40085 #29 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.825+0000 s20019| 2015-11-23T18:45:02.824+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40086 #30 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.826+0000 s20019| 2015-11-23T18:45:02.825+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40087 #31 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.827+0000 s20020| 2015-11-23T18:45:02.827+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60500 #33 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.829+0000 s20019| 2015-11-23T18:45:02.829+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40089 #32 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.831+0000 s20019| 2015-11-23T18:45:02.831+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40090 #33 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.838+0000 s20019| 2015-11-23T18:45:02.838+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40091 #34 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.841+0000 setting random seed: 1017320269 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.841+0000 setting random seed: 260806865 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.842+0000 setting random seed: 2869422577 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.842+0000 setting random seed: 2790782119 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.842+0000 setting random seed: 582298524 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.842+0000 setting random seed: 3457041541 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.842+0000 setting random seed: 222545149 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.843+0000 setting random seed: 424359633 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.843+0000 setting random seed: 547268168 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.843+0000 setting random seed: 725853792 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.850+0000 setting random seed: 2761469973 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.850+0000 setting random seed: 3719011677 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.850+0000 s20020| 2015-11-23T18:45:02.844+0000 I SHARDING [conn26] ChunkManager: time to load chunks for db5.coll5: 0ms sequenceNumber: 7 version: 2|5||56535eae6e3a936d394e6ccf based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.851+0000 s20020| 2015-11-23T18:45:02.846+0000 I NETWORK [conn24] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.851+0000 s20020| 2015-11-23T18:45:02.846+0000 I NETWORK [conn29] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.852+0000 s20020| 2015-11-23T18:45:02.846+0000 I NETWORK [conn26] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.852+0000 setting random seed: 3999606765 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.852+0000 setting random seed: 3678798296 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.852+0000 setting random seed: 785792294 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.852+0000 setting random seed: 1088066487 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.853+0000 setting random seed: 619867116 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.853+0000 setting random seed: 581592434 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.860+0000 setting random seed: 164005026 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.860+0000 setting random seed: 2230850871 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.878+0000 c20016| 2015-11-23T18:45:02.872+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48606 #45 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.900+0000 d20010| 2015-11-23T18:45:02.878+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39039 #68 (62 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.903+0000 s20020| 2015-11-23T18:45:02.900+0000 I NETWORK [conn29] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.907+0000 s20019| 2015-11-23T18:45:02.903+0000 I NETWORK [conn26] end connection 10.99.163.247:40073 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.908+0000 c20017| 2015-11-23T18:45:02.903+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55922 #43 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.918+0000 c20016| 2015-11-23T18:45:02.911+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48609 #46 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.918+0000 c20016| 2015-11-23T18:45:02.912+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48610 #47 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.921+0000 s20019| 2015-11-23T18:45:02.920+0000 I NETWORK [conn29] end connection 10.99.163.247:40085 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.928+0000 d20010| 2015-11-23T18:45:02.927+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39043 #69 (63 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.929+0000 d20010| 2015-11-23T18:45:02.929+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39044 #70 (64 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.930+0000 s20020| 2015-11-23T18:45:02.930+0000 I NETWORK [conn27] end connection 10.99.163.247:60489 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.933+0000 s20019| 2015-11-23T18:45:02.933+0000 I NETWORK [conn32] end connection 10.99.163.247:40089 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.934+0000 s20020| 2015-11-23T18:45:02.934+0000 I NETWORK [conn26] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.936+0000 d20010| 2015-11-23T18:45:02.936+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39045 #71 (65 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.938+0000 c20017| 2015-11-23T18:45:02.937+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55928 #44 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.938+0000 s20020| 2015-11-23T18:45:02.938+0000 I NETWORK [conn24] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.945+0000 s20020| 2015-11-23T18:45:02.944+0000 I NETWORK [conn29] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.948+0000 s20019| 2015-11-23T18:45:02.948+0000 I NETWORK [conn27] end connection 10.99.163.247:40080 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.949+0000 c20018| 2015-11-23T18:45:02.948+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37049 #43 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.952+0000 s20019| 2015-11-23T18:45:02.951+0000 I NETWORK [conn25] end connection 10.99.163.247:40072 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.952+0000 s20019| 2015-11-23T18:45:02.952+0000 I NETWORK [conn30] end connection 10.99.163.247:40086 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.955+0000 s20019| 2015-11-23T18:45:02.954+0000 I NETWORK [conn28] end connection 10.99.163.247:40084 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.955+0000 s20020| 2015-11-23T18:45:02.954+0000 I NETWORK [conn26] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.956+0000 c20018| 2015-11-23T18:45:02.955+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37050 #44 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.960+0000 c20017| 2015-11-23T18:45:02.959+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:55931 #45 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.961+0000 s20020| 2015-11-23T18:45:02.961+0000 I NETWORK [conn24] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.963+0000 c20018| 2015-11-23T18:45:02.962+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37052 #45 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.966+0000 s20020| 2015-11-23T18:45:02.966+0000 I NETWORK [conn30] end connection 10.99.163.247:60493 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.971+0000 s20020| 2015-11-23T18:45:02.971+0000 I NETWORK [conn33] end connection 10.99.163.247:60500 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:02.993+0000 s20019| 2015-11-23T18:45:02.992+0000 I NETWORK [conn33] end connection 10.99.163.247:40090 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.006+0000 s20019| 2015-11-23T18:45:03.005+0000 I NETWORK [conn34] end connection 10.99.163.247:40091 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.010+0000 s20019| 2015-11-23T18:45:03.009+0000 I NETWORK [conn31] end connection 10.99.163.247:40087 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.029+0000 s20020| 2015-11-23T18:45:03.029+0000 I NETWORK [conn31] end connection 10.99.163.247:60494 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.031+0000 s20020| 2015-11-23T18:45:03.030+0000 I NETWORK [conn28] end connection 10.99.163.247:60490 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.032+0000 s20020| 2015-11-23T18:45:03.031+0000 I NETWORK [conn25] end connection 10.99.163.247:60487 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.032+0000 s20020| 2015-11-23T18:45:03.032+0000 I NETWORK [conn26] end connection 10.99.163.247:60488 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.033+0000 s20020| 2015-11-23T18:45:03.032+0000 I NETWORK [conn24] end connection 10.99.163.247:60486 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.033+0000 s20020| 2015-11-23T18:45:03.033+0000 I NETWORK [conn29] end connection 10.99.163.247:60491 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.035+0000 s20020| 2015-11-23T18:45:03.035+0000 I NETWORK [conn32] end connection 10.99.163.247:60495 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.046+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.046+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.046+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.047+0000 Workload(s) completed in 656 ms: jstests/concurrency/fsm_workloads/findAndModify_inc.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.047+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.047+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.047+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.047+0000 s20019| 2015-11-23T18:45:03.046+0000 I COMMAND [conn1] DROP: db5.coll5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.048+0000 s20019| 2015-11-23T18:45:03.046+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.046+0000-56535eaf6e3a936d394e6cd1", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304303046), what: "dropCollection.start", ns: "db5.coll5", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.067+0000 d20010| 2015-11-23T18:45:03.067+0000 I COMMAND [conn46] CMD: drop db5.coll5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.068+0000 s20019| 2015-11-23T18:45:03.067+0000 I SHARDING [conn1] distributed lock 'db5.coll5/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535eaf6e3a936d394e6cd2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.069+0000 d20013| 2015-11-23T18:45:03.068+0000 I COMMAND [conn50] CMD: drop db5.coll5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.070+0000 d20012| 2015-11-23T18:45:03.068+0000 I COMMAND [repl writer worker 9] CMD: drop db5.coll5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.070+0000 d20011| 2015-11-23T18:45:03.069+0000 I COMMAND [repl writer worker 15] CMD: drop db5.coll5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.071+0000 d20014| 2015-11-23T18:45:03.070+0000 I COMMAND [repl writer worker 9] CMD: drop db5.coll5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.071+0000 d20015| 2015-11-23T18:45:03.070+0000 I COMMAND [repl writer worker 5] CMD: drop db5.coll5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.080+0000 d20010| 2015-11-23T18:45:03.078+0000 I SHARDING [conn46] remotely refreshing metadata for db5.coll5 with requested shard version 0|0||000000000000000000000000, current shard version is 2|3||56535eae6e3a936d394e6ccf, current metadata version is 2|3||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.080+0000 d20010| 2015-11-23T18:45:03.079+0000 I SHARDING [conn46] dropping metadata for db5.coll5 at shard version 2|3||56535eae6e3a936d394e6ccf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.081+0000 d20013| 2015-11-23T18:45:03.079+0000 I SHARDING [conn50] remotely refreshing metadata for db5.coll5 with requested shard version 0|0||000000000000000000000000, current shard version is 2|5||56535eae6e3a936d394e6ccf, current metadata version is 2|5||56535eae6e3a936d394e6ccf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.081+0000 d20013| 2015-11-23T18:45:03.079+0000 I SHARDING [conn50] dropping metadata for db5.coll5 at shard version 2|5||56535eae6e3a936d394e6ccf, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.082+0000 s20019| 2015-11-23T18:45:03.080+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.080+0000-56535eaf6e3a936d394e6cd3", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304303080), what: "dropCollection", ns: "db5.coll5", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.097+0000 s20019| 2015-11-23T18:45:03.096+0000 I SHARDING [conn1] distributed lock 'db5.coll5/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.098+0000 s20019| 2015-11-23T18:45:03.097+0000 I COMMAND [conn1] DROP DATABASE: db5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.098+0000 s20019| 2015-11-23T18:45:03.097+0000 I SHARDING [conn1] DBConfig::dropDatabase: db5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.099+0000 s20019| 2015-11-23T18:45:03.097+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.097+0000-56535eaf6e3a936d394e6cd4", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304303097), what: "dropDatabase.start", ns: "db5", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.104+0000 s20019| 2015-11-23T18:45:03.104+0000 I SHARDING [conn1] DBConfig::dropDatabase: db5 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.105+0000 d20013| 2015-11-23T18:45:03.104+0000 I COMMAND [conn45] dropDatabase db5 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.105+0000 d20013| 2015-11-23T18:45:03.104+0000 I COMMAND [conn45] dropDatabase db5 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.106+0000 s20019| 2015-11-23T18:45:03.104+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.104+0000-56535eaf6e3a936d394e6cd5", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304303104), what: "dropDatabase", ns: "db5", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.106+0000 d20014| 2015-11-23T18:45:03.105+0000 I COMMAND [repl writer worker 7] dropDatabase db5 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.106+0000 d20014| 2015-11-23T18:45:03.105+0000 I COMMAND [repl writer worker 7] dropDatabase db5 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.106+0000 d20015| 2015-11-23T18:45:03.105+0000 I COMMAND [repl writer worker 11] dropDatabase db5 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.107+0000 d20015| 2015-11-23T18:45:03.105+0000 I COMMAND [repl writer worker 11] dropDatabase db5 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.112+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.112+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.112+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.112+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.113+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.113+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.113+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.127+0000 d20010| 2015-11-23T18:45:03.127+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.127+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.128+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.128+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.128+0000 ip-10-99-163-247:20010: awaitReplication completed in 15 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.128+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.128+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.129+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.130+0000 d20012| 2015-11-23T18:45:03.128+0000 I COMMAND [repl writer worker 8] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.130+0000 d20011| 2015-11-23T18:45:03.128+0000 I COMMAND [repl writer worker 2] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.140+0000 d20013| 2015-11-23T18:45:03.139+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.140+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.140+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.140+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 ip-10-99-163-247:20013: awaitReplication completed in 13 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 Workload(s) started: jstests/concurrency/fsm_workloads/indexed_insert_2d.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.141+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.142+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.142+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.142+0000 d20015| 2015-11-23T18:45:03.140+0000 I COMMAND [repl writer worker 0] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.142+0000 d20014| 2015-11-23T18:45:03.140+0000 I COMMAND [repl writer worker 6] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.160+0000 s20019| 2015-11-23T18:45:03.160+0000 I SHARDING [conn1] distributed lock 'db6/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535eaf6e3a936d394e6cd6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.162+0000 s20019| 2015-11-23T18:45:03.161+0000 I SHARDING [conn1] Placing [db6] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.162+0000 s20019| 2015-11-23T18:45:03.161+0000 I SHARDING [conn1] Enabling sharding for database [db6] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.169+0000 s20019| 2015-11-23T18:45:03.168+0000 I SHARDING [conn1] distributed lock 'db6/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.177+0000 d20013| 2015-11-23T18:45:03.177+0000 I INDEX [conn53] build index on: db6.coll6 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.177+0000 d20013| 2015-11-23T18:45:03.177+0000 I INDEX [conn53] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.180+0000 d20013| 2015-11-23T18:45:03.179+0000 I INDEX [conn53] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.181+0000 s20019| 2015-11-23T18:45:03.180+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db6.coll6", key: { _id: "hashed" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.187+0000 d20015| 2015-11-23T18:45:03.186+0000 I INDEX [repl writer worker 10] build index on: db6.coll6 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.187+0000 d20015| 2015-11-23T18:45:03.186+0000 I INDEX [repl writer worker 10] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.188+0000 d20015| 2015-11-23T18:45:03.187+0000 I INDEX [repl writer worker 10] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.190+0000 d20014| 2015-11-23T18:45:03.189+0000 I INDEX [repl writer worker 3] build index on: db6.coll6 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.190+0000 d20014| 2015-11-23T18:45:03.189+0000 I INDEX [repl writer worker 3] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.191+0000 d20014| 2015-11-23T18:45:03.191+0000 I INDEX [repl writer worker 3] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.203+0000 s20019| 2015-11-23T18:45:03.203+0000 I SHARDING [conn1] distributed lock 'db6.coll6/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535eaf6e3a936d394e6cd7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.204+0000 s20019| 2015-11-23T18:45:03.203+0000 I SHARDING [conn1] enable sharding on: db6.coll6 with shard key: { _id: "hashed" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.205+0000 s20019| 2015-11-23T18:45:03.203+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.203+0000-56535eaf6e3a936d394e6cd8", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304303203), what: "shardCollection.start", ns: "db6.coll6", details: { shardKey: { _id: "hashed" }, collection: "db6.coll6", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 2 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.205+0000 s20019| 2015-11-23T18:45:03.204+0000 I SHARDING [conn1] going to create 2 chunk(s) for: db6.coll6 using new epoch 56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.212+0000 s20019| 2015-11-23T18:45:03.211+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db6.coll6: 0ms sequenceNumber: 27 version: 1|1||56535eaf6e3a936d394e6cd9 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.227+0000 s20019| 2015-11-23T18:45:03.227+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db6.coll6: 0ms sequenceNumber: 28 version: 1|1||56535eaf6e3a936d394e6cd9 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.228+0000 d20013| 2015-11-23T18:45:03.227+0000 I SHARDING [conn61] remotely refreshing metadata for db6.coll6 with requested shard version 1|1||56535eaf6e3a936d394e6cd9, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.228+0000 d20013| 2015-11-23T18:45:03.228+0000 I SHARDING [conn61] collection db6.coll6 was previously unsharded, new metadata loaded with shard version 1|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.229+0000 d20013| 2015-11-23T18:45:03.228+0000 I SHARDING [conn61] collection version was loaded at version 1|1||56535eaf6e3a936d394e6cd9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.230+0000 s20019| 2015-11-23T18:45:03.228+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.228+0000-56535eaf6e3a936d394e6cda", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304303228), what: "shardCollection.end", ns: "db6.coll6", details: { version: "1|1||56535eaf6e3a936d394e6cd9" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.256+0000 s20019| 2015-11-23T18:45:03.255+0000 I SHARDING [conn1] distributed lock 'db6.coll6/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.256+0000 s20019| 2015-11-23T18:45:03.256+0000 I SHARDING [conn1] moving chunk ns: db6.coll6 moving ( ns: db6.coll6, shard: test-rs1, lastmod: 1|0||56535eaf6e3a936d394e6cd9, min: { _id: MinKey }, max: { _id: 0 }) test-rs1 -> test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.257+0000 d20013| 2015-11-23T18:45:03.256+0000 I SHARDING [conn61] moveChunk waiting for full cleanup after move [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.258+0000 d20013| 2015-11-23T18:45:03.256+0000 I SHARDING [conn61] received moveChunk request: { moveChunk: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eaf6e3a936d394e6cd9') ], epoch: ObjectId('56535eaf6e3a936d394e6cd9') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.285+0000 d20013| 2015-11-23T18:45:03.284+0000 I SHARDING [conn61] distributed lock 'db6.coll6/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'migrating chunk [{ _id: MinKey }, { _id: 0 }) in db6.coll6', ts : 56535eaf6202d0bae2546315 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.285+0000 d20013| 2015-11-23T18:45:03.284+0000 I SHARDING [conn61] remotely refreshing metadata for db6.coll6 based on current shard version 1|1||56535eaf6e3a936d394e6cd9, current metadata version is 1|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.286+0000 d20013| 2015-11-23T18:45:03.284+0000 I SHARDING [conn61] metadata of collection db6.coll6 already up to date (shard version : 1|1||56535eaf6e3a936d394e6cd9, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.287+0000 d20013| 2015-11-23T18:45:03.284+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.284+0000-56535eaf6202d0bae2546316", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304303284), what: "moveChunk.start", ns: "db6.coll6", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.287+0000 d20013| 2015-11-23T18:45:03.285+0000 I SHARDING [conn61] moveChunk request accepted at version 1|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.287+0000 d20013| 2015-11-23T18:45:03.285+0000 I SHARDING [conn61] moveChunk number of documents: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.287+0000 d20010| 2015-11-23T18:45:03.285+0000 I SHARDING [conn67] remotely refreshing metadata for db6.coll6, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.288+0000 d20010| 2015-11-23T18:45:03.286+0000 I SHARDING [conn67] collection db6.coll6 was previously unsharded, new metadata loaded with shard version 0|0||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.288+0000 d20010| 2015-11-23T18:45:03.286+0000 I SHARDING [conn67] collection version was loaded at version 1|1||56535eaf6e3a936d394e6cd9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.289+0000 d20010| 2015-11-23T18:45:03.286+0000 I SHARDING [migrateThread] starting receiving-end of migration of chunk { _id: MinKey } -> { _id: 0 } for collection db6.coll6 from test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 at epoch 56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.289+0000 d20013| 2015-11-23T18:45:03.287+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.290+0000 d20013| 2015-11-23T18:45:03.290+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.295+0000 d20013| 2015-11-23T18:45:03.294+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.302+0000 d20010| 2015-11-23T18:45:03.301+0000 I INDEX [migrateThread] build index on: db6.coll6 properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.302+0000 d20010| 2015-11-23T18:45:03.301+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.303+0000 d20013| 2015-11-23T18:45:03.302+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.312+0000 d20010| 2015-11-23T18:45:03.312+0000 I INDEX [migrateThread] build index on: db6.coll6 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.313+0000 d20010| 2015-11-23T18:45:03.312+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.319+0000 d20013| 2015-11-23T18:45:03.318+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.321+0000 d20010| 2015-11-23T18:45:03.321+0000 I INDEX [migrateThread] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.322+0000 d20010| 2015-11-23T18:45:03.321+0000 I SHARDING [migrateThread] Deleter starting delete for: db6.coll6 from { _id: MinKey } -> { _id: 0 }, with opId: 44150 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.322+0000 d20010| 2015-11-23T18:45:03.321+0000 I SHARDING [migrateThread] rangeDeleter deleted 0 documents for db6.coll6 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.328+0000 d20011| 2015-11-23T18:45:03.328+0000 I INDEX [repl writer worker 5] build index on: db6.coll6 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.329+0000 d20011| 2015-11-23T18:45:03.328+0000 I INDEX [repl writer worker 5] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.333+0000 d20011| 2015-11-23T18:45:03.332+0000 I INDEX [repl writer worker 5] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.335+0000 d20012| 2015-11-23T18:45:03.334+0000 I INDEX [repl writer worker 3] build index on: db6.coll6 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.335+0000 d20012| 2015-11-23T18:45:03.334+0000 I INDEX [repl writer worker 3] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.339+0000 d20012| 2015-11-23T18:45:03.338+0000 I INDEX [repl writer worker 3] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.340+0000 d20010| 2015-11-23T18:45:03.339+0000 I SHARDING [migrateThread] Waiting for replication to catch up before entering critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.340+0000 d20010| 2015-11-23T18:45:03.339+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db6.coll6' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.352+0000 d20013| 2015-11-23T18:45:03.351+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "steady", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.352+0000 d20013| 2015-11-23T18:45:03.351+0000 I SHARDING [conn61] About to check if it is safe to enter critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.352+0000 d20013| 2015-11-23T18:45:03.351+0000 I SHARDING [conn61] About to enter migrate critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.352+0000 d20013| 2015-11-23T18:45:03.351+0000 I SHARDING [conn61] moveChunk setting version to: 2|0||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.360+0000 d20010| 2015-11-23T18:45:03.360+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db6.coll6' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.361+0000 d20010| 2015-11-23T18:45:03.360+0000 I SHARDING [migrateThread] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.360+0000-56535eafa5e5109e4dbf13e6", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304303360), what: "moveChunk.to", ns: "db6.coll6", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 5: 34, step 2 of 5: 17, step 3 of 5: 0, step 4 of 5: 0, step 5 of 5: 20, note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.362+0000 d20013| 2015-11-23T18:45:03.361+0000 I SHARDING [conn61] moveChunk migrate commit accepted by TO-shard: { active: false, ns: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "done", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.362+0000 d20013| 2015-11-23T18:45:03.361+0000 I SHARDING [conn61] moveChunk updating self version to: 2|1||56535eaf6e3a936d394e6cd9 through { _id: 0 } -> { _id: MaxKey } for collection 'db6.coll6' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.372+0000 d20013| 2015-11-23T18:45:03.370+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.370+0000-56535eaf6202d0bae2546317", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304303370), what: "moveChunk.commit", ns: "db6.coll6", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0", cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.372+0000 d20013| 2015-11-23T18:45:03.371+0000 I SHARDING [conn61] MigrateFromStatus::done About to acquire global lock to exit critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.372+0000 d20013| 2015-11-23T18:45:03.371+0000 I SHARDING [conn61] doing delete inline for cleanup of chunk data [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.373+0000 d20013| 2015-11-23T18:45:03.371+0000 I SHARDING [conn61] Deleter starting delete for: db6.coll6 from { _id: MinKey } -> { _id: 0 }, with opId: 11861 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.373+0000 d20013| 2015-11-23T18:45:03.371+0000 I SHARDING [conn61] rangeDeleter deleted 0 documents for db6.coll6 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.374+0000 d20013| 2015-11-23T18:45:03.371+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.371+0000-56535eaf6202d0bae2546318", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304303371), what: "moveChunk.from", ns: "db6.coll6", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 6: 0, step 2 of 6: 28, step 3 of 6: 1, step 4 of 6: 64, step 5 of 6: 20, step 6 of 6: 0, to: "test-rs0", from: "test-rs1", note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.399+0000 d20013| 2015-11-23T18:45:03.398+0000 I SHARDING [conn61] distributed lock 'db6.coll6/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.401+0000 d20013| 2015-11-23T18:45:03.398+0000 I COMMAND [conn61] command db6.coll6 command: moveChunk { moveChunk: "db6.coll6", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eaf6e3a936d394e6cd9') ], epoch: ObjectId('56535eaf6e3a936d394e6cd9') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:154 locks:{ Global: { acquireCount: { r: 11, w: 3, R: 2 } }, Database: { acquireCount: { r: 3, w: 3 } }, Collection: { acquireCount: { r: 3, w: 1, W: 2 } } } protocol:op_command 142ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.402+0000 s20019| 2015-11-23T18:45:03.399+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db6.coll6: 0ms sequenceNumber: 29 version: 2|1||56535eaf6e3a936d394e6cd9 based on: 1|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.403+0000 d20010| 2015-11-23T18:45:03.399+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db6.coll6", keyPattern: { _id: "hashed" }, min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs0", splitKeys: [ { _id: -4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eaf6e3a936d394e6cd9') ], epoch: ObjectId('56535eaf6e3a936d394e6cd9') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.413+0000 d20010| 2015-11-23T18:45:03.412+0000 I SHARDING [conn26] distributed lock 'db6.coll6/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ _id: MinKey }, { _id: 0 }) in db6.coll6', ts : 56535eafa5e5109e4dbf13e7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.413+0000 d20010| 2015-11-23T18:45:03.412+0000 I SHARDING [conn26] remotely refreshing metadata for db6.coll6 based on current shard version 0|0||56535eaf6e3a936d394e6cd9, current metadata version is 1|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.414+0000 d20010| 2015-11-23T18:45:03.412+0000 I SHARDING [conn26] updating metadata for db6.coll6 from shard version 0|0||56535eaf6e3a936d394e6cd9 to shard version 2|0||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.414+0000 d20010| 2015-11-23T18:45:03.412+0000 I SHARDING [conn26] collection version was loaded at version 2|1||56535eaf6e3a936d394e6cd9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.414+0000 d20010| 2015-11-23T18:45:03.412+0000 I SHARDING [conn26] splitChunk accepted at version 2|0||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.418+0000 d20010| 2015-11-23T18:45:03.416+0000 I SHARDING [conn26] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.416+0000-56535eafa5e5109e4dbf13e8", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38802", time: new Date(1448304303416), what: "split", ns: "db6.coll6", details: { before: { min: { _id: MinKey }, max: { _id: 0 } }, left: { min: { _id: MinKey }, max: { _id: -4611686018427387902 }, lastmod: Timestamp 2000|2, lastmodEpoch: ObjectId('56535eaf6e3a936d394e6cd9') }, right: { min: { _id: -4611686018427387902 }, max: { _id: 0 }, lastmod: Timestamp 2000|3, lastmodEpoch: ObjectId('56535eaf6e3a936d394e6cd9') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.424+0000 d20010| 2015-11-23T18:45:03.424+0000 I SHARDING [conn26] distributed lock 'db6.coll6/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.425+0000 s20019| 2015-11-23T18:45:03.424+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db6.coll6: 0ms sequenceNumber: 30 version: 2|3||56535eaf6e3a936d394e6cd9 based on: 2|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.426+0000 d20013| 2015-11-23T18:45:03.424+0000 I SHARDING [conn61] received splitChunk request: { splitChunk: "db6.coll6", keyPattern: { _id: "hashed" }, min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs1", splitKeys: [ { _id: 4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eaf6e3a936d394e6cd9') ], epoch: ObjectId('56535eaf6e3a936d394e6cd9') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.433+0000 d20013| 2015-11-23T18:45:03.433+0000 I SHARDING [conn61] distributed lock 'db6.coll6/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ _id: 0 }, { _id: MaxKey }) in db6.coll6', ts : 56535eaf6202d0bae2546319 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.434+0000 d20013| 2015-11-23T18:45:03.433+0000 I SHARDING [conn61] remotely refreshing metadata for db6.coll6 based on current shard version 2|0||56535eaf6e3a936d394e6cd9, current metadata version is 2|0||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.434+0000 d20013| 2015-11-23T18:45:03.433+0000 I SHARDING [conn61] updating metadata for db6.coll6 from shard version 2|0||56535eaf6e3a936d394e6cd9 to shard version 2|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.435+0000 d20013| 2015-11-23T18:45:03.433+0000 I SHARDING [conn61] collection version was loaded at version 2|3||56535eaf6e3a936d394e6cd9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.435+0000 d20013| 2015-11-23T18:45:03.433+0000 I SHARDING [conn61] splitChunk accepted at version 2|1||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.439+0000 d20013| 2015-11-23T18:45:03.438+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:03.438+0000-56535eaf6202d0bae254631a", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304303438), what: "split", ns: "db6.coll6", details: { before: { min: { _id: 0 }, max: { _id: MaxKey } }, left: { min: { _id: 0 }, max: { _id: 4611686018427387902 }, lastmod: Timestamp 2000|4, lastmodEpoch: ObjectId('56535eaf6e3a936d394e6cd9') }, right: { min: { _id: 4611686018427387902 }, max: { _id: MaxKey }, lastmod: Timestamp 2000|5, lastmodEpoch: ObjectId('56535eaf6e3a936d394e6cd9') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.447+0000 d20013| 2015-11-23T18:45:03.447+0000 I SHARDING [conn61] distributed lock 'db6.coll6/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.448+0000 s20019| 2015-11-23T18:45:03.447+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db6.coll6: 0ms sequenceNumber: 31 version: 2|5||56535eaf6e3a936d394e6cd9 based on: 2|3||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.448+0000 Using 20 threads (requested 20) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.451+0000 d20010| 2015-11-23T18:45:03.450+0000 I INDEX [conn26] build index on: db6.coll6 properties: { v: 1, key: { indexed_insert_2d: "2d" }, name: "indexed_insert_2d_2d", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.451+0000 d20010| 2015-11-23T18:45:03.450+0000 I INDEX [conn26] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.452+0000 d20013| 2015-11-23T18:45:03.450+0000 I INDEX [conn61] build index on: db6.coll6 properties: { v: 1, key: { indexed_insert_2d: "2d" }, name: "indexed_insert_2d_2d", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.452+0000 d20013| 2015-11-23T18:45:03.451+0000 I INDEX [conn61] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.452+0000 d20010| 2015-11-23T18:45:03.451+0000 I INDEX [conn26] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.453+0000 d20013| 2015-11-23T18:45:03.453+0000 I INDEX [conn61] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.454+0000 d20012| 2015-11-23T18:45:03.453+0000 I INDEX [repl writer worker 15] build index on: db6.coll6 properties: { v: 1, key: { indexed_insert_2d: "2d" }, name: "indexed_insert_2d_2d", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.455+0000 d20012| 2015-11-23T18:45:03.454+0000 I INDEX [repl writer worker 15] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.456+0000 d20011| 2015-11-23T18:45:03.453+0000 I INDEX [repl writer worker 7] build index on: db6.coll6 properties: { v: 1, key: { indexed_insert_2d: "2d" }, name: "indexed_insert_2d_2d", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.456+0000 d20011| 2015-11-23T18:45:03.453+0000 I INDEX [repl writer worker 7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.456+0000 d20011| 2015-11-23T18:45:03.454+0000 I INDEX [repl writer worker 7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.457+0000 d20012| 2015-11-23T18:45:03.455+0000 I INDEX [repl writer worker 15] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.461+0000 d20014| 2015-11-23T18:45:03.459+0000 I INDEX [repl writer worker 0] build index on: db6.coll6 properties: { v: 1, key: { indexed_insert_2d: "2d" }, name: "indexed_insert_2d_2d", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.461+0000 d20014| 2015-11-23T18:45:03.459+0000 I INDEX [repl writer worker 0] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.461+0000 d20014| 2015-11-23T18:45:03.460+0000 I INDEX [repl writer worker 0] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.462+0000 d20015| 2015-11-23T18:45:03.461+0000 I INDEX [repl writer worker 13] build index on: db6.coll6 properties: { v: 1, key: { indexed_insert_2d: "2d" }, name: "indexed_insert_2d_2d", ns: "db6.coll6" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.463+0000 d20015| 2015-11-23T18:45:03.461+0000 I INDEX [repl writer worker 13] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.463+0000 d20015| 2015-11-23T18:45:03.462+0000 I INDEX [repl writer worker 13] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.781+0000 s20020| 2015-11-23T18:45:03.780+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60517 #34 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.796+0000 s20019| 2015-11-23T18:45:03.795+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40106 #35 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.812+0000 s20019| 2015-11-23T18:45:03.812+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40107 #36 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.828+0000 s20020| 2015-11-23T18:45:03.828+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60520 #35 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.841+0000 s20020| 2015-11-23T18:45:03.840+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60521 #36 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.875+0000 s20020| 2015-11-23T18:45:03.865+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60522 #37 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.941+0000 s20019| 2015-11-23T18:45:03.874+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40111 #37 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.942+0000 s20020| 2015-11-23T18:45:03.874+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60524 #38 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.942+0000 s20020| 2015-11-23T18:45:03.875+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60525 #39 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.943+0000 s20019| 2015-11-23T18:45:03.875+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40114 #38 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.943+0000 s20019| 2015-11-23T18:45:03.876+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40115 #39 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.944+0000 s20019| 2015-11-23T18:45:03.884+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40116 #40 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.944+0000 s20020| 2015-11-23T18:45:03.889+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60529 #40 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.944+0000 s20020| 2015-11-23T18:45:03.891+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60530 #41 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.945+0000 s20019| 2015-11-23T18:45:03.893+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40120 #41 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.962+0000 s20020| 2015-11-23T18:45:03.897+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60533 #42 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.964+0000 s20019| 2015-11-23T18:45:03.909+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40122 #42 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.965+0000 s20020| 2015-11-23T18:45:03.915+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60535 #43 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.968+0000 s20019| 2015-11-23T18:45:03.924+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40124 #43 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.975+0000 s20019| 2015-11-23T18:45:03.927+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40125 #44 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.977+0000 setting random seed: 1052972210 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.977+0000 setting random seed: 1338667538 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.978+0000 setting random seed: 476293730 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:03.999+0000 setting random seed: 4286790510 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.015+0000 setting random seed: 3350303446 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.016+0000 setting random seed: 4291957576 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.017+0000 setting random seed: 1676764808 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.017+0000 setting random seed: 4103190137 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.037+0000 setting random seed: 2637075877 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.055+0000 s20020| 2015-11-23T18:45:03.935+0000 I SHARDING [conn36] ChunkManager: time to load chunks for db6.coll6: 0ms sequenceNumber: 8 version: 2|5||56535eaf6e3a936d394e6cd9 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.070+0000 setting random seed: 3524490753 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.070+0000 setting random seed: 1594039429 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.097+0000 setting random seed: 100044480 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.097+0000 setting random seed: 3722000645 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.097+0000 setting random seed: 112167580 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.113+0000 setting random seed: 1207127880 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.113+0000 setting random seed: 896793177 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.113+0000 setting random seed: 2781971983 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.134+0000 setting random seed: 2288370046 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.135+0000 setting random seed: 16072785 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.135+0000 setting random seed: 3985632558 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.554+0000 d20010| 2015-11-23T18:45:04.553+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39072 #72 (66 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.589+0000 s20020| 2015-11-23T18:45:04.589+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.615+0000 d20010| 2015-11-23T18:45:04.614+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39073 #73 (67 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.624+0000 d20010| 2015-11-23T18:45:04.623+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39074 #74 (68 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.653+0000 d20010| 2015-11-23T18:45:04.653+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39075 #75 (69 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.663+0000 d20010| 2015-11-23T18:45:04.662+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39076 #76 (70 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.668+0000 s20019| 2015-11-23T18:45:04.668+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.671+0000 s20019| 2015-11-23T18:45:04.671+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.702+0000 s20019| 2015-11-23T18:45:04.701+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:45:04.692+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20019:1448304274:1804289383', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.706+0000 s20019| 2015-11-23T18:45:04.705+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.713+0000 s20019| 2015-11-23T18:45:04.713+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-2-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.724+0000 d20013| 2015-11-23T18:45:04.723+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36659 #70 (64 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.779+0000 s20020| 2015-11-23T18:45:04.778+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-1-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.886+0000 s20020| 2015-11-23T18:45:04.885+0000 I NETWORK [conn42] end connection 10.99.163.247:60533 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.896+0000 s20019| 2015-11-23T18:45:04.895+0000 I NETWORK [conn38] end connection 10.99.163.247:40114 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.902+0000 s20020| 2015-11-23T18:45:04.902+0000 I NETWORK [conn38] end connection 10.99.163.247:60524 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.914+0000 s20019| 2015-11-23T18:45:04.914+0000 I NETWORK [conn35] end connection 10.99.163.247:40106 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.916+0000 s20019| 2015-11-23T18:45:04.916+0000 I NETWORK [conn39] end connection 10.99.163.247:40115 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.928+0000 s20020| 2015-11-23T18:45:04.927+0000 I NETWORK [conn40] end connection 10.99.163.247:60529 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.983+0000 s20019| 2015-11-23T18:45:04.982+0000 I NETWORK [conn36] end connection 10.99.163.247:40107 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.986+0000 s20019| 2015-11-23T18:45:04.985+0000 I NETWORK [conn37] end connection 10.99.163.247:40111 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:04.991+0000 s20019| 2015-11-23T18:45:04.991+0000 I NETWORK [conn41] end connection 10.99.163.247:40120 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.000+0000 s20020| 2015-11-23T18:45:05.000+0000 I NETWORK [conn35] end connection 10.99.163.247:60520 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.020+0000 s20020| 2015-11-23T18:45:05.019+0000 I NETWORK [conn41] end connection 10.99.163.247:60530 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.031+0000 s20020| 2015-11-23T18:45:05.031+0000 I NETWORK [conn43] end connection 10.99.163.247:60535 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.062+0000 s20019| 2015-11-23T18:45:05.061+0000 I NETWORK [conn40] end connection 10.99.163.247:40116 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.064+0000 s20019| 2015-11-23T18:45:05.063+0000 I NETWORK [conn42] end connection 10.99.163.247:40122 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.075+0000 s20020| 2015-11-23T18:45:05.074+0000 I NETWORK [conn34] end connection 10.99.163.247:60517 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.092+0000 s20019| 2015-11-23T18:45:05.091+0000 I NETWORK [conn43] end connection 10.99.163.247:40124 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.093+0000 s20019| 2015-11-23T18:45:05.093+0000 I NETWORK [conn44] end connection 10.99.163.247:40125 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.097+0000 s20020| 2015-11-23T18:45:05.097+0000 I NETWORK [conn36] end connection 10.99.163.247:60521 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.098+0000 s20020| 2015-11-23T18:45:05.098+0000 I NETWORK [conn37] end connection 10.99.163.247:60522 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.100+0000 s20020| 2015-11-23T18:45:05.099+0000 I NETWORK [conn39] end connection 10.99.163.247:60525 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.103+0000 s20020| 2015-11-23T18:45:05.103+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:45:05.075+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20020:1448304275:1804289383', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 Workload(s) completed in 1656 ms: jstests/concurrency/fsm_workloads/indexed_insert_2d.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.105+0000 s20019| 2015-11-23T18:45:05.105+0000 I COMMAND [conn1] DROP: db6.coll6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.106+0000 s20019| 2015-11-23T18:45:05.105+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.105+0000-56535eb16e3a936d394e6cdb", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304305105), what: "dropCollection.start", ns: "db6.coll6", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.135+0000 s20019| 2015-11-23T18:45:05.135+0000 I SHARDING [conn1] distributed lock 'db6.coll6/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535eb16e3a936d394e6cdc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.135+0000 d20010| 2015-11-23T18:45:05.135+0000 I COMMAND [conn46] CMD: drop db6.coll6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.136+0000 d20013| 2015-11-23T18:45:05.136+0000 I COMMAND [conn11] CMD: drop db6.coll6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.137+0000 d20012| 2015-11-23T18:45:05.136+0000 I COMMAND [repl writer worker 4] CMD: drop db6.coll6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.138+0000 d20011| 2015-11-23T18:45:05.136+0000 I COMMAND [repl writer worker 7] CMD: drop db6.coll6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.140+0000 d20015| 2015-11-23T18:45:05.137+0000 I COMMAND [repl writer worker 7] CMD: drop db6.coll6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.141+0000 d20014| 2015-11-23T18:45:05.138+0000 I COMMAND [repl writer worker 9] CMD: drop db6.coll6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.151+0000 d20010| 2015-11-23T18:45:05.149+0000 I SHARDING [conn46] remotely refreshing metadata for db6.coll6 with requested shard version 0|0||000000000000000000000000, current shard version is 2|3||56535eaf6e3a936d394e6cd9, current metadata version is 2|3||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.151+0000 d20010| 2015-11-23T18:45:05.150+0000 I SHARDING [conn46] dropping metadata for db6.coll6 at shard version 2|3||56535eaf6e3a936d394e6cd9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.152+0000 d20013| 2015-11-23T18:45:05.150+0000 I SHARDING [conn11] remotely refreshing metadata for db6.coll6 with requested shard version 0|0||000000000000000000000000, current shard version is 2|5||56535eaf6e3a936d394e6cd9, current metadata version is 2|5||56535eaf6e3a936d394e6cd9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.152+0000 d20013| 2015-11-23T18:45:05.150+0000 I SHARDING [conn11] dropping metadata for db6.coll6 at shard version 2|5||56535eaf6e3a936d394e6cd9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.153+0000 s20019| 2015-11-23T18:45:05.151+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.151+0000-56535eb16e3a936d394e6cdd", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304305151), what: "dropCollection", ns: "db6.coll6", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.182+0000 s20019| 2015-11-23T18:45:05.181+0000 I SHARDING [conn1] distributed lock 'db6.coll6/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.183+0000 s20019| 2015-11-23T18:45:05.182+0000 I COMMAND [conn1] DROP DATABASE: db6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.184+0000 s20019| 2015-11-23T18:45:05.182+0000 I SHARDING [conn1] DBConfig::dropDatabase: db6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.184+0000 s20019| 2015-11-23T18:45:05.182+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.182+0000-56535eb16e3a936d394e6cde", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304305182), what: "dropDatabase.start", ns: "db6", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.197+0000 s20019| 2015-11-23T18:45:05.197+0000 I SHARDING [conn1] DBConfig::dropDatabase: db6 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.198+0000 d20013| 2015-11-23T18:45:05.197+0000 I COMMAND [conn45] dropDatabase db6 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.198+0000 d20013| 2015-11-23T18:45:05.197+0000 I COMMAND [conn45] dropDatabase db6 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.199+0000 s20019| 2015-11-23T18:45:05.197+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.197+0000-56535eb16e3a936d394e6ce0", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304305197), what: "dropDatabase", ns: "db6", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.199+0000 d20015| 2015-11-23T18:45:05.198+0000 I COMMAND [repl writer worker 5] dropDatabase db6 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.199+0000 d20015| 2015-11-23T18:45:05.198+0000 I COMMAND [repl writer worker 5] dropDatabase db6 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.200+0000 d20014| 2015-11-23T18:45:05.198+0000 I COMMAND [repl writer worker 14] dropDatabase db6 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.200+0000 d20014| 2015-11-23T18:45:05.198+0000 I COMMAND [repl writer worker 14] dropDatabase db6 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.202+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.202+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.202+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.202+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.203+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.203+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.203+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.210+0000 s20019| 2015-11-23T18:45:05.209+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'doing balance round', ts : 56535eb16e3a936d394e6cdf [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.222+0000 s20019| 2015-11-23T18:45:05.221+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.221+0000-56535eb16e3a936d394e6ce1", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304305221), what: "balancer.round", ns: "", details: { executionTimeMillis: 53, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.233+0000 d20010| 2015-11-23T18:45:05.233+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.234+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.234+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.234+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.235+0000 ip-10-99-163-247:20010: awaitReplication completed in 32 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.235+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.235+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.235+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.235+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.236+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.236+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.236+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.236+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.236+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.236+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.236+0000 d20011| 2015-11-23T18:45:05.234+0000 I COMMAND [repl writer worker 4] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.237+0000 d20012| 2015-11-23T18:45:05.234+0000 I COMMAND [repl writer worker 10] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.245+0000 s20019| 2015-11-23T18:45:05.245+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.259+0000 d20013| 2015-11-23T18:45:05.259+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.259+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.260+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.260+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.260+0000 ip-10-99-163-247:20013: awaitReplication completed in 25 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.260+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.260+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.261+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.261+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.261+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.261+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.261+0000 Workload(s) started: jstests/concurrency/fsm_workloads/yield_id_hack.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.261+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.262+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.262+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.262+0000 d20014| 2015-11-23T18:45:05.259+0000 I COMMAND [repl writer worker 5] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.263+0000 d20015| 2015-11-23T18:45:05.260+0000 I COMMAND [repl writer worker 2] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.280+0000 s20019| 2015-11-23T18:45:05.280+0000 I SHARDING [conn1] distributed lock 'db7/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535eb16e3a936d394e6ce2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.282+0000 s20019| 2015-11-23T18:45:05.281+0000 I SHARDING [conn1] Placing [db7] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.282+0000 s20019| 2015-11-23T18:45:05.281+0000 I SHARDING [conn1] Enabling sharding for database [db7] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.292+0000 s20019| 2015-11-23T18:45:05.292+0000 I SHARDING [conn1] distributed lock 'db7/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.302+0000 d20013| 2015-11-23T18:45:05.301+0000 I INDEX [conn65] build index on: db7.coll7 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db7.coll7" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.303+0000 d20013| 2015-11-23T18:45:05.301+0000 I INDEX [conn65] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.303+0000 d20013| 2015-11-23T18:45:05.303+0000 I INDEX [conn65] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.304+0000 s20019| 2015-11-23T18:45:05.303+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db7.coll7", key: { _id: "hashed" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.308+0000 d20015| 2015-11-23T18:45:05.306+0000 I INDEX [repl writer worker 10] build index on: db7.coll7 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db7.coll7" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.308+0000 d20015| 2015-11-23T18:45:05.306+0000 I INDEX [repl writer worker 10] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.309+0000 d20015| 2015-11-23T18:45:05.307+0000 I INDEX [repl writer worker 10] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.309+0000 d20014| 2015-11-23T18:45:05.308+0000 I INDEX [repl writer worker 0] build index on: db7.coll7 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db7.coll7" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.309+0000 d20014| 2015-11-23T18:45:05.308+0000 I INDEX [repl writer worker 0] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.310+0000 d20014| 2015-11-23T18:45:05.310+0000 I INDEX [repl writer worker 0] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.317+0000 s20020| 2015-11-23T18:45:05.316+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' acquired for 'doing balance round', ts : 56535eb113c05a55d8fa6f94 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.321+0000 s20019| 2015-11-23T18:45:05.320+0000 I SHARDING [conn1] distributed lock 'db7.coll7/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535eb16e3a936d394e6ce3 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.321+0000 s20019| 2015-11-23T18:45:05.321+0000 I SHARDING [conn1] enable sharding on: db7.coll7 with shard key: { _id: "hashed" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.323+0000 s20019| 2015-11-23T18:45:05.321+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.321+0000-56535eb16e3a936d394e6ce4", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304305321), what: "shardCollection.start", ns: "db7.coll7", details: { shardKey: { _id: "hashed" }, collection: "db7.coll7", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 2 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.323+0000 s20020| 2015-11-23T18:45:05.321+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.321+0000-56535eb113c05a55d8fa6f95", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304305321), what: "balancer.round", ns: "", details: { executionTimeMillis: 21, errorOccured: false, candidateChunks: 0, chunksMoved: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.324+0000 s20019| 2015-11-23T18:45:05.322+0000 I SHARDING [conn1] going to create 2 chunk(s) for: db7.coll7 using new epoch 56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.328+0000 s20020| 2015-11-23T18:45:05.328+0000 I SHARDING [Balancer] distributed lock 'balancer/ip-10-99-163-247:20020:1448304275:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.331+0000 s20019| 2015-11-23T18:45:05.330+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db7.coll7: 0ms sequenceNumber: 32 version: 1|1||56535eb16e3a936d394e6ce5 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.335+0000 s20019| 2015-11-23T18:45:05.335+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db7.coll7: 0ms sequenceNumber: 33 version: 1|1||56535eb16e3a936d394e6ce5 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.336+0000 d20013| 2015-11-23T18:45:05.335+0000 I SHARDING [conn61] remotely refreshing metadata for db7.coll7 with requested shard version 1|1||56535eb16e3a936d394e6ce5, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.337+0000 d20013| 2015-11-23T18:45:05.336+0000 I SHARDING [conn61] collection db7.coll7 was previously unsharded, new metadata loaded with shard version 1|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.337+0000 d20013| 2015-11-23T18:45:05.336+0000 I SHARDING [conn61] collection version was loaded at version 1|1||56535eb16e3a936d394e6ce5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.338+0000 s20019| 2015-11-23T18:45:05.336+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.336+0000-56535eb16e3a936d394e6ce6", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304305336), what: "shardCollection.end", ns: "db7.coll7", details: { version: "1|1||56535eb16e3a936d394e6ce5" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.361+0000 s20019| 2015-11-23T18:45:05.361+0000 I SHARDING [conn1] distributed lock 'db7.coll7/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.363+0000 s20019| 2015-11-23T18:45:05.361+0000 I SHARDING [conn1] moving chunk ns: db7.coll7 moving ( ns: db7.coll7, shard: test-rs1, lastmod: 1|0||56535eb16e3a936d394e6ce5, min: { _id: MinKey }, max: { _id: 0 }) test-rs1 -> test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.363+0000 d20013| 2015-11-23T18:45:05.361+0000 I SHARDING [conn61] moveChunk waiting for full cleanup after move [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.364+0000 d20013| 2015-11-23T18:45:05.361+0000 I SHARDING [conn61] received moveChunk request: { moveChunk: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eb16e3a936d394e6ce5') ], epoch: ObjectId('56535eb16e3a936d394e6ce5') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.387+0000 d20013| 2015-11-23T18:45:05.386+0000 I SHARDING [conn61] distributed lock 'db7.coll7/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'migrating chunk [{ _id: MinKey }, { _id: 0 }) in db7.coll7', ts : 56535eb16202d0bae254631c [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.388+0000 d20013| 2015-11-23T18:45:05.386+0000 I SHARDING [conn61] remotely refreshing metadata for db7.coll7 based on current shard version 1|1||56535eb16e3a936d394e6ce5, current metadata version is 1|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.389+0000 d20013| 2015-11-23T18:45:05.386+0000 I SHARDING [conn61] metadata of collection db7.coll7 already up to date (shard version : 1|1||56535eb16e3a936d394e6ce5, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.389+0000 d20013| 2015-11-23T18:45:05.386+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.386+0000-56535eb16202d0bae254631d", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304305386), what: "moveChunk.start", ns: "db7.coll7", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.390+0000 d20013| 2015-11-23T18:45:05.387+0000 I SHARDING [conn61] moveChunk request accepted at version 1|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.390+0000 d20013| 2015-11-23T18:45:05.388+0000 I SHARDING [conn61] moveChunk number of documents: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.391+0000 d20010| 2015-11-23T18:45:05.388+0000 I SHARDING [conn67] remotely refreshing metadata for db7.coll7, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.391+0000 d20010| 2015-11-23T18:45:05.388+0000 I SHARDING [conn67] collection db7.coll7 was previously unsharded, new metadata loaded with shard version 0|0||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.391+0000 d20010| 2015-11-23T18:45:05.388+0000 I SHARDING [conn67] collection version was loaded at version 1|1||56535eb16e3a936d394e6ce5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.392+0000 d20010| 2015-11-23T18:45:05.388+0000 I SHARDING [migrateThread] starting receiving-end of migration of chunk { _id: MinKey } -> { _id: 0 } for collection db7.coll7 from test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 at epoch 56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.392+0000 d20013| 2015-11-23T18:45:05.390+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.393+0000 d20013| 2015-11-23T18:45:05.392+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.397+0000 d20013| 2015-11-23T18:45:05.396+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.405+0000 d20013| 2015-11-23T18:45:05.404+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.421+0000 d20013| 2015-11-23T18:45:05.421+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.447+0000 d20010| 2015-11-23T18:45:05.447+0000 I INDEX [migrateThread] build index on: db7.coll7 properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "db7.coll7" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.448+0000 d20010| 2015-11-23T18:45:05.448+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.454+0000 d20013| 2015-11-23T18:45:05.453+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.459+0000 d20010| 2015-11-23T18:45:05.459+0000 I INDEX [migrateThread] build index on: db7.coll7 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db7.coll7" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.460+0000 d20010| 2015-11-23T18:45:05.459+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.463+0000 d20010| 2015-11-23T18:45:05.463+0000 I INDEX [migrateThread] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.464+0000 d20010| 2015-11-23T18:45:05.463+0000 I SHARDING [migrateThread] Deleter starting delete for: db7.coll7 from { _id: MinKey } -> { _id: 0 }, with opId: 46260 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.464+0000 d20010| 2015-11-23T18:45:05.464+0000 I SHARDING [migrateThread] rangeDeleter deleted 0 documents for db7.coll7 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.472+0000 d20012| 2015-11-23T18:45:05.471+0000 I INDEX [repl writer worker 14] build index on: db7.coll7 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db7.coll7" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.473+0000 d20012| 2015-11-23T18:45:05.471+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.474+0000 d20012| 2015-11-23T18:45:05.474+0000 I INDEX [repl writer worker 14] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.479+0000 d20011| 2015-11-23T18:45:05.478+0000 I INDEX [repl writer worker 10] build index on: db7.coll7 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db7.coll7" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.479+0000 d20011| 2015-11-23T18:45:05.478+0000 I INDEX [repl writer worker 10] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.481+0000 d20011| 2015-11-23T18:45:05.480+0000 I INDEX [repl writer worker 10] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.481+0000 d20010| 2015-11-23T18:45:05.480+0000 I SHARDING [migrateThread] Waiting for replication to catch up before entering critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.481+0000 d20010| 2015-11-23T18:45:05.481+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db7.coll7' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.518+0000 d20013| 2015-11-23T18:45:05.517+0000 I SHARDING [conn61] moveChunk data transfer progress: { active: true, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "steady", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.518+0000 d20013| 2015-11-23T18:45:05.517+0000 I SHARDING [conn61] About to check if it is safe to enter critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.519+0000 d20013| 2015-11-23T18:45:05.517+0000 I SHARDING [conn61] About to enter migrate critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.519+0000 d20013| 2015-11-23T18:45:05.518+0000 I SHARDING [conn61] moveChunk setting version to: 2|0||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.522+0000 d20010| 2015-11-23T18:45:05.522+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db7.coll7' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.523+0000 d20010| 2015-11-23T18:45:05.522+0000 I SHARDING [migrateThread] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.522+0000-56535eb1a5e5109e4dbf13e9", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304305522), what: "moveChunk.to", ns: "db7.coll7", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 5: 74, step 2 of 5: 17, step 3 of 5: 0, step 4 of 5: 0, step 5 of 5: 41, note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.524+0000 d20013| 2015-11-23T18:45:05.523+0000 I SHARDING [conn61] moveChunk migrate commit accepted by TO-shard: { active: false, ns: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "done", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.524+0000 d20013| 2015-11-23T18:45:05.523+0000 I SHARDING [conn61] moveChunk updating self version to: 2|1||56535eb16e3a936d394e6ce5 through { _id: 0 } -> { _id: MaxKey } for collection 'db7.coll7' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.544+0000 d20013| 2015-11-23T18:45:05.542+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.542+0000-56535eb16202d0bae254631e", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304305542), what: "moveChunk.commit", ns: "db7.coll7", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0", cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.544+0000 d20013| 2015-11-23T18:45:05.543+0000 I SHARDING [conn61] MigrateFromStatus::done About to acquire global lock to exit critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.544+0000 d20013| 2015-11-23T18:45:05.543+0000 I SHARDING [conn61] doing delete inline for cleanup of chunk data [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.545+0000 d20013| 2015-11-23T18:45:05.543+0000 I SHARDING [conn61] Deleter starting delete for: db7.coll7 from { _id: MinKey } -> { _id: 0 }, with opId: 13939 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.545+0000 d20013| 2015-11-23T18:45:05.543+0000 I SHARDING [conn61] rangeDeleter deleted 0 documents for db7.coll7 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.546+0000 d20013| 2015-11-23T18:45:05.543+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.543+0000-56535eb16202d0bae254631f", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304305543), what: "moveChunk.from", ns: "db7.coll7", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 6: 0, step 2 of 6: 25, step 3 of 6: 1, step 4 of 6: 128, step 5 of 6: 25, step 6 of 6: 0, to: "test-rs0", from: "test-rs1", note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.556+0000 d20013| 2015-11-23T18:45:05.556+0000 I SHARDING [conn61] distributed lock 'db7.coll7/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.559+0000 d20013| 2015-11-23T18:45:05.556+0000 I COMMAND [conn61] command db7.coll7 command: moveChunk { moveChunk: "db7.coll7", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eb16e3a936d394e6ce5') ], epoch: ObjectId('56535eb16e3a936d394e6ce5') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:154 locks:{ Global: { acquireCount: { r: 11, w: 3, R: 2 } }, Database: { acquireCount: { r: 3, w: 3 } }, Collection: { acquireCount: { r: 3, w: 1, W: 2 } } } protocol:op_command 194ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.560+0000 s20019| 2015-11-23T18:45:05.557+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db7.coll7: 0ms sequenceNumber: 34 version: 2|1||56535eb16e3a936d394e6ce5 based on: 1|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.561+0000 d20010| 2015-11-23T18:45:05.557+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db7.coll7", keyPattern: { _id: "hashed" }, min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs0", splitKeys: [ { _id: -4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eb16e3a936d394e6ce5') ], epoch: ObjectId('56535eb16e3a936d394e6ce5') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.569+0000 d20010| 2015-11-23T18:45:05.568+0000 I SHARDING [conn26] distributed lock 'db7.coll7/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ _id: MinKey }, { _id: 0 }) in db7.coll7', ts : 56535eb1a5e5109e4dbf13ea [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.570+0000 d20010| 2015-11-23T18:45:05.568+0000 I SHARDING [conn26] remotely refreshing metadata for db7.coll7 based on current shard version 0|0||56535eb16e3a936d394e6ce5, current metadata version is 1|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.570+0000 d20010| 2015-11-23T18:45:05.568+0000 I SHARDING [conn26] updating metadata for db7.coll7 from shard version 0|0||56535eb16e3a936d394e6ce5 to shard version 2|0||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.571+0000 d20010| 2015-11-23T18:45:05.568+0000 I SHARDING [conn26] collection version was loaded at version 2|1||56535eb16e3a936d394e6ce5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.571+0000 d20010| 2015-11-23T18:45:05.568+0000 I SHARDING [conn26] splitChunk accepted at version 2|0||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.575+0000 d20010| 2015-11-23T18:45:05.573+0000 I SHARDING [conn26] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.573+0000-56535eb1a5e5109e4dbf13eb", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38802", time: new Date(1448304305573), what: "split", ns: "db7.coll7", details: { before: { min: { _id: MinKey }, max: { _id: 0 } }, left: { min: { _id: MinKey }, max: { _id: -4611686018427387902 }, lastmod: Timestamp 2000|2, lastmodEpoch: ObjectId('56535eb16e3a936d394e6ce5') }, right: { min: { _id: -4611686018427387902 }, max: { _id: 0 }, lastmod: Timestamp 2000|3, lastmodEpoch: ObjectId('56535eb16e3a936d394e6ce5') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.583+0000 d20010| 2015-11-23T18:45:05.583+0000 I SHARDING [conn26] distributed lock 'db7.coll7/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.584+0000 s20019| 2015-11-23T18:45:05.584+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db7.coll7: 0ms sequenceNumber: 35 version: 2|3||56535eb16e3a936d394e6ce5 based on: 2|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.586+0000 d20013| 2015-11-23T18:45:05.584+0000 I SHARDING [conn61] received splitChunk request: { splitChunk: "db7.coll7", keyPattern: { _id: "hashed" }, min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs1", splitKeys: [ { _id: 4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eb16e3a936d394e6ce5') ], epoch: ObjectId('56535eb16e3a936d394e6ce5') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.593+0000 d20013| 2015-11-23T18:45:05.592+0000 I SHARDING [conn61] distributed lock 'db7.coll7/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ _id: 0 }, { _id: MaxKey }) in db7.coll7', ts : 56535eb16202d0bae2546320 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.594+0000 d20013| 2015-11-23T18:45:05.592+0000 I SHARDING [conn61] remotely refreshing metadata for db7.coll7 based on current shard version 2|0||56535eb16e3a936d394e6ce5, current metadata version is 2|0||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.594+0000 d20013| 2015-11-23T18:45:05.593+0000 I SHARDING [conn61] updating metadata for db7.coll7 from shard version 2|0||56535eb16e3a936d394e6ce5 to shard version 2|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.595+0000 d20013| 2015-11-23T18:45:05.593+0000 I SHARDING [conn61] collection version was loaded at version 2|3||56535eb16e3a936d394e6ce5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.595+0000 d20013| 2015-11-23T18:45:05.593+0000 I SHARDING [conn61] splitChunk accepted at version 2|1||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.601+0000 d20013| 2015-11-23T18:45:05.599+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:05.599+0000-56535eb16202d0bae2546321", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304305599), what: "split", ns: "db7.coll7", details: { before: { min: { _id: 0 }, max: { _id: MaxKey } }, left: { min: { _id: 0 }, max: { _id: 4611686018427387902 }, lastmod: Timestamp 2000|4, lastmodEpoch: ObjectId('56535eb16e3a936d394e6ce5') }, right: { min: { _id: 4611686018427387902 }, max: { _id: MaxKey }, lastmod: Timestamp 2000|5, lastmodEpoch: ObjectId('56535eb16e3a936d394e6ce5') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.608+0000 d20013| 2015-11-23T18:45:05.607+0000 I SHARDING [conn61] distributed lock 'db7.coll7/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.609+0000 s20019| 2015-11-23T18:45:05.608+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db7.coll7: 0ms sequenceNumber: 36 version: 2|5||56535eb16e3a936d394e6ce5 based on: 2|3||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.609+0000 Using 5 threads (requested 5) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.738+0000 s20020| 2015-11-23T18:45:05.738+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60544 #44 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.743+0000 s20020| 2015-11-23T18:45:05.742+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60545 #45 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.746+0000 s20019| 2015-11-23T18:45:05.746+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40134 #45 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.749+0000 s20020| 2015-11-23T18:45:05.748+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60547 #46 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.757+0000 s20019| 2015-11-23T18:45:05.757+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40136 #46 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.764+0000 setting random seed: 3016646396 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.764+0000 setting random seed: 1805982035 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.764+0000 setting random seed: 2490554041 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.764+0000 setting random seed: 1891006457 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.765+0000 setting random seed: 2798847548 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:05.766+0000 s20020| 2015-11-23T18:45:05.765+0000 I SHARDING [conn44] ChunkManager: time to load chunks for db7.coll7: 0ms sequenceNumber: 9 version: 2|5||56535eb16e3a936d394e6ce5 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:06.174+0000 d20010| 2015-11-23T18:45:06.174+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39083 #77 (71 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:06.178+0000 s20020| 2015-11-23T18:45:06.178+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-0-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:06.847+0000 d20010| 2015-11-23T18:45:06.846+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:45:06.840+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20010:1448304275:1756418349', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:07.279+0000 d20013| 2015-11-23T18:45:07.278+0000 I SHARDING [LockPinger] cluster ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 pinged successfully at 2015-11-23T18:45:07.270+0000 by distributed lock pinger 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20013:1448304277:68810772', sleeping for 30000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:07.343+0000 d20010| 2015-11-23T18:45:07.342+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:39084 #78 (72 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:07.346+0000 s20020| 2015-11-23T18:45:07.346+0000 I ASIO [NetworkInterfaceASIO-TaskExecutorPool-1-1] Successfully connected to ip-10-99-163-247:20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:08.428+0000 s20019| 2015-11-23T18:45:08.427+0000 I NETWORK [conn45] end connection 10.99.163.247:40134 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:08.738+0000 s20019| 2015-11-23T18:45:08.738+0000 I NETWORK [conn46] end connection 10.99.163.247:40136 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:09.373+0000 s20020| 2015-11-23T18:45:09.372+0000 I NETWORK [conn46] end connection 10.99.163.247:60547 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:09.413+0000 s20020| 2015-11-23T18:45:09.412+0000 I NETWORK [conn45] end connection 10.99.163.247:60545 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.020+0000 s20020| 2015-11-23T18:45:10.020+0000 I NETWORK [conn44] end connection 10.99.163.247:60544 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.028+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.028+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.028+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.029+0000 Workload(s) completed in 4420 ms: jstests/concurrency/fsm_workloads/yield_id_hack.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.029+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.029+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.029+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.029+0000 s20019| 2015-11-23T18:45:10.028+0000 I COMMAND [conn1] DROP: db7.coll7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.030+0000 s20019| 2015-11-23T18:45:10.028+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.028+0000-56535eb66e3a936d394e6ce7", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310028), what: "dropCollection.start", ns: "db7.coll7", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.036+0000 s20019| 2015-11-23T18:45:10.035+0000 I SHARDING [conn1] distributed lock 'db7.coll7/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535eb66e3a936d394e6ce8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.036+0000 d20010| 2015-11-23T18:45:10.035+0000 I COMMAND [conn46] CMD: drop db7.coll7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.038+0000 d20013| 2015-11-23T18:45:10.037+0000 I COMMAND [conn11] CMD: drop db7.coll7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.039+0000 d20011| 2015-11-23T18:45:10.037+0000 I COMMAND [repl writer worker 11] CMD: drop db7.coll7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.039+0000 d20012| 2015-11-23T18:45:10.037+0000 I COMMAND [repl writer worker 1] CMD: drop db7.coll7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.040+0000 d20014| 2015-11-23T18:45:10.039+0000 I COMMAND [repl writer worker 11] CMD: drop db7.coll7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.041+0000 d20015| 2015-11-23T18:45:10.039+0000 I COMMAND [repl writer worker 13] CMD: drop db7.coll7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.044+0000 d20010| 2015-11-23T18:45:10.043+0000 I SHARDING [conn46] remotely refreshing metadata for db7.coll7 with requested shard version 0|0||000000000000000000000000, current shard version is 2|3||56535eb16e3a936d394e6ce5, current metadata version is 2|3||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.045+0000 d20010| 2015-11-23T18:45:10.044+0000 I SHARDING [conn46] dropping metadata for db7.coll7 at shard version 2|3||56535eb16e3a936d394e6ce5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.045+0000 d20013| 2015-11-23T18:45:10.044+0000 I SHARDING [conn11] remotely refreshing metadata for db7.coll7 with requested shard version 0|0||000000000000000000000000, current shard version is 2|5||56535eb16e3a936d394e6ce5, current metadata version is 2|5||56535eb16e3a936d394e6ce5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.046+0000 d20013| 2015-11-23T18:45:10.045+0000 I SHARDING [conn11] dropping metadata for db7.coll7 at shard version 2|5||56535eb16e3a936d394e6ce5, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.047+0000 s20019| 2015-11-23T18:45:10.045+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.045+0000-56535eb66e3a936d394e6ce9", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310045), what: "dropCollection", ns: "db7.coll7", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.049+0000 s20019| 2015-11-23T18:45:10.049+0000 I SHARDING [conn1] distributed lock 'db7.coll7/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.050+0000 s20019| 2015-11-23T18:45:10.050+0000 I COMMAND [conn1] DROP DATABASE: db7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.051+0000 s20019| 2015-11-23T18:45:10.050+0000 I SHARDING [conn1] DBConfig::dropDatabase: db7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.051+0000 s20019| 2015-11-23T18:45:10.050+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.050+0000-56535eb66e3a936d394e6cea", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310050), what: "dropDatabase.start", ns: "db7", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.053+0000 s20019| 2015-11-23T18:45:10.052+0000 I SHARDING [conn1] DBConfig::dropDatabase: db7 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.053+0000 d20013| 2015-11-23T18:45:10.052+0000 I COMMAND [conn45] dropDatabase db7 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.054+0000 d20013| 2015-11-23T18:45:10.052+0000 I COMMAND [conn45] dropDatabase db7 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.054+0000 s20019| 2015-11-23T18:45:10.053+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.053+0000-56535eb66e3a936d394e6ceb", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310053), what: "dropDatabase", ns: "db7", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.055+0000 d20015| 2015-11-23T18:45:10.053+0000 I COMMAND [repl writer worker 4] dropDatabase db7 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.055+0000 d20014| 2015-11-23T18:45:10.053+0000 I COMMAND [repl writer worker 1] dropDatabase db7 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.055+0000 d20014| 2015-11-23T18:45:10.053+0000 I COMMAND [repl writer worker 1] dropDatabase db7 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.055+0000 d20015| 2015-11-23T18:45:10.053+0000 I COMMAND [repl writer worker 4] dropDatabase db7 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.058+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.059+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.059+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.059+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.059+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.059+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.059+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.076+0000 d20010| 2015-11-23T18:45:10.076+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.077+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.077+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 ip-10-99-163-247:20010: awaitReplication completed in 18 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.078+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.079+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.079+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.079+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.079+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.079+0000 d20012| 2015-11-23T18:45:10.076+0000 I COMMAND [repl writer worker 0] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.080+0000 d20011| 2015-11-23T18:45:10.076+0000 I COMMAND [repl writer worker 10] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.094+0000 d20013| 2015-11-23T18:45:10.094+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.095+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.095+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.095+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.096+0000 ip-10-99-163-247:20013: awaitReplication completed in 18 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.096+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.096+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.096+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.097+0000 d20014| 2015-11-23T18:45:10.095+0000 I COMMAND [repl writer worker 2] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.097+0000 d20015| 2015-11-23T18:45:10.095+0000 I COMMAND [repl writer worker 6] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.097+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.097+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.097+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.098+0000 Workload(s) started: jstests/concurrency/fsm_workloads/explain_distinct.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.098+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.098+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.098+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.114+0000 s20019| 2015-11-23T18:45:10.113+0000 I SHARDING [conn1] distributed lock 'db8/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535eb66e3a936d394e6cec [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.115+0000 s20019| 2015-11-23T18:45:10.115+0000 I SHARDING [conn1] Placing [db8] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.115+0000 s20019| 2015-11-23T18:45:10.115+0000 I SHARDING [conn1] Enabling sharding for database [db8] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.124+0000 s20019| 2015-11-23T18:45:10.124+0000 I SHARDING [conn1] distributed lock 'db8/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.138+0000 d20013| 2015-11-23T18:45:10.137+0000 I INDEX [conn65] build index on: db8.coll8 properties: { v: 1, key: { j: 1.0 }, name: "j_1", ns: "db8.coll8" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.138+0000 d20013| 2015-11-23T18:45:10.138+0000 I INDEX [conn65] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.144+0000 d20013| 2015-11-23T18:45:10.143+0000 I INDEX [conn65] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.144+0000 s20019| 2015-11-23T18:45:10.144+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db8.coll8", key: { j: 1.0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.168+0000 d20014| 2015-11-23T18:45:10.168+0000 I INDEX [repl writer worker 14] build index on: db8.coll8 properties: { v: 1, key: { j: 1.0 }, name: "j_1", ns: "db8.coll8" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.169+0000 d20014| 2015-11-23T18:45:10.168+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.170+0000 d20015| 2015-11-23T18:45:10.168+0000 I INDEX [repl writer worker 14] build index on: db8.coll8 properties: { v: 1, key: { j: 1.0 }, name: "j_1", ns: "db8.coll8" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.170+0000 d20015| 2015-11-23T18:45:10.168+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.171+0000 d20014| 2015-11-23T18:45:10.169+0000 I INDEX [repl writer worker 14] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.171+0000 d20015| 2015-11-23T18:45:10.169+0000 I INDEX [repl writer worker 14] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.171+0000 s20019| 2015-11-23T18:45:10.169+0000 I SHARDING [conn1] distributed lock 'db8.coll8/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535eb66e3a936d394e6ced [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.172+0000 s20019| 2015-11-23T18:45:10.170+0000 I SHARDING [conn1] enable sharding on: db8.coll8 with shard key: { j: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.173+0000 s20019| 2015-11-23T18:45:10.170+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.170+0000-56535eb66e3a936d394e6cee", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310170), what: "shardCollection.start", ns: "db8.coll8", details: { shardKey: { j: 1.0 }, collection: "db8.coll8", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 1 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.173+0000 s20019| 2015-11-23T18:45:10.171+0000 I SHARDING [conn1] going to create 1 chunk(s) for: db8.coll8 using new epoch 56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.181+0000 s20019| 2015-11-23T18:45:10.181+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db8.coll8: 0ms sequenceNumber: 37 version: 1|0||56535eb66e3a936d394e6cef based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.188+0000 s20019| 2015-11-23T18:45:10.188+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db8.coll8: 0ms sequenceNumber: 38 version: 1|0||56535eb66e3a936d394e6cef based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.189+0000 d20013| 2015-11-23T18:45:10.188+0000 I SHARDING [conn61] remotely refreshing metadata for db8.coll8 with requested shard version 1|0||56535eb66e3a936d394e6cef, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.189+0000 d20013| 2015-11-23T18:45:10.189+0000 I SHARDING [conn61] collection db8.coll8 was previously unsharded, new metadata loaded with shard version 1|0||56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.190+0000 d20013| 2015-11-23T18:45:10.189+0000 I SHARDING [conn61] collection version was loaded at version 1|0||56535eb66e3a936d394e6cef, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.190+0000 s20019| 2015-11-23T18:45:10.189+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.189+0000-56535eb66e3a936d394e6cf0", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310189), what: "shardCollection.end", ns: "db8.coll8", details: { version: "1|0||56535eb66e3a936d394e6cef" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.202+0000 s20019| 2015-11-23T18:45:10.201+0000 I SHARDING [conn1] distributed lock 'db8.coll8/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.202+0000 Using 10 threads (requested 10) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.396+0000 s20020| 2015-11-23T18:45:10.395+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60551 #47 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.399+0000 s20019| 2015-11-23T18:45:10.398+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40140 #47 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.400+0000 s20019| 2015-11-23T18:45:10.400+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40141 #48 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.405+0000 s20020| 2015-11-23T18:45:10.405+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60554 #48 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.406+0000 s20020| 2015-11-23T18:45:10.405+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60555 #49 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.407+0000 s20019| 2015-11-23T18:45:10.406+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40144 #49 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.407+0000 s20020| 2015-11-23T18:45:10.406+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60557 #50 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.417+0000 s20019| 2015-11-23T18:45:10.417+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40146 #50 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.418+0000 s20020| 2015-11-23T18:45:10.417+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60559 #51 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.421+0000 s20019| 2015-11-23T18:45:10.421+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40148 #51 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.434+0000 setting random seed: 1156415595 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.434+0000 setting random seed: 3869296473 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.434+0000 setting random seed: 3237680611 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.434+0000 setting random seed: 3489103819 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.435+0000 setting random seed: 607291343 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.435+0000 setting random seed: 1537596128 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.435+0000 setting random seed: 3951134216 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.435+0000 setting random seed: 314857689 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.435+0000 setting random seed: 2959472007 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.436+0000 setting random seed: 459876963 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.437+0000 s20020| 2015-11-23T18:45:10.436+0000 I SHARDING [conn50] ChunkManager: time to load chunks for db8.coll8: 1ms sequenceNumber: 10 version: 1|0||56535eb66e3a936d394e6cef based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.459+0000 d20013| 2015-11-23T18:45:10.458+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36677 #71 (65 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.481+0000 d20013| 2015-11-23T18:45:10.480+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36678 #72 (66 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.482+0000 d20013| 2015-11-23T18:45:10.482+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36679 #73 (67 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.484+0000 d20013| 2015-11-23T18:45:10.483+0000 I SHARDING [conn11] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.485+0000 d20013| 2015-11-23T18:45:10.483+0000 I SHARDING [conn11] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.486+0000 d20013| 2015-11-23T18:45:10.484+0000 I SHARDING [conn61] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.489+0000 d20013| 2015-11-23T18:45:10.489+0000 I SHARDING [conn11] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.492+0000 d20013| 2015-11-23T18:45:10.491+0000 I SHARDING [conn66] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.494+0000 d20013| 2015-11-23T18:45:10.493+0000 I SHARDING [conn61] distributed lock 'db8.coll8/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8', ts : 56535eb66202d0bae2546323 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.494+0000 d20013| 2015-11-23T18:45:10.493+0000 W SHARDING [conn66] could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.495+0000 d20013| 2015-11-23T18:45:10.493+0000 I SHARDING [conn61] remotely refreshing metadata for db8.coll8 based on current shard version 1|0||56535eb66e3a936d394e6cef, current metadata version is 1|0||56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.497+0000 s20019| 2015-11-23T18:45:10.493+0000 W SHARDING [conn49] splitChunk failed - cmd: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.498+0000 d20013| 2015-11-23T18:45:10.497+0000 I SHARDING [conn61] metadata of collection db8.coll8 already up to date (shard version : 1|0||56535eb66e3a936d394e6cef, took 4ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.499+0000 d20013| 2015-11-23T18:45:10.497+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36680 #74 (68 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.499+0000 d20013| 2015-11-23T18:45:10.497+0000 I SHARDING [conn61] splitChunk accepted at version 1|0||56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.500+0000 d20013| 2015-11-23T18:45:10.499+0000 I SHARDING [conn67] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.505+0000 d20013| 2015-11-23T18:45:10.503+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.503+0000-56535eb66202d0bae2546325", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304310503), what: "multi-split", ns: "db8.coll8", details: { before: { min: { j: MinKey }, max: { j: MaxKey } }, number: 1, of: 3, chunk: { min: { j: MinKey }, max: { j: 0.0 }, lastmod: Timestamp 1000|1, lastmodEpoch: ObjectId('56535eb66e3a936d394e6cef') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.507+0000 d20013| 2015-11-23T18:45:10.506+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.506+0000-56535eb66202d0bae2546326", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304310506), what: "multi-split", ns: "db8.coll8", details: { before: { min: { j: MinKey }, max: { j: MaxKey } }, number: 2, of: 3, chunk: { min: { j: 0.0 }, max: { j: 4.0 }, lastmod: Timestamp 1000|2, lastmodEpoch: ObjectId('56535eb66e3a936d394e6cef') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.513+0000 d20013| 2015-11-23T18:45:10.508+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36681 #75 (69 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.514+0000 d20013| 2015-11-23T18:45:10.509+0000 I SHARDING [conn67] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.522+0000 d20013| 2015-11-23T18:45:10.509+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.529+0000 d20013| 2015-11-23T18:45:10.510+0000 I SHARDING [conn67] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.543+0000 d20013| 2015-11-23T18:45:10.511+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.511+0000-56535eb66202d0bae2546328", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304310511), what: "multi-split", ns: "db8.coll8", details: { before: { min: { j: MinKey }, max: { j: MaxKey } }, number: 3, of: 3, chunk: { min: { j: 4.0 }, max: { j: MaxKey }, lastmod: Timestamp 1000|3, lastmodEpoch: ObjectId('56535eb66e3a936d394e6cef') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.552+0000 d20013| 2015-11-23T18:45:10.511+0000 I SHARDING [conn38] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.556+0000 d20013| 2015-11-23T18:45:10.511+0000 W SHARDING [conn37] could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.563+0000 s20020| 2015-11-23T18:45:10.511+0000 W SHARDING [conn48] splitChunk failed - cmd: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.568+0000 d20013| 2015-11-23T18:45:10.512+0000 W SHARDING [conn38] could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.582+0000 s20020| 2015-11-23T18:45:10.513+0000 W SHARDING [conn50] splitChunk failed - cmd: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.588+0000 d20013| 2015-11-23T18:45:10.519+0000 I SHARDING [conn66] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.594+0000 d20013| 2015-11-23T18:45:10.520+0000 W SHARDING [conn66] could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.606+0000 s20019| 2015-11-23T18:45:10.520+0000 W SHARDING [conn48] splitChunk failed - cmd: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.609+0000 d20013| 2015-11-23T18:45:10.523+0000 I SHARDING [conn61] distributed lock 'db8.coll8/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.615+0000 s20019| 2015-11-23T18:45:10.524+0000 I SHARDING [conn47] ChunkManager: time to load chunks for db8.coll8: 0ms sequenceNumber: 39 version: 1|3||56535eb66e3a936d394e6cef based on: 1|0||56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.616+0000 s20019| 2015-11-23T18:45:10.524+0000 I SHARDING [conn47] autosplitted db8.coll8 shard: ns: db8.coll8, shard: test-rs1, lastmod: 1|0||56535eb66e3a936d394e6cef, min: { j: MinKey }, max: { j: MaxKey } into 3 (splitThreshold 921) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.621+0000 d20013| 2015-11-23T18:45:10.526+0000 I SHARDING [conn67] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.628+0000 d20013| 2015-11-23T18:45:10.526+0000 I SHARDING [conn54] request split points lookup for chunk db8.coll8 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.640+0000 d20013| 2015-11-23T18:45:10.527+0000 I SHARDING [conn38] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 8.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.644+0000 d20013| 2015-11-23T18:45:10.527+0000 I SHARDING [conn37] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 8.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.652+0000 s20020| 2015-11-23T18:45:10.532+0000 W SHARDING [conn50] splitChunk failed - cmd: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 }, { j: 8.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.653+0000 d20013| 2015-11-23T18:45:10.532+0000 I SHARDING [conn38] could not acquire lock 'db8.coll8/ip-10-99-163-247:20013:1448304277:68810772' (another update won) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.657+0000 d20013| 2015-11-23T18:45:10.532+0000 I SHARDING [conn38] distributed lock 'db8.coll8/ip-10-99-163-247:20013:1448304277:68810772' was not acquired. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.661+0000 d20013| 2015-11-23T18:45:10.532+0000 W SHARDING [conn38] could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.667+0000 d20013| 2015-11-23T18:45:10.535+0000 I SHARDING [conn75] received splitChunk request: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.670+0000 d20013| 2015-11-23T18:45:10.537+0000 I SHARDING [conn37] distributed lock 'db8.coll8/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8', ts : 56535eb66202d0bae254632c [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.671+0000 d20013| 2015-11-23T18:45:10.537+0000 I SHARDING [conn37] remotely refreshing metadata for db8.coll8 based on current shard version 1|3||56535eb66e3a936d394e6cef, current metadata version is 1|3||56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.676+0000 d20013| 2015-11-23T18:45:10.538+0000 I SHARDING [conn37] metadata of collection db8.coll8 already up to date (shard version : 1|3||56535eb66e3a936d394e6cef, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.680+0000 d20013| 2015-11-23T18:45:10.538+0000 W SHARDING [conn37] splitChunk cannot find chunk [{ j: MinKey },{ j: MaxKey }) to split, the chunk boundaries may be stale [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.686+0000 d20013| 2015-11-23T18:45:10.542+0000 I SHARDING [conn37] distributed lock 'db8.coll8/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.693+0000 s20020| 2015-11-23T18:45:10.542+0000 I SHARDING [conn48] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.700+0000 s20020| 2015-11-23T18:45:10.542+0000 W SHARDING [conn48] could not autosplit collection db8.coll8 :: caused by :: 9996 stale config in runCommand ( ns : db8.coll8, received : 1|0||56535eb66e3a936d394e6cef, wanted : 1|3||56535eb66e3a936d394e6cef, recv ) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.700+0000 d20013| 2015-11-23T18:45:10.543+0000 I NETWORK [conn37] end connection 10.99.163.247:36524 (68 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.713+0000 d20013| 2015-11-23T18:45:10.548+0000 W SHARDING [conn75] could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, { j: MaxKey }) in db8.coll8 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.736+0000 s20020| 2015-11-23T18:45:10.548+0000 W SHARDING [conn49] splitChunk failed - cmd: { splitChunk: "db8.coll8", keyPattern: { j: 1.0 }, min: { j: MinKey }, max: { j: MaxKey }, from: "test-rs1", splitKeys: [ { j: 0.0 }, { j: 4.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb66e3a936d394e6cef') ], epoch: ObjectId('56535eb66e3a936d394e6cef') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db8.coll8 to split chunk [{ j: MinKey },{ j: MaxKey }) :: caused by :: Lock for splitting chunk [{ j: MinKey }, ..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.752+0000 s20020| 2015-11-23T18:45:10.553+0000 I SHARDING [conn49] ChunkManager: time to load chunks for db8.coll8: 0ms sequenceNumber: 11 version: 1|3||56535eb66e3a936d394e6cef based on: 1|0||56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.758+0000 s20020| 2015-11-23T18:45:10.757+0000 I NETWORK [conn50] end connection 10.99.163.247:60557 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.783+0000 s20019| 2015-11-23T18:45:10.782+0000 I NETWORK [conn49] end connection 10.99.163.247:40144 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.791+0000 s20019| 2015-11-23T18:45:10.791+0000 I NETWORK [conn48] end connection 10.99.163.247:40141 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.807+0000 s20020| 2015-11-23T18:45:10.806+0000 I NETWORK [conn49] end connection 10.99.163.247:60555 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.816+0000 s20019| 2015-11-23T18:45:10.816+0000 I NETWORK [conn50] end connection 10.99.163.247:40146 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.818+0000 s20019| 2015-11-23T18:45:10.817+0000 I NETWORK [conn47] end connection 10.99.163.247:40140 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.823+0000 s20019| 2015-11-23T18:45:10.823+0000 I NETWORK [conn51] end connection 10.99.163.247:40148 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.826+0000 s20020| 2015-11-23T18:45:10.826+0000 I NETWORK [conn51] end connection 10.99.163.247:60559 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.830+0000 s20020| 2015-11-23T18:45:10.829+0000 I NETWORK [conn48] end connection 10.99.163.247:60554 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.843+0000 s20020| 2015-11-23T18:45:10.843+0000 I NETWORK [conn47] end connection 10.99.163.247:60551 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.846+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.846+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.846+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.846+0000 Workload(s) completed in 644 ms: jstests/concurrency/fsm_workloads/explain_distinct.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.847+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.847+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.847+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.847+0000 s20019| 2015-11-23T18:45:10.846+0000 I COMMAND [conn1] DROP: db8.coll8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.848+0000 s20019| 2015-11-23T18:45:10.846+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.846+0000-56535eb66e3a936d394e6cf1", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310846), what: "dropCollection.start", ns: "db8.coll8", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.879+0000 s20019| 2015-11-23T18:45:10.878+0000 I SHARDING [conn1] distributed lock 'db8.coll8/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535eb66e3a936d394e6cf2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.879+0000 d20010| 2015-11-23T18:45:10.878+0000 I COMMAND [conn46] CMD: drop db8.coll8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.880+0000 d20013| 2015-11-23T18:45:10.879+0000 I COMMAND [conn11] CMD: drop db8.coll8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.881+0000 d20015| 2015-11-23T18:45:10.880+0000 I COMMAND [repl writer worker 4] CMD: drop db8.coll8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.881+0000 d20014| 2015-11-23T18:45:10.881+0000 I COMMAND [repl writer worker 6] CMD: drop db8.coll8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.903+0000 d20013| 2015-11-23T18:45:10.902+0000 I SHARDING [conn11] remotely refreshing metadata for db8.coll8 with requested shard version 0|0||000000000000000000000000, current shard version is 1|3||56535eb66e3a936d394e6cef, current metadata version is 1|3||56535eb66e3a936d394e6cef [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.903+0000 d20013| 2015-11-23T18:45:10.902+0000 I SHARDING [conn11] dropping metadata for db8.coll8 at shard version 1|3||56535eb66e3a936d394e6cef, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.904+0000 s20019| 2015-11-23T18:45:10.902+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.902+0000-56535eb66e3a936d394e6cf3", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310902), what: "dropCollection", ns: "db8.coll8", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.924+0000 s20019| 2015-11-23T18:45:10.923+0000 I SHARDING [conn1] distributed lock 'db8.coll8/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.924+0000 s20019| 2015-11-23T18:45:10.924+0000 I COMMAND [conn1] DROP DATABASE: db8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.925+0000 s20019| 2015-11-23T18:45:10.924+0000 I SHARDING [conn1] DBConfig::dropDatabase: db8 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.925+0000 s20019| 2015-11-23T18:45:10.924+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.924+0000-56535eb66e3a936d394e6cf4", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310924), what: "dropDatabase.start", ns: "db8", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.928+0000 s20019| 2015-11-23T18:45:10.927+0000 I SHARDING [conn1] DBConfig::dropDatabase: db8 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.928+0000 d20013| 2015-11-23T18:45:10.927+0000 I COMMAND [conn45] dropDatabase db8 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.929+0000 d20013| 2015-11-23T18:45:10.927+0000 I COMMAND [conn45] dropDatabase db8 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.929+0000 s20019| 2015-11-23T18:45:10.927+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:10.927+0000-56535eb66e3a936d394e6cf5", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304310927), what: "dropDatabase", ns: "db8", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.929+0000 d20014| 2015-11-23T18:45:10.928+0000 I COMMAND [repl writer worker 13] dropDatabase db8 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.930+0000 d20014| 2015-11-23T18:45:10.928+0000 I COMMAND [repl writer worker 13] dropDatabase db8 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.930+0000 d20015| 2015-11-23T18:45:10.928+0000 I COMMAND [repl writer worker 10] dropDatabase db8 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.930+0000 d20015| 2015-11-23T18:45:10.928+0000 I COMMAND [repl writer worker 10] dropDatabase db8 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.932+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.932+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.932+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.933+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.933+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.933+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.933+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.959+0000 d20010| 2015-11-23T18:45:10.958+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.959+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.959+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.959+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 ip-10-99-163-247:20010: awaitReplication completed in 26 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.960+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.961+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.961+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.961+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.962+0000 d20012| 2015-11-23T18:45:10.959+0000 I COMMAND [repl writer worker 5] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.962+0000 d20011| 2015-11-23T18:45:10.959+0000 I COMMAND [repl writer worker 8] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.979+0000 d20013| 2015-11-23T18:45:10.978+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.979+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.979+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.979+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.980+0000 ip-10-99-163-247:20013: awaitReplication completed in 21 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.980+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.980+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.981+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.981+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.981+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.981+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.981+0000 Workload(s) started: jstests/concurrency/fsm_workloads/update_ordered_bulk_inc.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.981+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.982+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.982+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.982+0000 d20015| 2015-11-23T18:45:10.979+0000 I COMMAND [repl writer worker 15] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.982+0000 d20014| 2015-11-23T18:45:10.980+0000 I COMMAND [repl writer worker 11] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.990+0000 s20019| 2015-11-23T18:45:10.989+0000 I SHARDING [conn1] distributed lock 'db9/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535eb66e3a936d394e6cf6 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.991+0000 s20019| 2015-11-23T18:45:10.991+0000 I SHARDING [conn1] Placing [db9] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.991+0000 s20019| 2015-11-23T18:45:10.991+0000 I SHARDING [conn1] Enabling sharding for database [db9] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:10.995+0000 s20019| 2015-11-23T18:45:10.995+0000 I SHARDING [conn1] distributed lock 'db9/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.006+0000 d20013| 2015-11-23T18:45:11.003+0000 I INDEX [conn65] build index on: db9.coll9 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db9.coll9" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.007+0000 d20013| 2015-11-23T18:45:11.003+0000 I INDEX [conn65] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.007+0000 d20013| 2015-11-23T18:45:11.004+0000 I INDEX [conn65] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.007+0000 s20019| 2015-11-23T18:45:11.005+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db9.coll9", key: { _id: "hashed" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.019+0000 s20019| 2015-11-23T18:45:11.019+0000 I SHARDING [conn1] distributed lock 'db9.coll9/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535eb76e3a936d394e6cf7 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.020+0000 d20015| 2015-11-23T18:45:11.019+0000 I INDEX [repl writer worker 11] build index on: db9.coll9 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db9.coll9" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.020+0000 d20015| 2015-11-23T18:45:11.019+0000 I INDEX [repl writer worker 11] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.021+0000 d20014| 2015-11-23T18:45:11.019+0000 I INDEX [repl writer worker 9] build index on: db9.coll9 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db9.coll9" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.021+0000 d20014| 2015-11-23T18:45:11.019+0000 I INDEX [repl writer worker 9] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.021+0000 s20019| 2015-11-23T18:45:11.019+0000 I SHARDING [conn1] enable sharding on: db9.coll9 with shard key: { _id: "hashed" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.022+0000 s20019| 2015-11-23T18:45:11.019+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.019+0000-56535eb76e3a936d394e6cf8", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304311019), what: "shardCollection.start", ns: "db9.coll9", details: { shardKey: { _id: "hashed" }, collection: "db9.coll9", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 2 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.023+0000 d20015| 2015-11-23T18:45:11.020+0000 I INDEX [repl writer worker 11] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.024+0000 d20014| 2015-11-23T18:45:11.020+0000 I INDEX [repl writer worker 9] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.024+0000 s20019| 2015-11-23T18:45:11.021+0000 I SHARDING [conn1] going to create 2 chunk(s) for: db9.coll9 using new epoch 56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.027+0000 s20019| 2015-11-23T18:45:11.027+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db9.coll9: 0ms sequenceNumber: 40 version: 1|1||56535eb76e3a936d394e6cf9 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.029+0000 s20019| 2015-11-23T18:45:11.029+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db9.coll9: 0ms sequenceNumber: 41 version: 1|1||56535eb76e3a936d394e6cf9 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.030+0000 d20013| 2015-11-23T18:45:11.030+0000 I SHARDING [conn22] remotely refreshing metadata for db9.coll9 with requested shard version 1|1||56535eb76e3a936d394e6cf9, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.031+0000 d20013| 2015-11-23T18:45:11.030+0000 I SHARDING [conn22] collection db9.coll9 was previously unsharded, new metadata loaded with shard version 1|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.032+0000 d20013| 2015-11-23T18:45:11.030+0000 I SHARDING [conn22] collection version was loaded at version 1|1||56535eb76e3a936d394e6cf9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.032+0000 s20019| 2015-11-23T18:45:11.030+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.030+0000-56535eb76e3a936d394e6cfa", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304311030), what: "shardCollection.end", ns: "db9.coll9", details: { version: "1|1||56535eb76e3a936d394e6cf9" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.035+0000 s20019| 2015-11-23T18:45:11.035+0000 I SHARDING [conn1] distributed lock 'db9.coll9/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.035+0000 s20019| 2015-11-23T18:45:11.035+0000 I SHARDING [conn1] moving chunk ns: db9.coll9 moving ( ns: db9.coll9, shard: test-rs1, lastmod: 1|0||56535eb76e3a936d394e6cf9, min: { _id: MinKey }, max: { _id: 0 }) test-rs1 -> test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.036+0000 d20013| 2015-11-23T18:45:11.035+0000 I SHARDING [conn22] moveChunk waiting for full cleanup after move [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.037+0000 d20013| 2015-11-23T18:45:11.035+0000 I SHARDING [conn22] received moveChunk request: { moveChunk: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eb76e3a936d394e6cf9') ], epoch: ObjectId('56535eb76e3a936d394e6cf9') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.046+0000 d20013| 2015-11-23T18:45:11.046+0000 I SHARDING [conn22] distributed lock 'db9.coll9/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'migrating chunk [{ _id: MinKey }, { _id: 0 }) in db9.coll9', ts : 56535eb76202d0bae254632f [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.047+0000 d20013| 2015-11-23T18:45:11.046+0000 I SHARDING [conn22] remotely refreshing metadata for db9.coll9 based on current shard version 1|1||56535eb76e3a936d394e6cf9, current metadata version is 1|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.047+0000 d20013| 2015-11-23T18:45:11.046+0000 I SHARDING [conn22] metadata of collection db9.coll9 already up to date (shard version : 1|1||56535eb76e3a936d394e6cf9, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.048+0000 d20013| 2015-11-23T18:45:11.046+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.046+0000-56535eb76202d0bae2546330", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304311046), what: "moveChunk.start", ns: "db9.coll9", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.049+0000 d20010| 2015-11-23T18:45:11.048+0000 I SHARDING [conn62] remotely refreshing metadata for db9.coll9, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.049+0000 d20013| 2015-11-23T18:45:11.047+0000 I SHARDING [conn22] moveChunk request accepted at version 1|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.049+0000 d20013| 2015-11-23T18:45:11.048+0000 I SHARDING [conn22] moveChunk number of documents: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.050+0000 d20010| 2015-11-23T18:45:11.048+0000 I SHARDING [conn62] collection db9.coll9 was previously unsharded, new metadata loaded with shard version 0|0||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.051+0000 d20010| 2015-11-23T18:45:11.048+0000 I SHARDING [conn62] collection version was loaded at version 1|1||56535eb76e3a936d394e6cf9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.051+0000 d20010| 2015-11-23T18:45:11.048+0000 I SHARDING [migrateThread] starting receiving-end of migration of chunk { _id: MinKey } -> { _id: 0 } for collection db9.coll9 from test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 at epoch 56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.053+0000 d20013| 2015-11-23T18:45:11.050+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.054+0000 d20013| 2015-11-23T18:45:11.052+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.056+0000 d20010| 2015-11-23T18:45:11.055+0000 I INDEX [migrateThread] build index on: db9.coll9 properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "db9.coll9" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.056+0000 d20010| 2015-11-23T18:45:11.056+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.057+0000 d20013| 2015-11-23T18:45:11.056+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.066+0000 d20013| 2015-11-23T18:45:11.065+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.066+0000 d20010| 2015-11-23T18:45:11.065+0000 I INDEX [migrateThread] build index on: db9.coll9 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db9.coll9" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.067+0000 d20010| 2015-11-23T18:45:11.065+0000 I INDEX [migrateThread] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.071+0000 d20010| 2015-11-23T18:45:11.070+0000 I INDEX [migrateThread] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.072+0000 d20010| 2015-11-23T18:45:11.071+0000 I SHARDING [migrateThread] Deleter starting delete for: db9.coll9 from { _id: MinKey } -> { _id: 0 }, with opId: 53840 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.072+0000 d20010| 2015-11-23T18:45:11.072+0000 I SHARDING [migrateThread] rangeDeleter deleted 0 documents for db9.coll9 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.077+0000 d20012| 2015-11-23T18:45:11.076+0000 I INDEX [repl writer worker 14] build index on: db9.coll9 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db9.coll9" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.078+0000 d20012| 2015-11-23T18:45:11.076+0000 I INDEX [repl writer worker 14] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.082+0000 d20013| 2015-11-23T18:45:11.081+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "ready", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.082+0000 d20011| 2015-11-23T18:45:11.081+0000 I INDEX [repl writer worker 15] build index on: db9.coll9 properties: { v: 1, key: { _id: "hashed" }, name: "_id_hashed", ns: "db9.coll9" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.082+0000 d20011| 2015-11-23T18:45:11.081+0000 I INDEX [repl writer worker 15] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.083+0000 d20012| 2015-11-23T18:45:11.082+0000 I INDEX [repl writer worker 14] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.085+0000 d20011| 2015-11-23T18:45:11.085+0000 I INDEX [repl writer worker 15] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.086+0000 d20010| 2015-11-23T18:45:11.086+0000 I SHARDING [migrateThread] Waiting for replication to catch up before entering critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.087+0000 d20010| 2015-11-23T18:45:11.086+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db9.coll9' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.115+0000 d20013| 2015-11-23T18:45:11.114+0000 I SHARDING [conn22] moveChunk data transfer progress: { active: true, ns: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "steady", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } my mem used: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.115+0000 d20013| 2015-11-23T18:45:11.114+0000 I SHARDING [conn22] About to check if it is safe to enter critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.115+0000 d20013| 2015-11-23T18:45:11.114+0000 I SHARDING [conn22] About to enter migrate critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.115+0000 d20013| 2015-11-23T18:45:11.114+0000 I SHARDING [conn22] moveChunk setting version to: 2|0||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.118+0000 d20010| 2015-11-23T18:45:11.117+0000 I SHARDING [migrateThread] migrate commit succeeded flushing to secondaries for 'db9.coll9' { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.119+0000 d20010| 2015-11-23T18:45:11.117+0000 I SHARDING [migrateThread] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.117+0000-56535eb7a5e5109e4dbf13ec", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304311117), what: "moveChunk.to", ns: "db9.coll9", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 5: 22, step 2 of 5: 14, step 3 of 5: 0, step 4 of 5: 0, step 5 of 5: 31, note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.119+0000 d20013| 2015-11-23T18:45:11.118+0000 I SHARDING [conn22] moveChunk migrate commit accepted by TO-shard: { active: false, ns: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", min: { _id: MinKey }, max: { _id: 0 }, shardKeyPattern: { _id: "hashed" }, state: "done", counts: { cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 }, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.120+0000 d20013| 2015-11-23T18:45:11.118+0000 I SHARDING [conn22] moveChunk updating self version to: 2|1||56535eb76e3a936d394e6cf9 through { _id: 0 } -> { _id: MaxKey } for collection 'db9.coll9' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.125+0000 d20013| 2015-11-23T18:45:11.124+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.124+0000-56535eb76202d0bae2546331", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304311124), what: "moveChunk.commit", ns: "db9.coll9", details: { min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs1", to: "test-rs0", cloned: 0, clonedBytes: 0, catchup: 0, steady: 0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.126+0000 d20013| 2015-11-23T18:45:11.125+0000 I SHARDING [conn22] MigrateFromStatus::done About to acquire global lock to exit critical section [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.126+0000 d20013| 2015-11-23T18:45:11.125+0000 I SHARDING [conn22] doing delete inline for cleanup of chunk data [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.126+0000 d20013| 2015-11-23T18:45:11.125+0000 I SHARDING [conn22] Deleter starting delete for: db9.coll9 from { _id: MinKey } -> { _id: 0 }, with opId: 25525 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.127+0000 d20013| 2015-11-23T18:45:11.125+0000 I SHARDING [conn22] rangeDeleter deleted 0 documents for db9.coll9 from { _id: MinKey } -> { _id: 0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.127+0000 d20013| 2015-11-23T18:45:11.125+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.125+0000-56535eb76202d0bae2546332", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304311125), what: "moveChunk.from", ns: "db9.coll9", details: { min: { _id: MinKey }, max: { _id: 0 }, step 1 of 6: 0, step 2 of 6: 11, step 3 of 6: 1, step 4 of 6: 65, step 5 of 6: 11, step 6 of 6: 0, to: "test-rs0", from: "test-rs1", note: "success" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.141+0000 d20013| 2015-11-23T18:45:11.141+0000 I SHARDING [conn22] distributed lock 'db9.coll9/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.143+0000 d20013| 2015-11-23T18:45:11.141+0000 I COMMAND [conn22] command db9.coll9 command: moveChunk { moveChunk: "db9.coll9", from: "test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", to: "test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012", fromShard: "test-rs1", toShard: "test-rs0", min: { _id: MinKey }, max: { _id: 0 }, maxChunkSizeBytes: 52428800, configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", secondaryThrottle: false, waitForDelete: true, maxTimeMS: 0, shardVersion: [ Timestamp 1000|1, ObjectId('56535eb76e3a936d394e6cf9') ], epoch: ObjectId('56535eb76e3a936d394e6cf9') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:154 locks:{ Global: { acquireCount: { r: 11, w: 3, R: 2 } }, Database: { acquireCount: { r: 3, w: 3 } }, Collection: { acquireCount: { r: 3, w: 1, W: 2 } } } protocol:op_command 105ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.143+0000 s20019| 2015-11-23T18:45:11.142+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db9.coll9: 0ms sequenceNumber: 42 version: 2|1||56535eb76e3a936d394e6cf9 based on: 1|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.144+0000 d20010| 2015-11-23T18:45:11.142+0000 I SHARDING [conn26] received splitChunk request: { splitChunk: "db9.coll9", keyPattern: { _id: "hashed" }, min: { _id: MinKey }, max: { _id: 0 }, from: "test-rs0", splitKeys: [ { _id: -4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eb76e3a936d394e6cf9') ], epoch: ObjectId('56535eb76e3a936d394e6cf9') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.157+0000 d20010| 2015-11-23T18:45:11.157+0000 I SHARDING [conn26] distributed lock 'db9.coll9/ip-10-99-163-247:20010:1448304275:1756418349' acquired for 'splitting chunk [{ _id: MinKey }, { _id: 0 }) in db9.coll9', ts : 56535eb7a5e5109e4dbf13ed [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.158+0000 d20010| 2015-11-23T18:45:11.157+0000 I SHARDING [conn26] remotely refreshing metadata for db9.coll9 based on current shard version 0|0||56535eb76e3a936d394e6cf9, current metadata version is 1|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.158+0000 d20010| 2015-11-23T18:45:11.157+0000 I SHARDING [conn26] updating metadata for db9.coll9 from shard version 0|0||56535eb76e3a936d394e6cf9 to shard version 2|0||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.159+0000 d20010| 2015-11-23T18:45:11.157+0000 I SHARDING [conn26] collection version was loaded at version 2|1||56535eb76e3a936d394e6cf9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.159+0000 d20010| 2015-11-23T18:45:11.157+0000 I SHARDING [conn26] splitChunk accepted at version 2|0||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.162+0000 d20010| 2015-11-23T18:45:11.160+0000 I SHARDING [conn26] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.160+0000-56535eb7a5e5109e4dbf13ee", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:38802", time: new Date(1448304311160), what: "split", ns: "db9.coll9", details: { before: { min: { _id: MinKey }, max: { _id: 0 } }, left: { min: { _id: MinKey }, max: { _id: -4611686018427387902 }, lastmod: Timestamp 2000|2, lastmodEpoch: ObjectId('56535eb76e3a936d394e6cf9') }, right: { min: { _id: -4611686018427387902 }, max: { _id: 0 }, lastmod: Timestamp 2000|3, lastmodEpoch: ObjectId('56535eb76e3a936d394e6cf9') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.172+0000 d20010| 2015-11-23T18:45:11.172+0000 I SHARDING [conn26] distributed lock 'db9.coll9/ip-10-99-163-247:20010:1448304275:1756418349' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.174+0000 s20019| 2015-11-23T18:45:11.173+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db9.coll9: 0ms sequenceNumber: 43 version: 2|3||56535eb76e3a936d394e6cf9 based on: 2|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.175+0000 d20013| 2015-11-23T18:45:11.173+0000 I SHARDING [conn22] received splitChunk request: { splitChunk: "db9.coll9", keyPattern: { _id: "hashed" }, min: { _id: 0 }, max: { _id: MaxKey }, from: "test-rs1", splitKeys: [ { _id: 4611686018427387902 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 2000|1, ObjectId('56535eb76e3a936d394e6cf9') ], epoch: ObjectId('56535eb76e3a936d394e6cf9') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.181+0000 d20013| 2015-11-23T18:45:11.180+0000 I SHARDING [conn22] distributed lock 'db9.coll9/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ _id: 0 }, { _id: MaxKey }) in db9.coll9', ts : 56535eb76202d0bae2546333 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.182+0000 d20013| 2015-11-23T18:45:11.180+0000 I SHARDING [conn22] remotely refreshing metadata for db9.coll9 based on current shard version 2|0||56535eb76e3a936d394e6cf9, current metadata version is 2|0||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.182+0000 d20013| 2015-11-23T18:45:11.181+0000 I SHARDING [conn22] updating metadata for db9.coll9 from shard version 2|0||56535eb76e3a936d394e6cf9 to shard version 2|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.182+0000 d20013| 2015-11-23T18:45:11.181+0000 I SHARDING [conn22] collection version was loaded at version 2|3||56535eb76e3a936d394e6cf9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.183+0000 d20013| 2015-11-23T18:45:11.181+0000 I SHARDING [conn22] splitChunk accepted at version 2|1||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.197+0000 d20013| 2015-11-23T18:45:11.196+0000 I SHARDING [conn22] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.196+0000-56535eb76202d0bae2546334", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36459", time: new Date(1448304311196), what: "split", ns: "db9.coll9", details: { before: { min: { _id: 0 }, max: { _id: MaxKey } }, left: { min: { _id: 0 }, max: { _id: 4611686018427387902 }, lastmod: Timestamp 2000|4, lastmodEpoch: ObjectId('56535eb76e3a936d394e6cf9') }, right: { min: { _id: 4611686018427387902 }, max: { _id: MaxKey }, lastmod: Timestamp 2000|5, lastmodEpoch: ObjectId('56535eb76e3a936d394e6cf9') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.202+0000 d20013| 2015-11-23T18:45:11.202+0000 I SHARDING [conn22] distributed lock 'db9.coll9/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.203+0000 s20019| 2015-11-23T18:45:11.203+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db9.coll9: 0ms sequenceNumber: 44 version: 2|5||56535eb76e3a936d394e6cf9 based on: 2|3||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.204+0000 Using 10 threads (requested 10) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.411+0000 s20019| 2015-11-23T18:45:11.410+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40154 #52 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.414+0000 s20020| 2015-11-23T18:45:11.414+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60567 #52 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.425+0000 s20020| 2015-11-23T18:45:11.425+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60568 #53 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.426+0000 s20019| 2015-11-23T18:45:11.425+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40157 #53 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.426+0000 s20020| 2015-11-23T18:45:11.426+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60570 #54 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.433+0000 s20020| 2015-11-23T18:45:11.432+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60571 #55 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.436+0000 s20020| 2015-11-23T18:45:11.435+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60572 #56 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.440+0000 s20019| 2015-11-23T18:45:11.440+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40161 #54 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.444+0000 s20019| 2015-11-23T18:45:11.444+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40162 #55 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.445+0000 s20019| 2015-11-23T18:45:11.444+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40163 #56 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.449+0000 setting random seed: 758902531 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.450+0000 setting random seed: 967594698 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.450+0000 setting random seed: 3742762743 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.452+0000 setting random seed: 2645669617 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.452+0000 setting random seed: 4278104374 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.453+0000 setting random seed: 2928641833 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.453+0000 setting random seed: 3254025679 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.453+0000 setting random seed: 3404475635 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.453+0000 setting random seed: 3023672448 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.455+0000 setting random seed: 987587689 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.458+0000 s20020| 2015-11-23T18:45:11.455+0000 I SHARDING [conn52] ChunkManager: time to load chunks for db9.coll9: 1ms sequenceNumber: 12 version: 2|5||56535eb76e3a936d394e6cf9 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.792+0000 s20019| 2015-11-23T18:45:11.791+0000 I NETWORK [conn54] end connection 10.99.163.247:40161 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.810+0000 s20020| 2015-11-23T18:45:11.810+0000 I NETWORK [conn52] end connection 10.99.163.247:60567 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.812+0000 s20020| 2015-11-23T18:45:11.812+0000 I NETWORK [conn55] end connection 10.99.163.247:60571 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.815+0000 s20020| 2015-11-23T18:45:11.815+0000 I NETWORK [conn54] end connection 10.99.163.247:60570 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.831+0000 s20019| 2015-11-23T18:45:11.830+0000 I NETWORK [conn55] end connection 10.99.163.247:40162 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.832+0000 s20020| 2015-11-23T18:45:11.832+0000 I NETWORK [conn53] end connection 10.99.163.247:60568 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.839+0000 s20019| 2015-11-23T18:45:11.838+0000 I NETWORK [conn53] end connection 10.99.163.247:40157 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.843+0000 s20020| 2015-11-23T18:45:11.843+0000 I NETWORK [conn56] end connection 10.99.163.247:60572 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.861+0000 s20019| 2015-11-23T18:45:11.861+0000 I NETWORK [conn52] end connection 10.99.163.247:40154 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.876+0000 s20019| 2015-11-23T18:45:11.876+0000 I NETWORK [conn56] end connection 10.99.163.247:40163 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.879+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.879+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.879+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.880+0000 Workload(s) completed in 676 ms: jstests/concurrency/fsm_workloads/update_ordered_bulk_inc.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.880+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.880+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.880+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.880+0000 s20019| 2015-11-23T18:45:11.879+0000 I COMMAND [conn1] DROP: db9.coll9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.881+0000 s20019| 2015-11-23T18:45:11.879+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.879+0000-56535eb76e3a936d394e6cfb", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304311879), what: "dropCollection.start", ns: "db9.coll9", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.903+0000 s20019| 2015-11-23T18:45:11.902+0000 I SHARDING [conn1] distributed lock 'db9.coll9/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'drop', ts : 56535eb76e3a936d394e6cfc [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.903+0000 d20010| 2015-11-23T18:45:11.903+0000 I COMMAND [conn46] CMD: drop db9.coll9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.905+0000 d20013| 2015-11-23T18:45:11.904+0000 I COMMAND [conn11] CMD: drop db9.coll9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.906+0000 d20011| 2015-11-23T18:45:11.904+0000 I COMMAND [repl writer worker 9] CMD: drop db9.coll9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.906+0000 d20012| 2015-11-23T18:45:11.904+0000 I COMMAND [repl writer worker 3] CMD: drop db9.coll9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.908+0000 d20014| 2015-11-23T18:45:11.906+0000 I COMMAND [repl writer worker 1] CMD: drop db9.coll9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.909+0000 d20015| 2015-11-23T18:45:11.906+0000 I COMMAND [repl writer worker 4] CMD: drop db9.coll9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.912+0000 d20010| 2015-11-23T18:45:11.911+0000 I SHARDING [conn46] remotely refreshing metadata for db9.coll9 with requested shard version 0|0||000000000000000000000000, current shard version is 2|3||56535eb76e3a936d394e6cf9, current metadata version is 2|3||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.913+0000 d20010| 2015-11-23T18:45:11.911+0000 I SHARDING [conn46] dropping metadata for db9.coll9 at shard version 2|3||56535eb76e3a936d394e6cf9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.914+0000 d20013| 2015-11-23T18:45:11.912+0000 I SHARDING [conn11] remotely refreshing metadata for db9.coll9 with requested shard version 0|0||000000000000000000000000, current shard version is 2|5||56535eb76e3a936d394e6cf9, current metadata version is 2|5||56535eb76e3a936d394e6cf9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.914+0000 d20013| 2015-11-23T18:45:11.912+0000 I SHARDING [conn11] dropping metadata for db9.coll9 at shard version 2|5||56535eb76e3a936d394e6cf9, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.915+0000 s20019| 2015-11-23T18:45:11.913+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.913+0000-56535eb76e3a936d394e6cfd", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304311913), what: "dropCollection", ns: "db9.coll9", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.931+0000 s20019| 2015-11-23T18:45:11.931+0000 I SHARDING [conn1] distributed lock 'db9.coll9/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.932+0000 s20019| 2015-11-23T18:45:11.932+0000 I COMMAND [conn1] DROP DATABASE: db9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.932+0000 s20019| 2015-11-23T18:45:11.932+0000 I SHARDING [conn1] DBConfig::dropDatabase: db9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.933+0000 s20019| 2015-11-23T18:45:11.932+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.932+0000-56535eb76e3a936d394e6cfe", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304311932), what: "dropDatabase.start", ns: "db9", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.936+0000 s20019| 2015-11-23T18:45:11.936+0000 I SHARDING [conn1] DBConfig::dropDatabase: db9 dropped sharded collections: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.937+0000 d20013| 2015-11-23T18:45:11.936+0000 I COMMAND [conn45] dropDatabase db9 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.937+0000 d20013| 2015-11-23T18:45:11.936+0000 I COMMAND [conn45] dropDatabase db9 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.938+0000 s20019| 2015-11-23T18:45:11.936+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:11.936+0000-56535eb76e3a936d394e6cff", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304311936), what: "dropDatabase", ns: "db9", details: {} } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.938+0000 d20014| 2015-11-23T18:45:11.937+0000 I COMMAND [repl writer worker 14] dropDatabase db9 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.938+0000 d20014| 2015-11-23T18:45:11.937+0000 I COMMAND [repl writer worker 14] dropDatabase db9 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.939+0000 d20015| 2015-11-23T18:45:11.937+0000 I COMMAND [repl writer worker 13] dropDatabase db9 starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.939+0000 d20015| 2015-11-23T18:45:11.937+0000 I COMMAND [repl writer worker 13] dropDatabase db9 finished [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.940+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.940+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.940+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.940+0000 ip-10-99-163-247:20010: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.941+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.941+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.941+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.956+0000 d20010| 2015-11-23T18:45:11.955+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 ip-10-99-163-247:20010: awaitReplication completed in 16 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.957+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.958+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.958+0000 ip-10-99-163-247:20013: awaitReplication started [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.958+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.958+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.958+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.958+0000 d20012| 2015-11-23T18:45:11.956+0000 I COMMAND [repl writer worker 5] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.959+0000 d20011| 2015-11-23T18:45:11.956+0000 I COMMAND [repl writer worker 7] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.968+0000 d20013| 2015-11-23T18:45:11.968+0000 I COMMAND [conn1] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.969+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.969+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.969+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 ip-10-99-163-247:20013: awaitReplication completed in 12 ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.971+0000 Workload(s) started: jstests/concurrency/fsm_workloads/findAndModify_update.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.972+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.972+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.972+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.973+0000 d20014| 2015-11-23T18:45:11.969+0000 I COMMAND [repl writer worker 6] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.973+0000 d20015| 2015-11-23T18:45:11.969+0000 I COMMAND [repl writer worker 3] CMD: drop test.fsm_teardown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.979+0000 s20019| 2015-11-23T18:45:11.979+0000 I SHARDING [conn1] distributed lock 'db10/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'enableSharding', ts : 56535eb76e3a936d394e6d00 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.981+0000 s20019| 2015-11-23T18:45:11.980+0000 I SHARDING [conn1] Placing [db10] on: test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.981+0000 s20019| 2015-11-23T18:45:11.981+0000 I SHARDING [conn1] Enabling sharding for database [db10] in config db [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.985+0000 s20019| 2015-11-23T18:45:11.985+0000 I SHARDING [conn1] distributed lock 'db10/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.993+0000 d20013| 2015-11-23T18:45:11.993+0000 I INDEX [conn53] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0 }, name: "tid_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.994+0000 d20013| 2015-11-23T18:45:11.993+0000 I INDEX [conn53] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.995+0000 d20013| 2015-11-23T18:45:11.994+0000 I INDEX [conn53] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:11.995+0000 s20019| 2015-11-23T18:45:11.995+0000 I COMMAND [conn1] CMD: shardcollection: { shardcollection: "db10.coll10", key: { tid: 1.0 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.004+0000 d20014| 2015-11-23T18:45:12.004+0000 I INDEX [repl writer worker 15] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0 }, name: "tid_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.006+0000 d20014| 2015-11-23T18:45:12.004+0000 I INDEX [repl writer worker 15] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.007+0000 d20014| 2015-11-23T18:45:12.005+0000 I INDEX [repl writer worker 15] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.007+0000 d20015| 2015-11-23T18:45:12.004+0000 I INDEX [repl writer worker 7] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0 }, name: "tid_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.007+0000 d20015| 2015-11-23T18:45:12.004+0000 I INDEX [repl writer worker 7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.008+0000 d20015| 2015-11-23T18:45:12.005+0000 I INDEX [repl writer worker 7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.013+0000 s20019| 2015-11-23T18:45:12.012+0000 I SHARDING [conn1] distributed lock 'db10.coll10/ip-10-99-163-247:20019:1448304274:1804289383' acquired for 'shardCollection', ts : 56535eb76e3a936d394e6d01 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.013+0000 s20019| 2015-11-23T18:45:12.013+0000 I SHARDING [conn1] enable sharding on: db10.coll10 with shard key: { tid: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.015+0000 s20019| 2015-11-23T18:45:12.013+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:12.013+0000-56535eb86e3a936d394e6d02", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304312013), what: "shardCollection.start", ns: "db10.coll10", details: { shardKey: { tid: 1.0 }, collection: "db10.coll10", primary: "test-rs1:test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015", initShards: [], numChunks: 1 } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.015+0000 s20019| 2015-11-23T18:45:12.014+0000 I SHARDING [conn1] going to create 1 chunk(s) for: db10.coll10 using new epoch 56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.019+0000 s20019| 2015-11-23T18:45:12.019+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 45 version: 1|0||56535eb86e3a936d394e6d03 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.025+0000 s20019| 2015-11-23T18:45:12.025+0000 I SHARDING [conn1] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 46 version: 1|0||56535eb86e3a936d394e6d03 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.026+0000 d20013| 2015-11-23T18:45:12.025+0000 I SHARDING [conn22] remotely refreshing metadata for db10.coll10 with requested shard version 1|0||56535eb86e3a936d394e6d03, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.026+0000 d20013| 2015-11-23T18:45:12.026+0000 I SHARDING [conn22] collection db10.coll10 was previously unsharded, new metadata loaded with shard version 1|0||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.027+0000 d20013| 2015-11-23T18:45:12.026+0000 I SHARDING [conn22] collection version was loaded at version 1|0||56535eb86e3a936d394e6d03, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.027+0000 s20019| 2015-11-23T18:45:12.026+0000 I SHARDING [conn1] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:12.026+0000-56535eb86e3a936d394e6d04", server: "ip-10-99-163-247", clientAddr: "127.0.0.1:51291", time: new Date(1448304312026), what: "shardCollection.end", ns: "db10.coll10", details: { version: "1|0||56535eb86e3a936d394e6d03" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.039+0000 s20019| 2015-11-23T18:45:12.039+0000 I SHARDING [conn1] distributed lock 'db10.coll10/ip-10-99-163-247:20019:1448304274:1804289383' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.039+0000 Using 20 threads (requested 20) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.051+0000 d20013| 2015-11-23T18:45:12.051+0000 I INDEX [conn22] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0, value: 1.0 }, name: "tid_1_value_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.052+0000 d20013| 2015-11-23T18:45:12.051+0000 I INDEX [conn22] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.056+0000 d20013| 2015-11-23T18:45:12.056+0000 I INDEX [conn22] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.062+0000 d20015| 2015-11-23T18:45:12.062+0000 I INDEX [repl writer worker 12] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0, value: 1.0 }, name: "tid_1_value_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.063+0000 d20015| 2015-11-23T18:45:12.062+0000 I INDEX [repl writer worker 12] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.067+0000 d20015| 2015-11-23T18:45:12.067+0000 I INDEX [repl writer worker 12] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.068+0000 d20010| 2015-11-23T18:45:12.067+0000 I INDEX [conn26] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0, value: 1.0 }, name: "tid_1_value_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.068+0000 d20010| 2015-11-23T18:45:12.067+0000 I INDEX [conn26] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.077+0000 d20010| 2015-11-23T18:45:12.075+0000 I INDEX [conn26] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.077+0000 d20014| 2015-11-23T18:45:12.075+0000 I INDEX [repl writer worker 7] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0, value: 1.0 }, name: "tid_1_value_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.077+0000 d20014| 2015-11-23T18:45:12.075+0000 I INDEX [repl writer worker 7] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.082+0000 d20014| 2015-11-23T18:45:12.081+0000 I INDEX [repl writer worker 7] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.091+0000 d20012| 2015-11-23T18:45:12.090+0000 I INDEX [repl writer worker 10] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0, value: 1.0 }, name: "tid_1_value_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.091+0000 d20012| 2015-11-23T18:45:12.090+0000 I INDEX [repl writer worker 10] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.093+0000 d20012| 2015-11-23T18:45:12.092+0000 I INDEX [repl writer worker 10] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.110+0000 d20011| 2015-11-23T18:45:12.109+0000 I INDEX [repl writer worker 11] build index on: db10.coll10 properties: { v: 1, key: { tid: 1.0, value: 1.0 }, name: "tid_1_value_1", ns: "db10.coll10" } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.110+0000 d20011| 2015-11-23T18:45:12.109+0000 I INDEX [repl writer worker 11] building index using bulk method [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.111+0000 d20011| 2015-11-23T18:45:12.111+0000 I INDEX [repl writer worker 11] build index done. scanned 0 total records. 0 secs [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.443+0000 s20019| 2015-11-23T18:45:12.442+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40164 #57 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.443+0000 s20019| 2015-11-23T18:45:12.442+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40165 #58 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.452+0000 s20019| 2015-11-23T18:45:12.451+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40166 #59 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.452+0000 s20020| 2015-11-23T18:45:12.451+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60578 #57 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.470+0000 s20019| 2015-11-23T18:45:12.469+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40168 #60 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.475+0000 s20019| 2015-11-23T18:45:12.474+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40169 #61 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.478+0000 s20020| 2015-11-23T18:45:12.478+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60582 #58 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.484+0000 s20019| 2015-11-23T18:45:12.483+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40171 #62 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.487+0000 s20020| 2015-11-23T18:45:12.487+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60584 #59 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.490+0000 s20020| 2015-11-23T18:45:12.489+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60585 #60 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.495+0000 s20020| 2015-11-23T18:45:12.495+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60586 #61 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.503+0000 s20019| 2015-11-23T18:45:12.502+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40175 #63 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.506+0000 s20019| 2015-11-23T18:45:12.505+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40176 #64 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.509+0000 s20019| 2015-11-23T18:45:12.508+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40177 #65 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.510+0000 s20020| 2015-11-23T18:45:12.510+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60590 #62 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.515+0000 s20020| 2015-11-23T18:45:12.514+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60591 #63 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.516+0000 s20019| 2015-11-23T18:45:12.516+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:40180 #66 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.517+0000 s20020| 2015-11-23T18:45:12.516+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60593 #64 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.519+0000 s20020| 2015-11-23T18:45:12.519+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60594 #65 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.521+0000 s20020| 2015-11-23T18:45:12.521+0000 I NETWORK [mongosMain] connection accepted from 10.99.163.247:60595 #66 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.525+0000 setting random seed: 2638355945 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.526+0000 setting random seed: 3911323193 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.526+0000 setting random seed: 900736642 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.526+0000 setting random seed: 757301083 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.526+0000 setting random seed: 2144410341 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.531+0000 setting random seed: 240902168 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.532+0000 setting random seed: 2194150361 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.532+0000 setting random seed: 1841123515 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.532+0000 setting random seed: 3652125693 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.532+0000 setting random seed: 1844331217 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.533+0000 setting random seed: 1792967786 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.533+0000 setting random seed: 1973481611 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.533+0000 setting random seed: 189759555 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.533+0000 setting random seed: 1313219518 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.533+0000 setting random seed: 2094718352 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.534+0000 setting random seed: 2491889584 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.534+0000 s20020| 2015-11-23T18:45:12.532+0000 I SHARDING [conn57] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 13 version: 1|0||56535eb86e3a936d394e6d03 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.538+0000 setting random seed: 2679449129 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.538+0000 setting random seed: 949813967 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.538+0000 setting random seed: 3153802872 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.567+0000 setting random seed: 289134405 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.568+0000 d20013| 2015-11-23T18:45:12.541+0000 I SHARDING [conn11] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.568+0000 d20013| 2015-11-23T18:45:12.541+0000 I SHARDING [conn50] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.569+0000 d20013| 2015-11-23T18:45:12.541+0000 I SHARDING [conn54] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.569+0000 d20013| 2015-11-23T18:45:12.542+0000 I SHARDING [conn67] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.570+0000 d20013| 2015-11-23T18:45:12.542+0000 I SHARDING [conn11] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.570+0000 d20013| 2015-11-23T18:45:12.542+0000 I SHARDING [conn50] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.571+0000 d20013| 2015-11-23T18:45:12.542+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36712 #76 (69 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.573+0000 d20013| 2015-11-23T18:45:12.542+0000 I SHARDING [conn40] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 2.0 }, { tid: 18.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.576+0000 d20013| 2015-11-23T18:45:12.543+0000 I SHARDING [conn61] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 2.0 }, { tid: 18.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.577+0000 d20013| 2015-11-23T18:45:12.543+0000 I SHARDING [conn11] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.578+0000 d20013| 2015-11-23T18:45:12.544+0000 I SHARDING [conn66] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 2.0 }, { tid: 17.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.578+0000 d20013| 2015-11-23T18:45:12.544+0000 I NETWORK [conn66] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.579+0000 d20013| 2015-11-23T18:45:12.545+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36713 #77 (70 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.579+0000 d20013| 2015-11-23T18:45:12.546+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36714 #78 (71 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.580+0000 d20013| 2015-11-23T18:45:12.546+0000 I SHARDING [conn24] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.594+0000 d20013| 2015-11-23T18:45:12.546+0000 I SHARDING [conn54] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.616+0000 d20013| 2015-11-23T18:45:12.546+0000 I SHARDING [conn64] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.616+0000 d20013| 2015-11-23T18:45:12.546+0000 I NETWORK [conn64] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.618+0000 d20013| 2015-11-23T18:45:12.547+0000 I SHARDING [conn41] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 18.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.618+0000 d20013| 2015-11-23T18:45:12.547+0000 I NETWORK [conn41] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.620+0000 d20013| 2015-11-23T18:45:12.547+0000 I SHARDING [conn22] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.620+0000 d20013| 2015-11-23T18:45:12.547+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.673+0000 d20013| 2015-11-23T18:45:12.548+0000 I SHARDING [conn38] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 17.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.675+0000 d20013| 2015-11-23T18:45:12.548+0000 I NETWORK [conn38] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.675+0000 d20013| 2015-11-23T18:45:12.548+0000 I SHARDING [conn67] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.676+0000 d20013| 2015-11-23T18:45:12.549+0000 I SHARDING [conn75] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 13.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.676+0000 d20013| 2015-11-23T18:45:12.549+0000 I NETWORK [conn75] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.684+0000 d20013| 2015-11-23T18:45:12.551+0000 I SHARDING [conn40] could not acquire lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' (another update won) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.684+0000 d20013| 2015-11-23T18:45:12.551+0000 I SHARDING [conn40] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' was not acquired. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.686+0000 d20013| 2015-11-23T18:45:12.551+0000 W SHARDING [conn40] could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.688+0000 s20020| 2015-11-23T18:45:12.552+0000 W SHARDING [conn61] splitChunk failed - cmd: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 2.0 }, { tid: 18.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: Mi..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.689+0000 c20016| 2015-11-23T18:45:12.553+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48701 #48 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.690+0000 d20013| 2015-11-23T18:45:12.555+0000 I SHARDING [conn67] request split points lookup for chunk db10.coll10 { : MinKey } -->> { : MaxKey } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.692+0000 c20016| 2015-11-23T18:45:12.556+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48702 #49 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.692+0000 d20013| 2015-11-23T18:45:12.558+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36717 #79 (72 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.693+0000 d20013| 2015-11-23T18:45:12.558+0000 I SHARDING [conn61] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10', ts : 56535eb86202d0bae2546337 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.694+0000 d20013| 2015-11-23T18:45:12.558+0000 I SHARDING [conn61] remotely refreshing metadata for db10.coll10 based on current shard version 1|0||56535eb86e3a936d394e6d03, current metadata version is 1|0||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.695+0000 c20016| 2015-11-23T18:45:12.559+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48704 #50 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.698+0000 c20016| 2015-11-23T18:45:12.559+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48705 #51 (48 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.700+0000 c20016| 2015-11-23T18:45:12.559+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48706 #52 (49 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.700+0000 d20013| 2015-11-23T18:45:12.559+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36721 #80 (73 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.702+0000 d20013| 2015-11-23T18:45:12.560+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36722 #81 (74 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.702+0000 d20013| 2015-11-23T18:45:12.560+0000 I SHARDING [conn61] metadata of collection db10.coll10 already up to date (shard version : 1|0||56535eb86e3a936d394e6d03, took 1ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.709+0000 d20013| 2015-11-23T18:45:12.560+0000 I SHARDING [conn61] splitChunk accepted at version 1|0||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.710+0000 s20019| 2015-11-23T18:45:12.561+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.711+0000 s20019| 2015-11-23T18:45:12.561+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.711+0000 d20013| 2015-11-23T18:45:12.561+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36723 #82 (75 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.712+0000 d20013| 2015-11-23T18:45:12.562+0000 I NETWORK [conn64] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.713+0000 d20013| 2015-11-23T18:45:12.562+0000 I NETWORK [conn38] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.717+0000 d20013| 2015-11-23T18:45:12.562+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.718+0000 d20013| 2015-11-23T18:45:12.563+0000 I NETWORK [conn66] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.734+0000 d20013| 2015-11-23T18:45:12.563+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36724 #83 (76 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.735+0000 d20013| 2015-11-23T18:45:12.564+0000 I NETWORK [conn41] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.735+0000 d20013| 2015-11-23T18:45:12.565+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36725 #84 (77 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.739+0000 c20017| 2015-11-23T18:45:12.567+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56026 #46 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.739+0000 d20013| 2015-11-23T18:45:12.567+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36727 #85 (78 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.740+0000 d20013| 2015-11-23T18:45:12.568+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36729 #86 (79 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.740+0000 c20017| 2015-11-23T18:45:12.568+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56028 #47 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.740+0000 d20013| 2015-11-23T18:45:12.569+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36730 #87 (80 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.750+0000 s20019| 2015-11-23T18:45:12.570+0000 I SHARDING [conn57] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 47 version: 1|3||56535eb86e3a936d394e6d03 based on: 1|0||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.753+0000 s20019| 2015-11-23T18:45:12.570+0000 I SHARDING [conn57] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.753+0000 s20019| 2015-11-23T18:45:12.570+0000 I SHARDING [conn57] retrying command: { findandmodify: "coll10", query: { tid: 2.0 }, sort: { value: -1.0 }, update: { $min: { value: -2.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.755+0000 s20020| 2015-11-23T18:45:12.571+0000 I SHARDING [conn58] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 14 version: 1|3||56535eb86e3a936d394e6d03 based on: 1|0||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.756+0000 s20020| 2015-11-23T18:45:12.571+0000 I SHARDING [conn58] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.757+0000 s20020| 2015-11-23T18:45:12.571+0000 I SHARDING [conn58] retrying command: { findandmodify: "coll10", query: { tid: 7.0 }, sort: { value: -1.0 }, update: { $min: { value: -7.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.758+0000 d20010| 2015-11-23T18:45:12.571+0000 I SHARDING [conn26] remotely refreshing metadata for db10.coll10 with requested shard version 0|0||56535eb86e3a936d394e6d03, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.758+0000 d20010| 2015-11-23T18:45:12.572+0000 I SHARDING [conn26] collection db10.coll10 was previously unsharded, new metadata loaded with shard version 0|0||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.759+0000 d20010| 2015-11-23T18:45:12.572+0000 I SHARDING [conn26] collection version was loaded at version 1|3||56535eb86e3a936d394e6d03, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.759+0000 d20013| 2015-11-23T18:45:12.573+0000 I NETWORK [conn79] end connection 10.99.163.247:36717 (79 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.759+0000 s20019| 2015-11-23T18:45:12.573+0000 I SHARDING [conn63] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.760+0000 s20019| 2015-11-23T18:45:12.573+0000 I SHARDING [conn63] retrying command: { findandmodify: "coll10", query: { tid: 16.0 }, sort: { value: 1.0 }, update: { $max: { value: 16.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.760+0000 d20013| 2015-11-23T18:45:12.574+0000 I NETWORK [conn84] end connection 10.99.163.247:36725 (78 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.761+0000 s20020| 2015-11-23T18:45:12.574+0000 I SHARDING [conn64] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.761+0000 s20020| 2015-11-23T18:45:12.574+0000 I SHARDING [conn64] retrying command: { findandmodify: "coll10", query: { tid: 9.0 }, sort: { value: -1.0 }, update: { $min: { value: -9.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.761+0000 d20013| 2015-11-23T18:45:12.574+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36731 #88 (79 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.761+0000 d20013| 2015-11-23T18:45:12.575+0000 I NETWORK [conn81] end connection 10.99.163.247:36722 (78 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.762+0000 d20013| 2015-11-23T18:45:12.576+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36732 #89 (79 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.762+0000 c20017| 2015-11-23T18:45:12.576+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56033 #48 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.762+0000 d20013| 2015-11-23T18:45:12.577+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36734 #90 (80 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.763+0000 s20019| 2015-11-23T18:45:12.577+0000 I SHARDING [conn65] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.764+0000 s20019| 2015-11-23T18:45:12.577+0000 I SHARDING [conn65] retrying command: { findandmodify: "coll10", query: { tid: 6.0 }, sort: { value: -1.0 }, update: { $min: { value: -6.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.764+0000 d20013| 2015-11-23T18:45:12.577+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36735 #91 (81 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.765+0000 d20013| 2015-11-23T18:45:12.577+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:12.577+0000-56535eb86202d0bae254633e", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304312577), what: "multi-split", ns: "db10.coll10", details: { before: { min: { tid: MinKey }, max: { tid: MaxKey } }, number: 1, of: 3, chunk: { min: { tid: MinKey }, max: { tid: 2.0 }, lastmod: Timestamp 1000|1, lastmodEpoch: ObjectId('56535eb86e3a936d394e6d03') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.766+0000 d20013| 2015-11-23T18:45:12.578+0000 I SHARDING [conn83] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 9.0 }, { tid: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.766+0000 d20013| 2015-11-23T18:45:12.580+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36736 #92 (82 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.767+0000 d20013| 2015-11-23T18:45:12.580+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:12.580+0000-56535eb86202d0bae2546340", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304312580), what: "multi-split", ns: "db10.coll10", details: { before: { min: { tid: MinKey }, max: { tid: MaxKey } }, number: 2, of: 3, chunk: { min: { tid: 2.0 }, max: { tid: 18.0 }, lastmod: Timestamp 1000|2, lastmodEpoch: ObjectId('56535eb86e3a936d394e6d03') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.768+0000 d20013| 2015-11-23T18:45:12.580+0000 I NETWORK [conn64] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.768+0000 d20013| 2015-11-23T18:45:12.581+0000 W SHARDING [conn83] could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.770+0000 s20020| 2015-11-23T18:45:12.581+0000 W SHARDING [conn61] splitChunk failed - cmd: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 9.0 }, { tid: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: Mi..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.771+0000 d20013| 2015-11-23T18:45:12.582+0000 I SHARDING [conn61] about to log metadata event into changelog: { _id: "ip-10-99-163-247-2015-11-23T18:45:12.582+0000-56535eb86202d0bae2546341", server: "ip-10-99-163-247", clientAddr: "10.99.163.247:36586", time: new Date(1448304312582), what: "multi-split", ns: "db10.coll10", details: { before: { min: { tid: MinKey }, max: { tid: MaxKey } }, number: 3, of: 3, chunk: { min: { tid: 18.0 }, max: { tid: MaxKey }, lastmod: Timestamp 1000|3, lastmodEpoch: ObjectId('56535eb86e3a936d394e6d03') } } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.771+0000 s20019| 2015-11-23T18:45:12.582+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.772+0000 d20013| 2015-11-23T18:45:12.584+0000 I NETWORK [conn85] end connection 10.99.163.247:36727 (81 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.772+0000 c20016| 2015-11-23T18:45:12.585+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48723 #53 (50 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.772+0000 s20020| 2015-11-23T18:45:12.586+0000 I SHARDING [conn60] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.773+0000 s20020| 2015-11-23T18:45:12.586+0000 I SHARDING [conn60] retrying command: { findandmodify: "coll10", query: { tid: 1.0 }, sort: { value: -1.0 }, update: { $min: { value: -1.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.773+0000 d20013| 2015-11-23T18:45:12.586+0000 I SHARDING [conn86] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 12.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.774+0000 d20013| 2015-11-23T18:45:12.587+0000 I NETWORK [conn80] end connection 10.99.163.247:36721 (80 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.774+0000 s20019| 2015-11-23T18:45:12.587+0000 I SHARDING [conn58] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.774+0000 s20019| 2015-11-23T18:45:12.587+0000 I SHARDING [conn58] retrying command: { findandmodify: "coll10", query: { tid: 4.0 }, sort: { value: -1.0 }, update: { $min: { value: -4.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.775+0000 d20013| 2015-11-23T18:45:12.588+0000 I NETWORK [conn87] end connection 10.99.163.247:36730 (79 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.775+0000 c20017| 2015-11-23T18:45:12.588+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56038 #49 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.775+0000 c20017| 2015-11-23T18:45:12.589+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56039 #50 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.775+0000 d20013| 2015-11-23T18:45:12.588+0000 I NETWORK [conn82] end connection 10.99.163.247:36723 (79 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.776+0000 d20013| 2015-11-23T18:45:12.591+0000 I NETWORK [conn41] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.776+0000 d20013| 2015-11-23T18:45:12.589+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36740 #93 (81 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.776+0000 d20013| 2015-11-23T18:45:12.592+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36741 #94 (80 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.776+0000 d20013| 2015-11-23T18:45:12.592+0000 I NETWORK [conn75] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.777+0000 c20018| 2015-11-23T18:45:12.592+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37162 #46 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.777+0000 s20020| 2015-11-23T18:45:12.592+0000 I SHARDING [conn65] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.777+0000 s20020| 2015-11-23T18:45:12.592+0000 I SHARDING [conn65] retrying command: { findandmodify: "coll10", query: { tid: 11.0 }, sort: { value: -1.0 }, update: { $min: { value: -11.0 } }, new: true } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.778+0000 d20013| 2015-11-23T18:45:12.592+0000 I NETWORK [conn88] end connection 10.99.163.247:36731 (79 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.778+0000 d20013| 2015-11-23T18:45:12.593+0000 I SHARDING [conn61] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.778+0000 d20013| 2015-11-23T18:45:12.593+0000 I NETWORK [conn66] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.779+0000 s20019| 2015-11-23T18:45:12.594+0000 I SHARDING [conn64] autosplitted db10.coll10 shard: ns: db10.coll10, shard: test-rs1, lastmod: 1|0||56535eb86e3a936d394e6d03, min: { tid: MinKey }, max: { tid: MaxKey } into 3 (splitThreshold 921) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.779+0000 d20013| 2015-11-23T18:45:12.594+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36743 #95 (80 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.780+0000 d20013| 2015-11-23T18:45:12.594+0000 I SHARDING [conn92] received splitChunk request: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 13.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.780+0000 c20018| 2015-11-23T18:45:12.594+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37164 #47 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.780+0000 d20013| 2015-11-23T18:45:12.597+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36745 #96 (81 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.781+0000 d20013| 2015-11-23T18:45:12.597+0000 I NETWORK [conn38] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.781+0000 c20017| 2015-11-23T18:45:12.598+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56046 #51 (48 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.781+0000 c20018| 2015-11-23T18:45:12.599+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37167 #48 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.783+0000 d20013| 2015-11-23T18:45:12.600+0000 W SHARDING [conn92] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772 did not propagate properly. :: caused by :: 8017 update not consistent ns: config.locks query: { _id: "db10.coll10", state: 0, ts: ObjectId('56535eb86202d0bae2546337') } update: { $set: { state: 1, who: "ip-10-99-163-247:20013:1448304277:68810772:conn92:1575714183", process: "ip-10-99-163-247:20013:1448304277:68810772", when: new Date(1448304312594), why: "splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10", ts: ObjectId('56535eb86202d0bae2546343') } } gle1: { connectionId: 37, updatedExisting: true, n: 1, syncMillis: 0, writtenTo: null, err: null, waited: 0, ok: 1.0 } gle2: { connectionId: 35, updatedExisting: false, n: 0, syncMillis: 0, writtenTo: null, err: null, waited: 0, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.784+0000 d20013| 2015-11-23T18:45:12.601+0000 W SHARDING [conn86] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772 did not propagate properly. :: caused by :: 8017 update not consistent ns: config.locks query: { _id: "db10.coll10", state: 0, ts: ObjectId('56535eb86202d0bae2546337') } update: { $set: { state: 1, who: "ip-10-99-163-247:20013:1448304277:68810772:conn86:528048843", process: "ip-10-99-163-247:20013:1448304277:68810772", when: new Date(1448304312590), why: "splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10", ts: ObjectId('56535eb86202d0bae2546342') } } gle1: { connectionId: 38, updatedExisting: false, n: 0, syncMillis: 0, writtenTo: null, err: null, waited: 0, ok: 1.0 } gle2: { connectionId: 36, updatedExisting: true, n: 1, syncMillis: 0, writtenTo: null, err: null, waited: 0, ok: 1.0 } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.784+0000 d20013| 2015-11-23T18:45:12.603+0000 I SHARDING [conn92] lock update won, completing lock propagation for 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.785+0000 d20013| 2015-11-23T18:45:12.610+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36748 #97 (82 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.785+0000 d20013| 2015-11-23T18:45:12.611+0000 I NETWORK [conn22] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.785+0000 c20016| 2015-11-23T18:45:12.612+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48735 #54 (51 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.786+0000 d20013| 2015-11-23T18:45:12.618+0000 I SHARDING [conn92] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10', ts : 56535eb86202d0bae2546343 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.786+0000 d20013| 2015-11-23T18:45:12.618+0000 I SHARDING [conn92] remotely refreshing metadata for db10.coll10 based on current shard version 1|3||56535eb86e3a936d394e6d03, current metadata version is 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.786+0000 d20013| 2015-11-23T18:45:12.618+0000 I SHARDING [conn92] metadata of collection db10.coll10 already up to date (shard version : 1|3||56535eb86e3a936d394e6d03, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.787+0000 d20013| 2015-11-23T18:45:12.618+0000 W SHARDING [conn92] splitChunk cannot find chunk [{ tid: MinKey },{ tid: MaxKey }) to split, the chunk boundaries may be stale [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.787+0000 d20013| 2015-11-23T18:45:12.620+0000 W SHARDING [conn64] could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.788+0000 s20019| 2015-11-23T18:45:12.620+0000 W SHARDING [conn62] splitChunk failed - cmd: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: Mi..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.788+0000 d20013| 2015-11-23T18:45:12.623+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:36750 #98 (83 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.789+0000 d20013| 2015-11-23T18:45:12.624+0000 I NETWORK [conn75] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.789+0000 s20020| 2015-11-23T18:45:12.625+0000 I NETWORK [conn65] end connection 10.99.163.247:60594 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.789+0000 c20018| 2015-11-23T18:45:12.651+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37171 #49 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.789+0000 s20020| 2015-11-23T18:45:12.652+0000 I NETWORK [conn61] end connection 10.99.163.247:60586 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.790+0000 s20019| 2015-11-23T18:45:12.652+0000 I NETWORK [conn66] end connection 10.99.163.247:40180 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.790+0000 c20018| 2015-11-23T18:45:12.654+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37172 #50 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.790+0000 s20020| 2015-11-23T18:45:12.655+0000 I NETWORK [conn64] end connection 10.99.163.247:60593 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.791+0000 s20020| 2015-11-23T18:45:12.656+0000 I NETWORK [conn66] end connection 10.99.163.247:60595 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.791+0000 d20013| 2015-11-23T18:45:12.658+0000 I SHARDING [conn92] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.791+0000 s20020| 2015-11-23T18:45:12.658+0000 I SHARDING [conn63] sharded connection to test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.792+0000 s20020| 2015-11-23T18:45:12.658+0000 W SHARDING [conn63] could not autosplit collection db10.coll10 :: caused by :: 9996 stale config in runCommand ( ns : db10.coll10, received : 1|0||56535eb86e3a936d394e6d03, wanted : 1|3||56535eb86e3a936d394e6d03, recv ) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.792+0000 s20019| 2015-11-23T18:45:12.660+0000 I NETWORK [conn63] end connection 10.99.163.247:40175 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.792+0000 d20013| 2015-11-23T18:45:12.661+0000 I NETWORK [conn92] end connection 10.99.163.247:36736 (82 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.792+0000 s20019| 2015-11-23T18:45:12.663+0000 I NETWORK [conn57] end connection 10.99.163.247:40164 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.793+0000 c20016| 2015-11-23T18:45:12.672+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48739 #55 (52 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.793+0000 c20018| 2015-11-23T18:45:12.673+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37174 #51 (48 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.793+0000 d20013| 2015-11-23T18:45:12.678+0000 W SHARDING [conn75] could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.794+0000 d20013| 2015-11-23T18:45:12.678+0000 I COMMAND [conn75] command admin.$cmd command: splitChunk { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 13.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:346 locks:{} protocol:op_command 128ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.795+0000 s20020| 2015-11-23T18:45:12.678+0000 W SHARDING [conn59] splitChunk failed - cmd: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 13.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: Mi..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.796+0000 d20013| 2015-11-23T18:45:12.679+0000 I SHARDING [conn86] lock update won, completing lock propagation for 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.796+0000 d20013| 2015-11-23T18:45:12.679+0000 W SHARDING [conn66] could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.797+0000 d20013| 2015-11-23T18:45:12.679+0000 I COMMAND [conn66] command admin.$cmd command: splitChunk { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 2.0 }, { tid: 17.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:366 locks:{} protocol:op_command 135ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.798+0000 s20019| 2015-11-23T18:45:12.680+0000 W SHARDING [conn59] splitChunk failed - cmd: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 2.0 }, { tid: 17.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: Mi..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.799+0000 d20013| 2015-11-23T18:45:12.689+0000 W SHARDING [conn38] could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.800+0000 d20013| 2015-11-23T18:45:12.689+0000 I COMMAND [conn38] command admin.$cmd command: splitChunk { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 17.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:366 locks:{} protocol:op_command 141ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.801+0000 s20020| 2015-11-23T18:45:12.689+0000 W SHARDING [conn57] splitChunk failed - cmd: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 17.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: Mi..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.801+0000 c20016| 2015-11-23T18:45:12.689+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48741 #56 (53 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.801+0000 d20013| 2015-11-23T18:45:12.700+0000 W SHARDING [conn22] could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10 is taken. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.802+0000 d20013| 2015-11-23T18:45:12.700+0000 I COMMAND [conn22] command admin.$cmd command: splitChunk { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } ntoreturn:1 ntoskip:0 keyUpdates:0 writeConflicts:0 numYields:0 reslen:366 locks:{} protocol:op_command 152ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.803+0000 s20019| 2015-11-23T18:45:12.700+0000 W SHARDING [conn60] splitChunk failed - cmd: { splitChunk: "db10.coll10", keyPattern: { tid: 1.0 }, min: { tid: MinKey }, max: { tid: MaxKey }, from: "test-rs1", splitKeys: [ { tid: 1.0 }, { tid: 16.0 } ], configdb: "ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018", shardVersion: [ Timestamp 1000|0, ObjectId('56535eb86e3a936d394e6d03') ], epoch: ObjectId('56535eb86e3a936d394e6d03') } result: { ok: 0.0, errmsg: "could not acquire collection lock for db10.coll10 to split chunk [{ tid: MinKey },{ tid: MaxKey }) :: caused by :: Lock for splitting chunk [{ tid: Mi..." } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.804+0000 s20020| 2015-11-23T18:45:12.701+0000 I NETWORK [conn58] end connection 10.99.163.247:60582 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.804+0000 s20020| 2015-11-23T18:45:12.705+0000 I NETWORK [conn60] end connection 10.99.163.247:60585 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.804+0000 s20019| 2015-11-23T18:45:12.718+0000 I NETWORK [conn65] end connection 10.99.163.247:40177 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.804+0000 s20019| 2015-11-23T18:45:12.723+0000 I NETWORK [conn58] end connection 10.99.163.247:40165 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.805+0000 s20019| 2015-11-23T18:45:12.723+0000 I NETWORK [conn62] end connection 10.99.163.247:40171 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.805+0000 s20019| 2015-11-23T18:45:12.732+0000 I NETWORK [conn59] end connection 10.99.163.247:40166 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.805+0000 s20019| 2015-11-23T18:45:12.733+0000 I NETWORK [conn64] end connection 10.99.163.247:40176 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.806+0000 d20013| 2015-11-23T18:45:12.739+0000 I SHARDING [conn86] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10', ts : 56535eb86202d0bae254633a [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.806+0000 d20013| 2015-11-23T18:45:12.739+0000 I SHARDING [conn86] remotely refreshing metadata for db10.coll10 based on current shard version 1|3||56535eb86e3a936d394e6d03, current metadata version is 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.806+0000 d20013| 2015-11-23T18:45:12.739+0000 I SHARDING [conn41] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' acquired for 'splitting chunk [{ tid: MinKey }, { tid: MaxKey }) in db10.coll10', ts : 56535eb86202d0bae254633a [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.807+0000 d20013| 2015-11-23T18:45:12.739+0000 I SHARDING [conn41] remotely refreshing metadata for db10.coll10 based on current shard version 1|3||56535eb86e3a936d394e6d03, current metadata version is 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.807+0000 d20013| 2015-11-23T18:45:12.740+0000 I SHARDING [conn41] metadata of collection db10.coll10 already up to date (shard version : 1|3||56535eb86e3a936d394e6d03, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.808+0000 d20013| 2015-11-23T18:45:12.740+0000 W SHARDING [conn41] splitChunk cannot find chunk [{ tid: MinKey },{ tid: MaxKey }) to split, the chunk boundaries may be stale [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.808+0000 d20013| 2015-11-23T18:45:12.740+0000 I SHARDING [conn86] metadata of collection db10.coll10 already up to date (shard version : 1|3||56535eb86e3a936d394e6d03, took 0ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.808+0000 d20013| 2015-11-23T18:45:12.740+0000 W SHARDING [conn86] splitChunk cannot find chunk [{ tid: MinKey },{ tid: MaxKey }) to split, the chunk boundaries may be stale [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.809+0000 d20013| 2015-11-23T18:45:12.740+0000 I - [conn86] Invariant failure iter != _lockMap.end() src/mongo/s/catalog/legacy/legacy_dist_lock_manager.cpp 182 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.809+0000 d20013| 2015-11-23T18:45:12.740+0000 I - [conn86] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.809+0000 d20013| [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.809+0000 d20013| ***aborting after invariant() failure [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.809+0000 d20013| [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.809+0000 d20013| [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.809+0000 s20020| 2015-11-23T18:45:12.748+0000 I NETWORK [conn63] end connection 10.99.163.247:60591 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.810+0000 s20019| 2015-11-23T18:45:12.750+0000 I NETWORK [conn60] end connection 10.99.163.247:40168 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.810+0000 d20013| 2015-11-23T18:45:12.750+0000 I SHARDING [conn41] distributed lock 'db10.coll10/ip-10-99-163-247:20013:1448304277:68810772' unlocked. [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.810+0000 d20013| 2015-11-23T18:45:12.753+0000 F - [conn86] Got signal: 6 (Aborted). [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.810+0000 d20013| [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.811+0000 d20013| 0x137f942 0x137e889 0x137f092 0x3aa100f790 0x3aa0c32625 0x3aa0c33e05 0x130774b 0x11c1590 0x118fd18 0xfe6d8b 0x126b3f7 0xba4eea 0xba5b86 0xaff3d0 0xcb921d 0x98a86c 0x132b9ed 0x3aa1007a51 0x3aa0ce893d [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.811+0000 d20013| ----- BEGIN BACKTRACE ----- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.822+0000 d20013| {"backtrace":[{"b":"400000","o":"F7F942"},{"b":"400000","o":"F7E889"},{"b":"400000","o":"F7F092"},{"b":"3AA1000000","o":"F790"},{"b":"3AA0C00000","o":"32625"},{"b":"3AA0C00000","o":"33E05"},{"b":"400000","o":"F0774B"},{"b":"400000","o":"DC1590"},{"b":"400000","o":"D8FD18"},{"b":"400000","o":"BE6D8B"},{"b":"400000","o":"E6B3F7"},{"b":"400000","o":"7A4EEA"},{"b":"400000","o":"7A5B86"},{"b":"400000","o":"6FF3D0"},{"b":"400000","o":"8B921D"},{"b":"400000","o":"58A86C"},{"b":"400000","o":"F2B9ED"},{"b":"3AA1000000","o":"7A51"},{"b":"3AA0C00000","o":"E893D"}],"processInfo":{ "mongodbVersion" : "3.2.0-rc3-103-ge3cd63f", "gitVersion" : "e3cd63fcae3deb1140941a51c85564f098062a23", "compiledModules" : [ "enterprise" ], "uname" : { "sysname" : "Linux", "release" : "2.6.32-220.el6.x86_64", "version" : "#1 SMP Wed Nov 9 08:03:13 EST 2011", "machine" : "x86_64" }, "somap" : [ { "elfType" : 2, "b" : "400000", "buildId" : "CE2EBCEACDDF77CDBBC11D3208BF858DEBC7137B" }, { "b" : "7FFF87AFF000", "elfType" : 3, "buildId" : "08F634A1D22DEFF00461D50A7699DACDC97657BF" }, { "b" : "7FCAEB3F2000", "path" : "/usr/lib64/libnetsnmpagent.so.20", "elfType" : 3, "buildId" : "E4E49DE2554F02ACF2728D1748874101B0709B3A" }, { "b" : "7FCAEB1CB000", "path" : "/usr/lib64/libnetsnmphelpers.so.20", "elfType" : 3, "buildId" : "17A35AEE324676929C7A5C8B4CE54443ED10AC07" }, { "b" : "7FCAEAD03000", "path" : "/usr/lib64/libnetsnmpmibs.so.20", "elfType" : 3, "buildId" : "78A49421FA60389F8C774BE68F5EF17DF2BD9CE3" }, { "b" : "7FCAEAA29000", "path" : "/usr/lib64/libnetsnmp.so.20", "elfType" : 3, "buildId" : "4CB6272BCAC2270393F559F67E8ED321690F79D5" }, { "path" : "/usr/lib64/libsasl2.so.2", "elfType" : 3, "buildId" : "E0AEE889D5BF1373F2F9EE0D448DBF3F5B5113F0" }, { "path" : "/lib64/libgssapi_krb5.so.2", "elfType" : 3, "buildId" : "0C72521270790A1BD52C8F6B989EEA5A575085BF" }, { "b" : "7FCAEA7BC000", "path" : "/usr/lib64/libssl.so.10", "elfType" : 3, "buildId" : "93610457BCF424BEBBF1F3FB44E51B51B50F2B55" }, { "b" : "7FCAEA3D8000", "path" : "/usr/lib64/libcrypto.so.10", "elfType" : 3, "buildId" : "06DDBB192AF74F99DB58F2150BFB83F42F5EBAD3" }, { "path" : "/lib64/librt.so.1", "elfType" : 3, "buildId" : "58C5A5FF5C82D7BE3113BE36DD87C7004E3C4DB1" }, { "path" : "/lib64/libdl.so.2", "elfType" : 3, "buildId" : "B5AE05CEDC0CE917F50A3A468CFA2ACD8592E8F6" }, { "path" : "/usr/lib64/libstdc++.so.6", "elfType" : 3, "buildId" : "28AF9321EBEA9D172CA43E11A60E02D0F7014870" }, { "path" : "/lib64/libm.so.6", "elfType" : 3, "buildId" : "989FE3A42CA8CEBDCC185A743896F23A0CF537ED" }, { "path" : "/lib64/libgcc_s.so.1", "elfType" : 3, "buildId" : "2AC15B051D1B8B53937E3341EA931D0E96F745D9" }, { "path" : "/lib64/libpthread.so.0", "elfType" : 3, "buildId" : "D467973C46E563CDCF64B5F12B2D6A50C7A25BA1" }, { "path" : "/lib64/libc.so.6", "elfType" : 3, "buildId" : "A6D15926E61580E250ED91F84FF7517F3970CD83" }, { "path" : "/lib64/ld-linux-x86-64.so.2", "elfType" : 3, "buildId" : "04202A4A8BE624D2193E812A25589E2DD02D5B5C" }, { "b" : "7FCAEA1CB000", "path" : "/lib64/libwrap.so.0", "elfType" : 3, "buildId" : "8C0C7CAB7F028E4592A8581EB2122FBECAB26B97" }, { "b" : "7F904765F000", "path" : "/usr/lib64/perl5/CORE/libperl.so", "elfType" : 3, "buildId" : "0A8E7D74369C1AF1F7C33B8DF8387DE5013898A4" }, { "path" : "/lib64/libresolv.so.2", "elfType" : 3, "buildId" : "F704FA7D21D05EF31E90FB4890FCA7F3D91DA138" }, { "path" : "/lib64/libnsl.so.1", "elfType" : 3, "buildId" : "CAD1498B2AA3531958C579F5CB39D8D6BFB5675B" }, { "path" : "/lib64/libcrypt.so.1", "elfType" : 3, "buildId" : "128802B73016BE233837EA9F2DCBC2153ACC2D6A" }, { "b" : "7F904845B000", "path" : "/lib64/libutil.so.1", "elfType" : 3, "buildId" : "565D9CDC6BD59EFE0156BAFE21033BE070F014DA" }, { "b" : "7F9042DEF000", "path" : "/usr/lib64/librpm.so.1", "elfType" : 3, "buildId" : "0B73153AA2E650B19153B7E8A57F9C7A965072CD" }, { "path" : "/usr/lib64/librpmio.so.1", "elfType" : 3, "buildId" : "7D821C87BEF03F9D7BBFE7FEE591EC5929D1C22C" }, { "b" : "7F9046BE6000", "path" : "/lib64/libpopt.so.0", "elfType" : 3, "buildId" : "E7B49911F1136073DD7DC58E8118CD9A4FBE2A19" }, { "b" : "7F90481CF000", "path" : "/lib64/libz.so.1", "elfType" : 3, "buildId" : "D053BB4FF0C2FC983842F81598813B9B931AD0D1" }, { "b" : "7FCAE93BF000", "path" : "/usr/lib64/libsensors.so.4", "elfType" : 3, "buildId" : "6855E5BF5B3634C15F01B1043BD892D727EE3C08" }, { "path" : "/lib64/libkrb5.so.3", "elfType" : 3, "buildId" : "DC11D5D89BDC77FF242481122D51E5A08DB60DA8" }, { "path" : "/lib64/libk5crypto.so.3", "elfType" : 3, "buildId" : "15782495E3AF093E67DDAE9A86436FFC6B3CC4D3" }, { "b" : "7F90469BA000", "path" : "/lib64/libcom_err.so.2", "elfType" : 3, "buildId" : "13FFCD68952B7715DDF34C9321D82E3041EA9006" }, { "b" : "7F90427AE000", "path" : "/lib64/libkrb5support.so.0", "elfType" : 3, "buildId" : "44A3A1C1891B4C8170C3DB80E7117A022E5EECD0" }, { "path" : "/lib64/libkeyutils.so.1", "elfType" : 3, "buildId" : "3BCCABE75DC61BBA81AAE45D164E26EF4F9F55DB" }, { "path" : "/lib64/libfreebl3.so", "elfType" : 3, "buildId" : "58BAC04A1DB3964A8F594EFFBE4838AD01214EDC" }, { "path" : "/usr/lib64/libnss3.so", "elfType" : 3, "buildId" : "A719876DB720919EA694995B0CB4E703E78F561F" }, { "b" : "7F904599C000", "path" : "/lib64/libbz2.so.1", "elfType" : 3, "buildId" : "1250B1D041DD7552F0C870BB188DC3A34DF2651D" }, { "b" : "7F9045385000", "path" : "/usr/lib64/libelf.so.1", "elfType" : 3, "buildId" : "1C2B39A5003E9DA8FD9C55972C06245E731E6546" }, { "path" : "/usr/lib64/liblzma.so.0", "elfType" : 3, "buildId" : "6FF9BAEEEE9DDEEF2DFA5CBD36147A75891C0AD4" }, { "b" : "7F9042558000", "path" : "/usr/lib64/liblua-5.1.so", "elfType" : 3, "buildId" : "6BDB4E1990D6EBA12A5C8D39A7650DB8798BF568" }, { "b" : "7F9046738000", "path" : "/lib64/libselinux.so.1", "elfType" : 3, "buildId" : "2D0F26E648D9661ABD83ED8B4BBE8F2CFA50393B" }, { "path" : "/lib64/libcap.so.2", "elfType" : 3, "buildId" : "A436538388F1F25113FDA834CA2EED524EFA17D6" }, { "b" : "7F9042930000", "path" : "/lib64/libacl.so.1", "elfType" : 3, "buildId" : "26CC708AC7C0FC1797A2340C024F0ADD0CE054D8" }, { "path" : "/lib64/libdb-4.7.so", "elfType" : 3, "buildId" : "437CA0AB593A7383FF1A1700D14AF4998FE93CF3" }, { "path" : "/usr/lib64/libnssutil3.so", "elfType" : 3, "buildId" : "3D18834CC92D576DCB1CD0F44BA62D3BFFFD52B7" }, { "path" : "/lib64/libplc4.so", "elfType" : 3, "buildId" : "C53F8B39797A277F40F582D8D11D3C2FFF7E5D1E" }, { "path" : "/lib64/libplds4.so", "elfType" : 3, "buildId" : "97F07716D324E086D43CC4D05873E1A16E020468" }, { "path" : "/lib64/libnspr4.so", "elfType" : 3, "buildId" : "7CD7DD1B6C294C61F494519CE3E0D7E114DFB36D" }, { "b" : "7F9044728000", "path" : "/lib64/libattr.so.1", "elfType" : 3, "buildId" : "8EF0683858704EF173AB11B1E27076F37F82B7B6" }, { "b" : "7FCAE811A000", "path" : "/usr/lib64/sasl2/libsasldb.so", "elfType" : 3, "buildId" : "4514552B5354286A143770420B38F2D5985D7FA1" }, { "b" : "7FCAE7F15000", "path" : "/usr/lib64/sasl2/libanonymous.so", "elfType" : 3, "buildId" : "EEAA33A75735D35F4BF25C3C2830B8C90ABDD8B5" }, { "b" : "7FCAE7D08000", "path" : "/usr/lib64/sasl2/libdigestmd5.so", "elfType" : 3, "buildId" : "34D8E3E2565DEF4A685D6976831B0372AD456993" }, { "b" : "7FCAE7B02000", "path" : "/usr/lib64/sasl2/libcrammd5.so", "elfType" : 3, "buildId" : "4CC7E695963F5C8B772EDFF456DB67F89E58FBD6" }, { "b" : "7FCAE78FD000", "path" : "/usr/lib64/sasl2/libplain.so", "elfType" : 3, "buildId" : "F8DDC7A3CA1CE5B75719AE0DC821647B609D17B6" }, { "b" : "7FCAE76F8000", "path" : "/usr/lib64/sasl2/liblogin.so", "elfType" : 3, "buildId" : "9D19F93E342AA4EE2D646E64642625F365056E5C" }, { "b" : "7FCAE74F0000", "path" : "/usr/lib64/sasl2/libgssapiv2.so", "elfType" : 3, "buildId" : "F7BCE9C6BFF4EAF0CB3142B299CF22D094CE4F04" } ] }} [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.822+0000 d20013| mongod(_ZN5mongo15printStackTraceERSo+0x32) [0x137f942] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.823+0000 d20013| mongod(+0xF7E889) [0x137e889] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.823+0000 d20013| mongod(+0xF7F092) [0x137f092] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.823+0000 d20013| libpthread.so.0(+0xF790) [0x3aa100f790] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.823+0000 d20013| libc.so.6(gsignal+0x35) [0x3aa0c32625] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.823+0000 d20013| libc.so.6(abort+0x175) [0x3aa0c33e05] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.823+0000 d20013| mongod(_ZN5mongo15invariantFailedEPKcS1_j+0xCB) [0x130774b] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.824+0000 d20013| mongod(_ZN5mongo21LegacyDistLockManager6unlockEPNS_16OperationContextERKNS_3OIDE+0x320) [0x11c1590] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.824+0000 d20013| mongod(_ZN5mongo24ForwardingCatalogManager14ScopedDistLockD1Ev+0x48) [0x118fd18] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.824+0000 d20013| mongod(_ZN5mongo10StatusWithINS_24ForwardingCatalogManager14ScopedDistLockEED1Ev+0x1B) [0xfe6d8b] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.824+0000 d20013| mongod(_ZN5mongo17SplitChunkCommand3runEPNS_16OperationContextERKSsRNS_7BSONObjEiRSsRNS_14BSONObjBuilderE+0x2CC7) [0x126b3f7] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.825+0000 d20013| mongod(_ZN5mongo7Command3runEPNS_16OperationContextERKNS_3rpc16RequestInterfaceEPNS3_21ReplyBuilderInterfaceE+0x40A) [0xba4eea] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.825+0000 d20013| mongod(_ZN5mongo7Command11execCommandEPNS_16OperationContextEPS0_RKNS_3rpc16RequestInterfaceEPNS4_21ReplyBuilderInterfaceE+0x3E6) [0xba5b86] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.825+0000 d20013| mongod(_ZN5mongo11runCommandsEPNS_16OperationContextERKNS_3rpc16RequestInterfaceEPNS2_21ReplyBuilderInterfaceE+0x1F0) [0xaff3d0] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.826+0000 d20013| mongod(_ZN5mongo16assembleResponseEPNS_16OperationContextERNS_7MessageERNS_10DbResponseERKNS_11HostAndPortE+0xC2D) [0xcb921d] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.826+0000 d20013| mongod(_ZN5mongo16MyMessageHandler7processERNS_7MessageEPNS_21AbstractMessagingPortE+0xEC) [0x98a86c] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.826+0000 d20013| mongod(_ZN5mongo17PortMessageServer17handleIncomingMsgEPv+0x26D) [0x132b9ed] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.826+0000 d20013| libpthread.so.0(+0x7A51) [0x3aa1007a51] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.826+0000 d20013| libc.so.6(clone+0x6D) [0x3aa0ce893d] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.826+0000 d20013| ----- END BACKTRACE ----- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:12.827+0000 s20020| 2015-11-23T18:45:12.756+0000 I NETWORK [conn57] end connection 10.99.163.247:60578 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.801+0000 d20015| 2015-11-23T18:45:16.801+0000 I NETWORK [conn3] end connection 10.99.163.247:56527 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.802+0000 d20015| 2015-11-23T18:45:16.801+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.803+0000 d20014| 2015-11-23T18:45:16.801+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.803+0000 d20014| 2015-11-23T18:45:16.801+0000 I NETWORK [conn6] end connection 10.99.163.247:37214 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.803+0000 d20014| 2015-11-23T18:45:16.801+0000 I NETWORK [conn7] end connection 10.99.163.247:37215 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.804+0000 2015-11-23T18:45:16.802+0000 I NETWORK [ReplicaSetMonitorWatcher] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.804+0000 2015-11-23T18:45:16.802+0000 I NETWORK [ReplicaSetMonitorWatcher] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20013] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.805+0000 2015-11-23T18:45:16.802+0000 I NETWORK [ReplicaSetMonitorWatcher] Detected bad connection created at 1448304273738988 microSec, clearing pool for ip-10-99-163-247:20013 of 0 connections [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.806+0000 d20014| 2015-11-23T18:45:16.802+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.806+0000 s20019| 2015-11-23T18:45:16.802+0000 I NETWORK [ReplicaSetMonitorWatcher] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.807+0000 s20019| 2015-11-23T18:45:16.802+0000 I NETWORK [ReplicaSetMonitorWatcher] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20013] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.808+0000 s20019| 2015-11-23T18:45:16.802+0000 I NETWORK [ReplicaSetMonitorWatcher] Detected bad connection created at 1448304275230317 microSec, clearing pool for ip-10-99-163-247:20013 of 0 connections [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.809+0000 d20014| 2015-11-23T18:45:16.803+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.809+0000 2015-11-23T18:45:16.803+0000 W NETWORK [ReplicaSetMonitorWatcher] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.810+0000 s20019| 2015-11-23T18:45:16.803+0000 W NETWORK [ReplicaSetMonitorWatcher] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.811+0000 c20016| 2015-11-23T18:45:16.804+0000 I NETWORK [conn35] end connection 10.99.163.247:48442 (52 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.811+0000 c20017| 2015-11-23T18:45:16.804+0000 I NETWORK [conn33] end connection 10.99.163.247:55757 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.811+0000 c20018| 2015-11-23T18:45:16.804+0000 I NETWORK [conn33] end connection 10.99.163.247:36878 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.812+0000 d20015| 2015-11-23T18:45:16.802+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.812+0000 d20015| 2015-11-23T18:45:16.803+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.813+0000 d20015| 2015-11-23T18:45:16.805+0000 I NETWORK [SyncSourceFeedback] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.814+0000 d20015| 2015-11-23T18:45:16.805+0000 I NETWORK [SyncSourceFeedback] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20013] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.816+0000 d20015| 2015-11-23T18:45:16.805+0000 I REPL [SyncSourceFeedback] SyncSourceFeedback error sending update: network error while attempting to run command 'replSetUpdatePosition' on host 'ip-10-99-163-247:20013' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.817+0000 d20014| 2015-11-23T18:45:16.805+0000 I NETWORK [SyncSourceFeedback] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.819+0000 d20015| 2015-11-23T18:45:16.805+0000 I REPL [SyncSourceFeedback] updateUpstream failed: HostUnreachable network error while attempting to run command 'replSetUpdatePosition' on host 'ip-10-99-163-247:20013' , will retry [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.819+0000 d20014| 2015-11-23T18:45:16.805+0000 I NETWORK [SyncSourceFeedback] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20013] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.820+0000 d20010| 2015-11-23T18:45:16.805+0000 I NETWORK [ReplicaSetMonitorWatcher] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.820+0000 d20010| 2015-11-23T18:45:16.805+0000 I NETWORK [ReplicaSetMonitorWatcher] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20013] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.821+0000 d20010| 2015-11-23T18:45:16.805+0000 I NETWORK [ReplicaSetMonitorWatcher] Detected bad connection created at 1448304277111617 microSec, clearing pool for ip-10-99-163-247:20013 of 0 connections [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.822+0000 d20012| 2015-11-23T18:45:16.805+0000 I NETWORK [conn9] end connection 10.99.163.247:46070 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.822+0000 d20014| 2015-11-23T18:45:16.805+0000 I REPL [SyncSourceFeedback] SyncSourceFeedback error sending update: network error while attempting to run command 'replSetUpdatePosition' on host 'ip-10-99-163-247:20013' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.823+0000 d20010| 2015-11-23T18:45:16.806+0000 I NETWORK [conn49] end connection 10.99.163.247:38871 (71 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.823+0000 d20014| 2015-11-23T18:45:16.806+0000 I REPL [SyncSourceFeedback] updateUpstream failed: HostUnreachable network error while attempting to run command 'replSetUpdatePosition' on host 'ip-10-99-163-247:20013' , will retry [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.823+0000 d20010| 2015-11-23T18:45:16.806+0000 I NETWORK [conn50] end connection 10.99.163.247:38872 (70 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.824+0000 d20010| 2015-11-23T18:45:16.806+0000 W NETWORK [ReplicaSetMonitorWatcher] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.824+0000 c20016| 2015-11-23T18:45:16.806+0000 I NETWORK [conn36] end connection 10.99.163.247:48446 (51 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.824+0000 c20018| 2015-11-23T18:45:16.806+0000 I NETWORK [conn34] end connection 10.99.163.247:36882 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.824+0000 c20017| 2015-11-23T18:45:16.806+0000 I NETWORK [conn34] end connection 10.99.163.247:55761 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.825+0000 c20016| 2015-11-23T18:45:16.806+0000 I NETWORK [conn37] end connection 10.99.163.247:48449 (50 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.825+0000 c20016| 2015-11-23T18:45:16.807+0000 I NETWORK [conn38] end connection 10.99.163.247:48451 (49 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.825+0000 c20017| 2015-11-23T18:45:16.807+0000 I NETWORK [conn35] end connection 10.99.163.247:55764 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.825+0000 c20018| 2015-11-23T18:45:16.807+0000 I NETWORK [conn35] end connection 10.99.163.247:36887 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.826+0000 c20017| 2015-11-23T18:45:16.807+0000 I NETWORK [conn36] end connection 10.99.163.247:55766 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.826+0000 c20018| 2015-11-23T18:45:16.807+0000 I NETWORK [conn36] end connection 10.99.163.247:36888 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.826+0000 c20016| 2015-11-23T18:45:16.807+0000 I NETWORK [conn39] end connection 10.99.163.247:48455 (48 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.827+0000 c20017| 2015-11-23T18:45:16.807+0000 I NETWORK [conn37] end connection 10.99.163.247:55770 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.827+0000 c20018| 2015-11-23T18:45:16.807+0000 I NETWORK [conn37] end connection 10.99.163.247:36891 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.827+0000 s20020| 2015-11-23T18:45:16.808+0000 I NETWORK [ReplicaSetMonitorWatcher] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.827+0000 s20020| 2015-11-23T18:45:16.808+0000 I NETWORK [ReplicaSetMonitorWatcher] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20013] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.828+0000 s20020| 2015-11-23T18:45:16.808+0000 I NETWORK [ReplicaSetMonitorWatcher] Detected bad connection created at 1448304285126795 microSec, clearing pool for ip-10-99-163-247:20013 of 0 connections [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.828+0000 d20011| 2015-11-23T18:45:16.808+0000 I NETWORK [conn12] end connection 10.99.163.247:49875 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.828+0000 d20015| 2015-11-23T18:45:16.808+0000 I NETWORK [conn10] end connection 10.99.163.247:56763 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.829+0000 d20014| 2015-11-23T18:45:16.808+0000 I NETWORK [conn11] end connection 10.99.163.247:37436 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.829+0000 c20016| 2015-11-23T18:45:16.808+0000 I NETWORK [conn42] end connection 10.99.163.247:48523 (47 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.829+0000 s20020| 2015-11-23T18:45:16.808+0000 W NETWORK [ReplicaSetMonitorWatcher] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.829+0000 c20017| 2015-11-23T18:45:16.809+0000 I NETWORK [conn40] end connection 10.99.163.247:55838 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.830+0000 c20018| 2015-11-23T18:45:16.809+0000 I NETWORK [conn40] end connection 10.99.163.247:36959 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.830+0000 d20010| 2015-11-23T18:45:16.809+0000 I NETWORK [conn62] end connection 10.99.163.247:38923 (69 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.830+0000 d20010| 2015-11-23T18:45:16.809+0000 I NETWORK [conn63] end connection 10.99.163.247:38924 (69 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.830+0000 s20020| 2015-11-23T18:45:16.809+0000 I NETWORK [conn59] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.831+0000 s20020| 2015-11-23T18:45:16.809+0000 I NETWORK [conn59] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20013] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.831+0000 s20019| 2015-11-23T18:45:16.809+0000 I SHARDING [Balancer] caught exception while doing balance: Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.832+0000 s20019| 2015-11-23T18:45:16.810+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:45:16.810+0000-56535ebc6e3a936d394e6d05", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304316810), what: "balancer.round", ns: "", details: { executionTimeMillis: 1563, errorOccured: true, errmsg: "Connection reset by peer" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.832+0000 s20019| 2015-11-23T18:45:16.811+0000 W SHARDING [conn61] could not autosplit collection db10.coll10 :: caused by :: 6 network error while attempting to run command 'splitChunk' on host 'ip-10-99-163-247:20013' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.832+0000 d20010| 2015-11-23T18:45:16.812+0000 I NETWORK [conn64] end connection 10.99.163.247:38958 (67 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.833+0000 s20019| 2015-11-23T18:45:16.813+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.833+0000 s20019| 2015-11-23T18:45:16.813+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.833+0000 s20020| 2015-11-23T18:45:16.813+0000 I SHARDING [Balancer] caught exception while doing balance: Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.834+0000 s20020| 2015-11-23T18:45:16.813+0000 I SHARDING [Balancer] about to log metadata event into actionlog: { _id: "ip-10-99-163-247-2015-11-23T18:45:16.813+0000-56535ebc13c05a55d8fa6f96", server: "ip-10-99-163-247", clientAddr: "", time: new Date(1448304316813), what: "balancer.round", ns: "", details: { executionTimeMillis: 1483, errorOccured: true, errmsg: "Connection reset by peer" } } [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.834+0000 d20015| 2015-11-23T18:45:16.813+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.834+0000 d20014| 2015-11-23T18:45:16.814+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.835+0000 c20016| 2015-11-23T18:45:16.814+0000 I NETWORK [conn52] end connection 10.99.163.247:48706 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.835+0000 c20016| 2015-11-23T18:45:16.815+0000 I NETWORK [conn44] end connection 10.99.163.247:48582 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.835+0000 d20015| 2015-11-23T18:45:16.815+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.835+0000 d20014| 2015-11-23T18:45:16.815+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.836+0000 c20017| 2015-11-23T18:45:16.815+0000 I NETWORK [conn42] end connection 10.99.163.247:55897 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.836+0000 c20018| 2015-11-23T18:45:16.815+0000 I NETWORK [conn42] end connection 10.99.163.247:37018 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.836+0000 c20016| 2015-11-23T18:45:16.815+0000 I NETWORK [conn48] end connection 10.99.163.247:48701 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.836+0000 d20010| 2015-11-23T18:45:16.815+0000 I NETWORK [conn67] end connection 10.99.163.247:39017 (66 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.837+0000 d20014| 2015-11-23T18:45:16.815+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.837+0000 c20016| 2015-11-23T18:45:16.816+0000 I NETWORK [conn50] end connection 10.99.163.247:48704 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.837+0000 d20015| 2015-11-23T18:45:16.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.838+0000 c20016| 2015-11-23T18:45:16.816+0000 I NETWORK [conn51] end connection 10.99.163.247:48705 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.838+0000 c20016| 2015-11-23T18:45:16.816+0000 I NETWORK [conn49] end connection 10.99.163.247:48702 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.838+0000 d20014| 2015-11-23T18:45:16.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.838+0000 c20016| 2015-11-23T18:45:16.816+0000 I NETWORK [conn53] end connection 10.99.163.247:48723 (41 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.839+0000 s20020| 2015-11-23T18:45:16.816+0000 I NETWORK [conn59] end connection 10.99.163.247:60584 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.839+0000 c20018| 2015-11-23T18:45:16.816+0000 I NETWORK [conn47] end connection 10.99.163.247:37164 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.839+0000 c20018| 2015-11-23T18:45:16.816+0000 I NETWORK [conn48] end connection 10.99.163.247:37167 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.840+0000 d20015| 2015-11-23T18:45:16.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.840+0000 c20017| 2015-11-23T18:45:16.817+0000 I NETWORK [conn49] end connection 10.99.163.247:56038 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.840+0000 c20017| 2015-11-23T18:45:16.817+0000 I NETWORK [conn46] end connection 10.99.163.247:56026 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.840+0000 c20017| 2015-11-23T18:45:16.817+0000 I NETWORK [conn47] end connection 10.99.163.247:56028 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.841+0000 c20017| 2015-11-23T18:45:16.817+0000 I NETWORK [conn50] end connection 10.99.163.247:56039 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.841+0000 c20017| 2015-11-23T18:45:16.817+0000 I NETWORK [conn48] end connection 10.99.163.247:56033 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.841+0000 c20017| 2015-11-23T18:45:16.817+0000 I NETWORK [conn51] end connection 10.99.163.247:56046 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.841+0000 s20020| 2015-11-23T18:45:16.817+0000 W SHARDING [conn62] could not autosplit collection db10.coll10 :: caused by :: 6 network error while attempting to run command 'splitChunk' on host 'ip-10-99-163-247:20013' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.842+0000 c20016| 2015-11-23T18:45:16.817+0000 I NETWORK [conn56] end connection 10.99.163.247:48741 (39 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.842+0000 c20018| 2015-11-23T18:45:16.818+0000 I NETWORK [conn46] end connection 10.99.163.247:37162 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.842+0000 c20018| 2015-11-23T18:45:16.818+0000 I NETWORK [conn50] end connection 10.99.163.247:37172 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.842+0000 c20018| 2015-11-23T18:45:16.818+0000 I NETWORK [conn49] end connection 10.99.163.247:37171 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.843+0000 c20016| 2015-11-23T18:45:16.818+0000 I NETWORK [conn54] end connection 10.99.163.247:48735 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.843+0000 c20016| 2015-11-23T18:45:16.818+0000 I NETWORK [conn55] end connection 10.99.163.247:48739 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.843+0000 c20018| 2015-11-23T18:45:16.818+0000 I NETWORK [conn51] end connection 10.99.163.247:37174 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.844+0000 s20020| 2015-11-23T18:45:16.819+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:16.844+0000 s20020| 2015-11-23T18:45:16.820+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.315+0000 s20019| 2015-11-23T18:45:17.314+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.315+0000 s20019| 2015-11-23T18:45:17.315+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.321+0000 s20020| 2015-11-23T18:45:17.320+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.321+0000 s20020| 2015-11-23T18:45:17.321+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.817+0000 s20019| 2015-11-23T18:45:17.816+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.817+0000 s20019| 2015-11-23T18:45:17.816+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.822+0000 s20020| 2015-11-23T18:45:17.821+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:17.823+0000 s20020| 2015-11-23T18:45:17.822+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.318+0000 s20019| 2015-11-23T18:45:18.317+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.319+0000 s20019| 2015-11-23T18:45:18.318+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.324+0000 s20020| 2015-11-23T18:45:18.323+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.324+0000 s20020| 2015-11-23T18:45:18.323+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.816+0000 d20015| 2015-11-23T18:45:18.815+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.816+0000 d20014| 2015-11-23T18:45:18.815+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.817+0000 d20014| 2015-11-23T18:45:18.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.817+0000 d20015| 2015-11-23T18:45:18.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.817+0000 d20015| 2015-11-23T18:45:18.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.818+0000 d20014| 2015-11-23T18:45:18.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.819+0000 s20019| 2015-11-23T18:45:18.819+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.819+0000 s20019| 2015-11-23T18:45:18.819+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.824+0000 s20020| 2015-11-23T18:45:18.824+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:18.825+0000 s20020| 2015-11-23T18:45:18.824+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.321+0000 s20019| 2015-11-23T18:45:19.320+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.321+0000 s20019| 2015-11-23T18:45:19.321+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.326+0000 s20020| 2015-11-23T18:45:19.325+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.326+0000 s20020| 2015-11-23T18:45:19.325+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.822+0000 s20019| 2015-11-23T18:45:19.822+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.823+0000 s20019| 2015-11-23T18:45:19.822+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.827+0000 s20020| 2015-11-23T18:45:19.826+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:19.827+0000 s20020| 2015-11-23T18:45:19.827+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.324+0000 s20019| 2015-11-23T18:45:20.324+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.325+0000 s20019| 2015-11-23T18:45:20.325+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.328+0000 s20020| 2015-11-23T18:45:20.328+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.329+0000 s20020| 2015-11-23T18:45:20.328+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.816+0000 d20014| 2015-11-23T18:45:20.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.817+0000 d20015| 2015-11-23T18:45:20.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.817+0000 d20014| 2015-11-23T18:45:20.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.818+0000 d20014| 2015-11-23T18:45:20.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.818+0000 d20015| 2015-11-23T18:45:20.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.818+0000 d20015| 2015-11-23T18:45:20.818+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.826+0000 s20019| 2015-11-23T18:45:20.825+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.826+0000 s20019| 2015-11-23T18:45:20.826+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.830+0000 s20020| 2015-11-23T18:45:20.829+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:20.830+0000 s20020| 2015-11-23T18:45:20.829+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.328+0000 s20019| 2015-11-23T18:45:21.327+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.328+0000 s20019| 2015-11-23T18:45:21.328+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.330+0000 s20020| 2015-11-23T18:45:21.330+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.331+0000 s20020| 2015-11-23T18:45:21.331+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.829+0000 s20019| 2015-11-23T18:45:21.829+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.830+0000 s20019| 2015-11-23T18:45:21.829+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.832+0000 s20020| 2015-11-23T18:45:21.831+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:21.832+0000 s20020| 2015-11-23T18:45:21.832+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.331+0000 s20019| 2015-11-23T18:45:22.330+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.332+0000 s20019| 2015-11-23T18:45:22.331+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.333+0000 s20020| 2015-11-23T18:45:22.333+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.334+0000 s20020| 2015-11-23T18:45:22.333+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.817+0000 d20015| 2015-11-23T18:45:22.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.817+0000 d20014| 2015-11-23T18:45:22.816+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.818+0000 d20015| 2015-11-23T18:45:22.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.818+0000 d20014| 2015-11-23T18:45:22.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.819+0000 d20015| 2015-11-23T18:45:22.818+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.819+0000 d20014| 2015-11-23T18:45:22.818+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.833+0000 s20019| 2015-11-23T18:45:22.832+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.833+0000 s20019| 2015-11-23T18:45:22.833+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.834+0000 s20020| 2015-11-23T18:45:22.834+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.835+0000 s20020| 2015-11-23T18:45:22.835+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.837+0000 d20015| 2015-11-23T18:45:22.836+0000 I REPL [ReplicationExecutor] Starting an election, since we've seen no PRIMARY in the past 10000ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.838+0000 d20015| 2015-11-23T18:45:22.836+0000 I REPL [ReplicationExecutor] conducting a dry run election to see if we could be elected [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.839+0000 d20015| 2015-11-23T18:45:22.836+0000 I REPL [ReplicationExecutor] VoteRequester: Got failed response from ip-10-99-163-247:20013: HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.839+0000 d20015| 2015-11-23T18:45:22.837+0000 I REPL [ReplicationExecutor] dry election run succeeded, running for election [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.840+0000 d20015| 2015-11-23T18:45:22.837+0000 I REPL [ReplicationExecutor] election succeeded, assuming primary role in term 2 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.840+0000 d20015| 2015-11-23T18:45:22.837+0000 I REPL [ReplicationExecutor] transition to PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.840+0000 d20015| 2015-11-23T18:45:22.838+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.840+0000 d20015| 2015-11-23T18:45:22.838+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:22.841+0000 d20015| 2015-11-23T18:45:22.839+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.334+0000 s20019| 2015-11-23T18:45:23.334+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.335+0000 s20019| 2015-11-23T18:45:23.334+0000 W NETWORK [conn61] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.336+0000 s20020| 2015-11-23T18:45:23.335+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.336+0000 s20020| 2015-11-23T18:45:23.336+0000 W NETWORK [conn62] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.816+0000 d20015| 2015-11-23T18:45:23.816+0000 I REPL [rsSync] transition to primary complete; database writes are now permitted [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.836+0000 s20019| 2015-11-23T18:45:23.835+0000 W NETWORK [conn61] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.837+0000 d20015| 2015-11-23T18:45:23.836+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57089 #11 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.837+0000 s20020| 2015-11-23T18:45:23.836+0000 W NETWORK [conn62] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.838+0000 d20015| 2015-11-23T18:45:23.837+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57091 #12 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.838+0000 d20015| 2015-11-23T18:45:23.837+0000 I SHARDING [conn11] remote client 10.99.163.247:57089 initialized this host as shard test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.838+0000 d20015| 2015-11-23T18:45:23.838+0000 I SHARDING [ShardingState initialization] first cluster operation detected, adding sharding hook to enable versioning and authentication to remote servers [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.839+0000 d20015| 2015-11-23T18:45:23.838+0000 I SHARDING [ShardingState initialization] Updating config server connection string to: ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.840+0000 d20015| 2015-11-23T18:45:23.840+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.841+0000 c20016| 2015-11-23T18:45:23.840+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48808 #57 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.841+0000 d20015| 2015-11-23T18:45:23.841+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.842+0000 c20017| 2015-11-23T18:45:23.841+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56123 #52 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.843+0000 d20015| 2015-11-23T18:45:23.842+0000 I NETWORK [ShardingState initialization] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.843+0000 c20018| 2015-11-23T18:45:23.843+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37244 #52 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.844+0000 d20015| 2015-11-23T18:45:23.843+0000 I NETWORK [ShardingState initialization] Starting new replica set monitor for test-rs0/ip-10-99-163-247:20010,ip-10-99-163-247:20011,ip-10-99-163-247:20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.845+0000 d20015| 2015-11-23T18:45:23.843+0000 I NETWORK [ShardingState initialization] Starting new replica set monitor for test-rs1/ip-10-99-163-247:20013,ip-10-99-163-247:20014,ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.845+0000 d20015| 2015-11-23T18:45:23.843+0000 I NETWORK [ReplicaSetMonitorWatcher] starting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.846+0000 d20015| 2015-11-23T18:45:23.844+0000 I SHARDING [conn11] remotely refreshing metadata for db10.coll10 with requested shard version 1|3||56535eb86e3a936d394e6d03, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.847+0000 d20015| 2015-11-23T18:45:23.844+0000 I SHARDING [conn12] remotely refreshing metadata for db10.coll10 with requested shard version 1|3||56535eb86e3a936d394e6d03, current shard version is 0|0||000000000000000000000000, current metadata version is 0|0||000000000000000000000000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.847+0000 d20015| 2015-11-23T18:45:23.844+0000 I NETWORK [conn12] SyncClusterConnection connecting to [ip-10-99-163-247:20016] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.848+0000 c20016| 2015-11-23T18:45:23.844+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:48811 #58 (39 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.848+0000 d20015| 2015-11-23T18:45:23.845+0000 I SHARDING [conn11] collection db10.coll10 was previously unsharded, new metadata loaded with shard version 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.849+0000 d20015| 2015-11-23T18:45:23.845+0000 I SHARDING [conn11] collection version was loaded at version 1|3||56535eb86e3a936d394e6d03, took 0ms [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.849+0000 d20015| 2015-11-23T18:45:23.845+0000 I NETWORK [conn12] SyncClusterConnection connecting to [ip-10-99-163-247:20017] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.850+0000 c20017| 2015-11-23T18:45:23.845+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:56126 #53 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.850+0000 d20015| 2015-11-23T18:45:23.845+0000 I NETWORK [conn12] SyncClusterConnection connecting to [ip-10-99-163-247:20018] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.851+0000 c20018| 2015-11-23T18:45:23.846+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:37247 #53 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.852+0000 s20019| 2015-11-23T18:45:23.847+0000 I SHARDING [conn61] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 48 version: 1|3||56535eb86e3a936d394e6d03 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.852+0000 s20019| 2015-11-23T18:45:23.847+0000 I SHARDING [conn61] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 49 version: 1|3||56535eb86e3a936d394e6d03 based on: 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.853+0000 s20019| 2015-11-23T18:45:23.847+0000 W SHARDING [conn61] chunk manager reload forced for collection 'db10.coll10', config version is 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.854+0000 d20015| 2015-11-23T18:45:23.847+0000 I SHARDING [conn12] metadata of collection db10.coll10 already up to date (shard version : 1|3||56535eb86e3a936d394e6d03, took 3ms) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.854+0000 d20015| 2015-11-23T18:45:23.848+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57098 #13 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.855+0000 s20019| 2015-11-23T18:45:23.849+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.855+0000 s20020| 2015-11-23T18:45:23.849+0000 I SHARDING [conn62] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 15 version: 1|3||56535eb86e3a936d394e6d03 based on: (empty) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.856+0000 s20020| 2015-11-23T18:45:23.850+0000 I SHARDING [conn62] ChunkManager: time to load chunks for db10.coll10: 0ms sequenceNumber: 16 version: 1|3||56535eb86e3a936d394e6d03 based on: 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.856+0000 s20020| 2015-11-23T18:45:23.850+0000 W SHARDING [conn62] chunk manager reload forced for collection 'db10.coll10', config version is 1|3||56535eb86e3a936d394e6d03 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.857+0000 s20019| 2015-11-23T18:45:23.850+0000 I NETWORK [conn61] Socket closed remotely, no longer connected (idle 11 secs, remote host 10.99.163.247:20013) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.857+0000 d20015| 2015-11-23T18:45:23.851+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57099 #14 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.858+0000 d20015| 2015-11-23T18:45:23.851+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57100 #15 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.858+0000 s20020| 2015-11-23T18:45:23.851+0000 I ASIO [NetworkInterfaceASIO-ShardRegistry-1] Successfully connected to ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.859+0000 s20020| 2015-11-23T18:45:23.854+0000 I NETWORK [conn62] Socket closed remotely, no longer connected (idle 11 secs, remote host 10.99.163.247:20013) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.859+0000 d20015| 2015-11-23T18:45:23.854+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57101 #16 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.868+0000 s20019| 2015-11-23T18:45:23.868+0000 I NETWORK [conn61] end connection 10.99.163.247:40169 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.870+0000 s20020| 2015-11-23T18:45:23.869+0000 I NETWORK [conn62] end connection 10.99.163.247:60590 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.873+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.874+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.874+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.874+0000 Workload(s) completed in 11834 ms: jstests/concurrency/fsm_workloads/findAndModify_update.js [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.874+0000 ---- [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.875+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.875+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.876+0000 s20019| 2015-11-23T18:45:23.875+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.876+0000 s20019| 2015-11-23T18:45:23.875+0000 W SHARDING [LockPinger] removing distributed lock ping thread 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20019:1448304274:1804289383' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.878+0000 s20019| 2015-11-23T18:45:23.877+0000 I NETWORK [LockPinger] scoped connection to ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.878+0000 c20016| 2015-11-23T18:45:23.877+0000 I NETWORK [conn6] end connection 10.99.163.247:48290 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.878+0000 c20017| 2015-11-23T18:45:23.877+0000 I NETWORK [conn6] end connection 10.99.163.247:55605 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.879+0000 s20019| 2015-11-23T18:45:23.877+0000 I SHARDING [signalProcessingThread] dbexit: rc:0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.879+0000 c20018| 2015-11-23T18:45:23.877+0000 I NETWORK [conn6] end connection 10.99.163.247:36726 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.881+0000 c20017| 2015-11-23T18:45:23.880+0000 I NETWORK [conn3] end connection 10.99.163.247:55596 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.882+0000 c20017| 2015-11-23T18:45:23.880+0000 I NETWORK [conn4] end connection 10.99.163.247:55599 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.882+0000 c20018| 2015-11-23T18:45:23.880+0000 I NETWORK [conn3] end connection 10.99.163.247:36717 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.883+0000 c20018| 2015-11-23T18:45:23.880+0000 I NETWORK [conn7] end connection 10.99.163.247:36729 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.884+0000 c20018| 2015-11-23T18:45:23.880+0000 I NETWORK [conn4] end connection 10.99.163.247:36720 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.884+0000 c20016| 2015-11-23T18:45:23.880+0000 I NETWORK [conn8] end connection 10.99.163.247:48296 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.884+0000 c20016| 2015-11-23T18:45:23.880+0000 I NETWORK [conn5] end connection 10.99.163.247:48287 (36 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.885+0000 c20016| 2015-11-23T18:45:23.880+0000 I NETWORK [conn17] end connection 10.99.163.247:48330 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.885+0000 c20016| 2015-11-23T18:45:23.880+0000 I NETWORK [conn7] end connection 10.99.163.247:48293 (34 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.885+0000 c20016| 2015-11-23T18:45:23.881+0000 I NETWORK [conn3] end connection 10.99.163.247:48281 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.886+0000 c20016| 2015-11-23T18:45:23.881+0000 I NETWORK [conn9] end connection 10.99.163.247:48300 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.886+0000 c20016| 2015-11-23T18:45:23.881+0000 I NETWORK [conn4] end connection 10.99.163.247:48284 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.886+0000 c20018| 2015-11-23T18:45:23.881+0000 I NETWORK [conn17] end connection 10.99.163.247:36766 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.887+0000 d20010| 2015-11-23T18:45:23.880+0000 I NETWORK [conn16] end connection 10.99.163.247:38761 (65 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.887+0000 d20010| 2015-11-23T18:45:23.880+0000 I NETWORK [conn48] end connection 10.99.163.247:38865 (65 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.887+0000 d20010| 2015-11-23T18:45:23.880+0000 I NETWORK [conn20] end connection 10.99.163.247:38778 (64 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.887+0000 d20010| 2015-11-23T18:45:23.881+0000 I NETWORK [conn19] end connection 10.99.163.247:38777 (62 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.888+0000 c20018| 2015-11-23T18:45:23.881+0000 I NETWORK [conn5] end connection 10.99.163.247:36723 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.888+0000 c20017| 2015-11-23T18:45:23.880+0000 I NETWORK [conn9] end connection 10.99.163.247:55615 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.888+0000 c20017| 2015-11-23T18:45:23.880+0000 I NETWORK [conn5] end connection 10.99.163.247:55602 (32 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.888+0000 c20017| 2015-11-23T18:45:23.881+0000 I NETWORK [conn17] end connection 10.99.163.247:55645 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.889+0000 c20017| 2015-11-23T18:45:23.881+0000 I NETWORK [conn7] end connection 10.99.163.247:55608 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.889+0000 c20017| 2015-11-23T18:45:23.881+0000 I NETWORK [conn8] end connection 10.99.163.247:55611 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.889+0000 d20010| 2015-11-23T18:45:23.881+0000 I NETWORK [conn17] end connection 10.99.163.247:38768 (61 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.889+0000 d20010| 2015-11-23T18:45:23.881+0000 I NETWORK [conn18] end connection 10.99.163.247:38776 (60 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.890+0000 c20018| 2015-11-23T18:45:23.881+0000 I NETWORK [conn8] end connection 10.99.163.247:36732 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.890+0000 c20018| 2015-11-23T18:45:23.882+0000 I NETWORK [conn9] end connection 10.99.163.247:36736 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.890+0000 d20015| 2015-11-23T18:45:23.881+0000 I NETWORK [conn11] end connection 10.99.163.247:57089 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.890+0000 d20015| 2015-11-23T18:45:23.882+0000 I NETWORK [conn7] end connection 10.99.163.247:56617 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.891+0000 d20012| 2015-11-23T18:45:23.882+0000 I NETWORK [conn7] end connection 10.99.163.247:45959 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.891+0000 d20010| 2015-11-23T18:45:23.882+0000 I NETWORK [conn15] end connection 10.99.163.247:38760 (59 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.891+0000 d20010| 2015-11-23T18:45:23.882+0000 I NETWORK [conn21] end connection 10.99.163.247:38795 (58 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.891+0000 d20010| 2015-11-23T18:45:23.882+0000 I NETWORK [conn22] end connection 10.99.163.247:38796 (58 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.892+0000 d20010| 2015-11-23T18:45:23.882+0000 I NETWORK [conn24] end connection 10.99.163.247:38799 (58 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.892+0000 d20010| 2015-11-23T18:45:23.882+0000 I NETWORK [conn30] end connection 10.99.163.247:38806 (55 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.892+0000 d20010| 2015-11-23T18:45:23.882+0000 I NETWORK [conn26] end connection 10.99.163.247:38802 (55 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.892+0000 d20010| 2015-11-23T18:45:23.883+0000 I NETWORK [conn42] end connection 10.99.163.247:38818 (53 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.893+0000 d20010| 2015-11-23T18:45:23.883+0000 I NETWORK [conn43] end connection 10.99.163.247:38819 (52 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.893+0000 d20010| 2015-11-23T18:45:23.883+0000 I NETWORK [conn45] end connection 10.99.163.247:38823 (52 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.893+0000 d20010| 2015-11-23T18:45:23.883+0000 I NETWORK [conn46] end connection 10.99.163.247:38857 (50 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.893+0000 d20010| 2015-11-23T18:45:23.883+0000 I NETWORK [conn47] end connection 10.99.163.247:38858 (49 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.894+0000 d20010| 2015-11-23T18:45:23.883+0000 I NETWORK [conn55] end connection 10.99.163.247:38899 (49 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.894+0000 d20010| 2015-11-23T18:45:23.883+0000 I NETWORK [conn57] end connection 10.99.163.247:38901 (48 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.894+0000 d20010| 2015-11-23T18:45:23.884+0000 I NETWORK [conn56] end connection 10.99.163.247:38900 (46 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.894+0000 d20010| 2015-11-23T18:45:23.884+0000 I NETWORK [conn58] end connection 10.99.163.247:38902 (45 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.895+0000 d20014| 2015-11-23T18:45:23.884+0000 I NETWORK [conn9] end connection 10.99.163.247:37428 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.895+0000 d20015| 2015-11-23T18:45:23.884+0000 I NETWORK [conn13] end connection 10.99.163.247:57098 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.895+0000 d20011| 2015-11-23T18:45:23.884+0000 I NETWORK [conn10] end connection 10.99.163.247:49867 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.895+0000 d20010| 2015-11-23T18:45:23.884+0000 I NETWORK [conn65] end connection 10.99.163.247:38969 (44 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.896+0000 d20015| 2015-11-23T18:45:23.884+0000 I NETWORK [conn15] end connection 10.99.163.247:57100 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.896+0000 d20010| 2015-11-23T18:45:23.885+0000 I NETWORK [conn66] end connection 10.99.163.247:38970 (43 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.896+0000 d20010| 2015-11-23T18:45:23.885+0000 I NETWORK [conn73] end connection 10.99.163.247:39073 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.896+0000 d20010| 2015-11-23T18:45:23.885+0000 I NETWORK [conn75] end connection 10.99.163.247:39075 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.897+0000 d20010| 2015-11-23T18:45:23.885+0000 I NETWORK [conn76] end connection 10.99.163.247:39076 (42 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:23.897+0000 d20010| 2015-11-23T18:45:23.885+0000 I NETWORK [conn74] end connection 10.99.163.247:39074 (40 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.817+0000 d20014| 2015-11-23T18:45:24.816+0000 I REPL [ReplicationExecutor] Member ip-10-99-163-247:20015 is now in state PRIMARY [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.818+0000 d20014| 2015-11-23T18:45:24.817+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.818+0000 d20014| 2015-11-23T18:45:24.818+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.818+0000 d20014| 2015-11-23T18:45:24.818+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.840+0000 d20015| 2015-11-23T18:45:24.839+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.840+0000 d20015| 2015-11-23T18:45:24.840+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.841+0000 d20015| 2015-11-23T18:45:24.840+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.875+0000 2015-11-23T18:45:24.875+0000 I - [thread1] shell: stopped mongo program on port 20019 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.876+0000 s20020| 2015-11-23T18:45:24.875+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.877+0000 s20020| 2015-11-23T18:45:24.875+0000 W SHARDING [LockPinger] removing distributed lock ping thread 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20020:1448304275:1804289383' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.879+0000 s20020| 2015-11-23T18:45:24.878+0000 I NETWORK [LockPinger] scoped connection to ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018 not being returned to the pool [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.879+0000 c20017| 2015-11-23T18:45:24.878+0000 I NETWORK [conn44] end connection 10.99.163.247:55928 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.879+0000 c20018| 2015-11-23T18:45:24.878+0000 I NETWORK [conn44] end connection 10.99.163.247:37050 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.879+0000 c20016| 2015-11-23T18:45:24.878+0000 I NETWORK [conn47] end connection 10.99.163.247:48610 (30 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.880+0000 s20020| 2015-11-23T18:45:24.878+0000 I SHARDING [signalProcessingThread] dbexit: rc:0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.884+0000 c20016| 2015-11-23T18:45:24.881+0000 I NETWORK [conn10] end connection 10.99.163.247:48305 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.885+0000 c20016| 2015-11-23T18:45:24.881+0000 I NETWORK [conn16] end connection 10.99.163.247:48323 (29 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.885+0000 c20016| 2015-11-23T18:45:24.881+0000 I NETWORK [conn13] end connection 10.99.163.247:48314 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.885+0000 c20016| 2015-11-23T18:45:24.881+0000 I NETWORK [conn14] end connection 10.99.163.247:48317 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.886+0000 c20016| 2015-11-23T18:45:24.881+0000 I NETWORK [conn15] end connection 10.99.163.247:48320 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.886+0000 c20016| 2015-11-23T18:45:24.882+0000 I NETWORK [conn12] end connection 10.99.163.247:48311 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.886+0000 c20016| 2015-11-23T18:45:24.883+0000 I NETWORK [conn41] end connection 10.99.163.247:48514 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.887+0000 d20011| 2015-11-23T18:45:24.883+0000 I NETWORK [conn8] end connection 10.99.163.247:49758 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.887+0000 c20017| 2015-11-23T18:45:24.881+0000 I NETWORK [conn10] end connection 10.99.163.247:55620 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.887+0000 c20017| 2015-11-23T18:45:24.881+0000 I NETWORK [conn16] end connection 10.99.163.247:55638 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.887+0000 c20017| 2015-11-23T18:45:24.881+0000 I NETWORK [conn13] end connection 10.99.163.247:55629 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.888+0000 c20017| 2015-11-23T18:45:24.882+0000 I NETWORK [conn15] end connection 10.99.163.247:55635 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.888+0000 c20017| 2015-11-23T18:45:24.882+0000 I NETWORK [conn39] end connection 10.99.163.247:55829 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.888+0000 c20017| 2015-11-23T18:45:24.882+0000 I NETWORK [conn12] end connection 10.99.163.247:55626 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.888+0000 c20017| 2015-11-23T18:45:24.883+0000 I NETWORK [conn14] end connection 10.99.163.247:55632 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.889+0000 d20014| 2015-11-23T18:45:24.882+0000 I NETWORK [conn12] end connection 10.99.163.247:37438 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.889+0000 d20015| 2015-11-23T18:45:24.881+0000 I NETWORK [conn12] end connection 10.99.163.247:57091 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.889+0000 d20015| 2015-11-23T18:45:24.882+0000 I NETWORK [conn9] end connection 10.99.163.247:56757 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.889+0000 d20015| 2015-11-23T18:45:24.883+0000 I NETWORK [conn16] end connection 10.99.163.247:57101 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.890+0000 d20015| 2015-11-23T18:45:24.883+0000 I NETWORK [conn14] end connection 10.99.163.247:57099 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.890+0000 c20018| 2015-11-23T18:45:24.881+0000 I NETWORK [conn14] end connection 10.99.163.247:36753 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.890+0000 c20018| 2015-11-23T18:45:24.881+0000 I NETWORK [conn16] end connection 10.99.163.247:36759 (27 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.890+0000 c20018| 2015-11-23T18:45:24.881+0000 I NETWORK [conn12] end connection 10.99.163.247:36747 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.891+0000 c20018| 2015-11-23T18:45:24.882+0000 I NETWORK [conn15] end connection 10.99.163.247:36756 (24 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.891+0000 c20018| 2015-11-23T18:45:24.882+0000 I NETWORK [conn10] end connection 10.99.163.247:36741 (23 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.891+0000 c20018| 2015-11-23T18:45:24.882+0000 I NETWORK [conn13] end connection 10.99.163.247:36750 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.891+0000 d20010| 2015-11-23T18:45:24.881+0000 I NETWORK [conn28] end connection 10.99.163.247:38804 (38 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.892+0000 d20010| 2015-11-23T18:45:24.881+0000 I NETWORK [conn25] end connection 10.99.163.247:38801 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.892+0000 d20010| 2015-11-23T18:45:24.881+0000 I NETWORK [conn29] end connection 10.99.163.247:38805 (37 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.892+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn37] end connection 10.99.163.247:38813 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.892+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn35] end connection 10.99.163.247:38811 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.893+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn39] end connection 10.99.163.247:38815 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.893+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn38] end connection 10.99.163.247:38814 (33 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.893+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn27] end connection 10.99.163.247:38803 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.893+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn40] end connection 10.99.163.247:38816 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.894+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn36] end connection 10.99.163.247:38812 (31 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.894+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn51] end connection 10.99.163.247:38895 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.894+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn44] end connection 10.99.163.247:38821 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.894+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn41] end connection 10.99.163.247:38817 (28 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.895+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn54] end connection 10.99.163.247:38898 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.895+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn32] end connection 10.99.163.247:38808 (35 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.895+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn53] end connection 10.99.163.247:38897 (25 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.896+0000 d20010| 2015-11-23T18:45:24.882+0000 I NETWORK [conn52] end connection 10.99.163.247:38896 (26 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.896+0000 d20010| 2015-11-23T18:45:24.883+0000 I NETWORK [conn34] end connection 10.99.163.247:38810 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.896+0000 d20012| 2015-11-23T18:45:24.883+0000 I NETWORK [conn8] end connection 10.99.163.247:45998 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.896+0000 d20010| 2015-11-23T18:45:24.883+0000 I NETWORK [conn23] end connection 10.99.163.247:38797 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.897+0000 d20010| 2015-11-23T18:45:24.883+0000 I NETWORK [conn33] end connection 10.99.163.247:38809 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.897+0000 d20010| 2015-11-23T18:45:24.883+0000 I NETWORK [conn31] end connection 10.99.163.247:38807 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.897+0000 c20018| 2015-11-23T18:45:24.884+0000 I NETWORK [conn39] end connection 10.99.163.247:36950 (21 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.897+0000 d20010| 2015-11-23T18:45:24.884+0000 I NETWORK [conn77] end connection 10.99.163.247:39083 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.898+0000 d20010| 2015-11-23T18:45:24.884+0000 I NETWORK [conn78] end connection 10.99.163.247:39084 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.898+0000 c20016| 2015-11-23T18:45:24.884+0000 I NETWORK [conn46] end connection 10.99.163.247:48609 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.898+0000 c20016| 2015-11-23T18:45:24.884+0000 I NETWORK [conn45] end connection 10.99.163.247:48606 (22 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.898+0000 d20010| 2015-11-23T18:45:24.884+0000 I NETWORK [conn68] end connection 10.99.163.247:39039 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.899+0000 d20010| 2015-11-23T18:45:24.885+0000 I NETWORK [conn71] end connection 10.99.163.247:39045 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.899+0000 d20010| 2015-11-23T18:45:24.885+0000 I NETWORK [conn69] end connection 10.99.163.247:39043 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.899+0000 d20010| 2015-11-23T18:45:24.885+0000 I NETWORK [conn70] end connection 10.99.163.247:39044 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.899+0000 c20017| 2015-11-23T18:45:24.885+0000 I NETWORK [conn43] end connection 10.99.163.247:55922 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.900+0000 c20018| 2015-11-23T18:45:24.885+0000 I NETWORK [conn43] end connection 10.99.163.247:37049 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.900+0000 c20018| 2015-11-23T18:45:24.885+0000 I NETWORK [conn45] end connection 10.99.163.247:37052 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.900+0000 c20017| 2015-11-23T18:45:24.885+0000 I NETWORK [conn45] end connection 10.99.163.247:55931 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:24.900+0000 d20010| 2015-11-23T18:45:24.885+0000 I NETWORK [conn72] end connection 10.99.163.247:39072 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.817+0000 d20014| 2015-11-23T18:45:25.816+0000 I REPL [ReplicationExecutor] syncing from: ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.817+0000 d20015| 2015-11-23T18:45:25.817+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57109 #17 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.818+0000 d20014| 2015-11-23T18:45:25.818+0000 I REPL [SyncSourceFeedback] setting syncSourceFeedback to ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.819+0000 d20015| 2015-11-23T18:45:25.818+0000 I NETWORK [conn17] end connection 10.99.163.247:57109 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.819+0000 d20015| 2015-11-23T18:45:25.818+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57110 #18 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.819+0000 d20015| 2015-11-23T18:45:25.818+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57111 #19 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.819+0000 d20014| 2015-11-23T18:45:25.819+0000 I ASIO [NetworkInterfaceASIO-BGSync-1] Successfully connected to ip-10-99-163-247:20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.877+0000 2015-11-23T18:45:25.875+0000 I - [thread1] shell: stopped mongo program on port 20020 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.878+0000 ReplSetTest stop *** Shutting down mongod in port 20010 *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.879+0000 d20010| 2015-11-23T18:45:25.875+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.879+0000 d20010| 2015-11-23T18:45:25.875+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:25.879+0000 d20010| 2015-11-23T18:45:25.879+0000 I REPL [signalProcessingThread] Stopping replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.806+0000 2015-11-23T18:45:26.805+0000 W NETWORK [ReplicaSetMonitorWatcher] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.807+0000 2015-11-23T18:45:26.806+0000 I NETWORK [ReplicaSetMonitorWatcher] Detected bad connection created at 1448304273728110 microSec, clearing pool for ip-10-99-163-247:20010 of 0 connections [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.807+0000 d20010| 2015-11-23T18:45:26.806+0000 I STORAGE [conn14] got request after shutdown() [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.808+0000 2015-11-23T18:45:26.807+0000 W NETWORK [ReplicaSetMonitorWatcher] No primary detected for set test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.841+0000 d20015| 2015-11-23T18:45:26.840+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.841+0000 d20015| 2015-11-23T18:45:26.841+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.842+0000 d20015| 2015-11-23T18:45:26.842+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.897+0000 d20010| 2015-11-23T18:45:26.897+0000 I STORAGE [conn3] got request after shutdown() [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.898+0000 d20011| 2015-11-23T18:45:26.897+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable End of file [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.903+0000 d20010| 2015-11-23T18:45:26.903+0000 I STORAGE [conn2] got request after shutdown() [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.903+0000 d20012| 2015-11-23T18:45:26.903+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable End of file [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.928+0000 d20010| 2015-11-23T18:45:26.928+0000 W SHARDING [LockPinger] removing distributed lock ping thread 'ip-10-99-163-247:20016,ip-10-99-163-247:20017,ip-10-99-163-247:20018/ip-10-99-163-247:20010:1448304275:1756418349' [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.929+0000 d20010| 2015-11-23T18:45:26.928+0000 W SHARDING [LockPinger] Error encountered while stopping ping on ip-10-99-163-247:20010:1448304275:1756418349 :: caused by :: 17382 Can't use connection pool during shutdown [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.930+0000 d20011| 2015-11-23T18:45:26.929+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.931+0000 d20012| 2015-11-23T18:45:26.929+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.932+0000 d20010| 2015-11-23T18:45:26.929+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.932+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.933+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [signalProcessingThread] closing listening socket: 9 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.933+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [signalProcessingThread] closing listening socket: 10 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.934+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20010.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.935+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.935+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.936+0000 d20010| 2015-11-23T18:45:26.929+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.936+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [conn1] end connection 127.0.0.1:34887 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.937+0000 d20010| 2015-11-23T18:45:26.929+0000 I NETWORK [conn10] end connection 10.99.163.247:38680 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.937+0000 d20010| 2015-11-23T18:45:26.930+0000 I NETWORK [conn12] end connection 10.99.163.247:38682 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.937+0000 d20011| 2015-11-23T18:45:26.930+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.937+0000 d20012| 2015-11-23T18:45:26.930+0000 I REPL [ReplicationExecutor] could not find member to sync from [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.938+0000 c20017| 2015-11-23T18:45:26.930+0000 I NETWORK [conn41] end connection 10.99.163.247:55842 (18 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.938+0000 c20018| 2015-11-23T18:45:26.930+0000 I NETWORK [conn41] end connection 10.99.163.247:36963 (18 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.938+0000 c20016| 2015-11-23T18:45:26.930+0000 I NETWORK [conn43] end connection 10.99.163.247:48527 (20 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.938+0000 c20017| 2015-11-23T18:45:26.931+0000 I NETWORK [conn21] end connection 10.99.163.247:55713 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.939+0000 d20012| 2015-11-23T18:45:26.931+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.939+0000 c20017| 2015-11-23T18:45:26.931+0000 I NETWORK [conn27] end connection 10.99.163.247:55723 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.939+0000 d20011| 2015-11-23T18:45:26.931+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.940+0000 c20017| 2015-11-23T18:45:26.931+0000 I NETWORK [conn23] end connection 10.99.163.247:55716 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.940+0000 c20016| 2015-11-23T18:45:26.931+0000 I NETWORK [conn21] end connection 10.99.163.247:48393 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.940+0000 c20016| 2015-11-23T18:45:26.931+0000 I NETWORK [conn23] end connection 10.99.163.247:48395 (19 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.940+0000 c20016| 2015-11-23T18:45:26.931+0000 I NETWORK [conn30] end connection 10.99.163.247:48415 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.941+0000 c20018| 2015-11-23T18:45:26.931+0000 I NETWORK [conn23] end connection 10.99.163.247:36845 (17 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.941+0000 c20018| 2015-11-23T18:45:26.931+0000 I NETWORK [conn24] end connection 10.99.163.247:36846 (16 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.941+0000 c20016| 2015-11-23T18:45:26.931+0000 I NETWORK [conn24] end connection 10.99.163.247:48396 (16 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.941+0000 c20016| 2015-11-23T18:45:26.932+0000 I NETWORK [conn25] end connection 10.99.163.247:48397 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.942+0000 c20018| 2015-11-23T18:45:26.932+0000 I NETWORK [conn29] end connection 10.99.163.247:36856 (15 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.942+0000 c20017| 2015-11-23T18:45:26.932+0000 I NETWORK [conn24] end connection 10.99.163.247:55717 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.942+0000 c20016| 2015-11-23T18:45:26.932+0000 I NETWORK [conn31] end connection 10.99.163.247:48420 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.942+0000 c20017| 2015-11-23T18:45:26.932+0000 I NETWORK [conn28] end connection 10.99.163.247:55730 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.943+0000 c20017| 2015-11-23T18:45:26.932+0000 I NETWORK [conn29] end connection 10.99.163.247:55732 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.943+0000 c20018| 2015-11-23T18:45:26.932+0000 I NETWORK [conn28] end connection 10.99.163.247:36855 (14 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.943+0000 c20016| 2015-11-23T18:45:26.932+0000 I NETWORK [conn20] end connection 10.99.163.247:48388 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.943+0000 c20017| 2015-11-23T18:45:26.932+0000 I NETWORK [conn20] end connection 10.99.163.247:55704 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.944+0000 c20016| 2015-11-23T18:45:26.933+0000 I NETWORK [conn22] end connection 10.99.163.247:48394 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.944+0000 c20018| 2015-11-23T18:45:26.933+0000 I NETWORK [conn22] end connection 10.99.163.247:36844 (13 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.944+0000 c20017| 2015-11-23T18:45:26.933+0000 I NETWORK [conn22] end connection 10.99.163.247:55715 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.944+0000 c20017| 2015-11-23T18:45:26.933+0000 I NETWORK [conn26] end connection 10.99.163.247:55722 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.945+0000 d20010| 2015-11-23T18:45:26.933+0000 I NETWORK [conn59] end connection 10.99.163.247:38914 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.945+0000 c20016| 2015-11-23T18:45:26.933+0000 I NETWORK [conn26] end connection 10.99.163.247:48398 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.945+0000 c20018| 2015-11-23T18:45:26.933+0000 I NETWORK [conn21] end connection 10.99.163.247:36840 (12 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.945+0000 d20010| 2015-11-23T18:45:26.933+0000 I NETWORK [conn60] end connection 10.99.163.247:38919 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.946+0000 d20010| 2015-11-23T18:45:26.933+0000 I NETWORK [conn61] end connection 10.99.163.247:38920 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.946+0000 d20014| 2015-11-23T18:45:26.933+0000 I NETWORK [conn10] end connection 10.99.163.247:37430 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.946+0000 c20016| 2015-11-23T18:45:26.933+0000 I NETWORK [conn33] end connection 10.99.163.247:48427 (10 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.946+0000 c20016| 2015-11-23T18:45:26.934+0000 I NETWORK [conn34] end connection 10.99.163.247:48430 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.947+0000 c20016| 2015-11-23T18:45:26.934+0000 I NETWORK [conn28] end connection 10.99.163.247:48404 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.947+0000 c20017| 2015-11-23T18:45:26.934+0000 I NETWORK [conn32] end connection 10.99.163.247:55745 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.947+0000 c20018| 2015-11-23T18:45:26.934+0000 I NETWORK [conn20] end connection 10.99.163.247:36826 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.947+0000 c20018| 2015-11-23T18:45:26.934+0000 I NETWORK [conn32] end connection 10.99.163.247:36866 (11 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.948+0000 c20016| 2015-11-23T18:45:26.934+0000 I NETWORK [conn27] end connection 10.99.163.247:48400 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.948+0000 c20016| 2015-11-23T18:45:26.935+0000 I NETWORK [conn29] end connection 10.99.163.247:48405 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.948+0000 c20017| 2015-11-23T18:45:26.935+0000 I NETWORK [conn25] end connection 10.99.163.247:55721 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.948+0000 c20017| 2015-11-23T18:45:26.935+0000 I NETWORK [conn30] end connection 10.99.163.247:55733 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.949+0000 c20018| 2015-11-23T18:45:26.935+0000 I NETWORK [conn30] end connection 10.99.163.247:36857 (9 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.949+0000 c20018| 2015-11-23T18:45:26.935+0000 I NETWORK [conn25] end connection 10.99.163.247:36847 (8 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.949+0000 c20018| 2015-11-23T18:45:26.935+0000 I NETWORK [conn26] end connection 10.99.163.247:36848 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.949+0000 c20016| 2015-11-23T18:45:26.935+0000 I NETWORK [conn32] end connection 10.99.163.247:48424 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.950+0000 c20018| 2015-11-23T18:45:26.935+0000 I NETWORK [conn27] end connection 10.99.163.247:36851 (7 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.950+0000 d20011| 2015-11-23T18:45:26.935+0000 I NETWORK [conn11] end connection 10.99.163.247:49873 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.950+0000 d20015| 2015-11-23T18:45:26.936+0000 I NETWORK [conn8] end connection 10.99.163.247:56718 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.950+0000 d20012| 2015-11-23T18:45:26.936+0000 I NETWORK [conn10] end connection 10.99.163.247:46112 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.951+0000 c20018| 2015-11-23T18:45:26.936+0000 I NETWORK [conn31] end connection 10.99.163.247:36863 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.951+0000 c20017| 2015-11-23T18:45:26.936+0000 I NETWORK [conn31] end connection 10.99.163.247:55742 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.951+0000 c20017| 2015-11-23T18:45:26.936+0000 I NETWORK [conn38] end connection 10.99.163.247:55808 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.951+0000 c20016| 2015-11-23T18:45:26.936+0000 I NETWORK [conn40] end connection 10.99.163.247:48493 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.952+0000 c20018| 2015-11-23T18:45:26.936+0000 I NETWORK [conn38] end connection 10.99.163.247:36929 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.986+0000 d20010| 2015-11-23T18:45:26.983+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.986+0000 d20010| 2015-11-23T18:45:26.983+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.997+0000 d20012| 2015-11-23T18:45:26.997+0000 I NETWORK [conn3] end connection 10.99.163.247:45852 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.997+0000 d20011| 2015-11-23T18:45:26.997+0000 I NETWORK [conn7] end connection 10.99.163.247:49623 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:26.998+0000 d20011| 2015-11-23T18:45:26.997+0000 I NETWORK [conn6] end connection 10.99.163.247:49622 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:27.876+0000 2015-11-23T18:45:27.876+0000 I - [thread1] shell: stopped mongo program on port 20010 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:27.877+0000 ReplSetTest stop *** Mongod in port 20010 shutdown with code (0) *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:27.877+0000 ReplSetTest stop *** Shutting down mongod in port 20011 *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:27.877+0000 d20011| 2015-11-23T18:45:27.876+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:27.877+0000 d20011| 2015-11-23T18:45:27.876+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:27.878+0000 d20011| 2015-11-23T18:45:27.878+0000 I REPL [signalProcessingThread] Stopping replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.844+0000 d20015| 2015-11-23T18:45:28.843+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.845+0000 d20015| 2015-11-23T18:45:28.844+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.845+0000 d20015| 2015-11-23T18:45:28.844+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.932+0000 d20011| 2015-11-23T18:45:28.931+0000 I STORAGE [conn5] got request after shutdown() [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.933+0000 d20012| 2015-11-23T18:45:28.931+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20011; HostUnreachable End of file [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.933+0000 d20012| 2015-11-23T18:45:28.932+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.934+0000 d20012| 2015-11-23T18:45:28.932+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:28.934+0000 d20012| 2015-11-23T18:45:28.933+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20010; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.081+0000 d20011| 2015-11-23T18:45:29.080+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.082+0000 d20011| 2015-11-23T18:45:29.080+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.082+0000 d20011| 2015-11-23T18:45:29.080+0000 I NETWORK [signalProcessingThread] closing listening socket: 12 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.082+0000 d20011| 2015-11-23T18:45:29.081+0000 I NETWORK [signalProcessingThread] closing listening socket: 13 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.083+0000 d20011| 2015-11-23T18:45:29.081+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20011.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.083+0000 d20011| 2015-11-23T18:45:29.081+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.083+0000 d20011| 2015-11-23T18:45:29.081+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.083+0000 d20011| 2015-11-23T18:45:29.081+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.084+0000 d20011| 2015-11-23T18:45:29.081+0000 I NETWORK [conn9] end connection 10.99.163.247:49862 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.084+0000 d20011| 2015-11-23T18:45:29.081+0000 I NETWORK [conn1] end connection 127.0.0.1:39276 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.084+0000 d20012| 2015-11-23T18:45:29.081+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20011; HostUnreachable Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.084+0000 d20012| 2015-11-23T18:45:29.082+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20011; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.153+0000 d20011| 2015-11-23T18:45:29.153+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.154+0000 d20011| 2015-11-23T18:45:29.153+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.167+0000 d20012| 2015-11-23T18:45:29.167+0000 I NETWORK [conn5] end connection 10.99.163.247:45860 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.820+0000 d20014| 2015-11-23T18:45:29.819+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.820+0000 d20014| 2015-11-23T18:45:29.820+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.821+0000 d20014| 2015-11-23T18:45:29.820+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.877+0000 2015-11-23T18:45:29.876+0000 I - [thread1] shell: stopped mongo program on port 20011 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.877+0000 ReplSetTest stop *** Mongod in port 20011 shutdown with code (0) *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.877+0000 ReplSetTest stop *** Shutting down mongod in port 20012 *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.877+0000 d20012| 2015-11-23T18:45:29.877+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.878+0000 d20012| 2015-11-23T18:45:29.877+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:29.879+0000 d20012| 2015-11-23T18:45:29.878+0000 I REPL [signalProcessingThread] Stopping replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:30.845+0000 d20015| 2015-11-23T18:45:30.845+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:30.846+0000 d20015| 2015-11-23T18:45:30.845+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:30.846+0000 d20015| 2015-11-23T18:45:30.846+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.082+0000 d20012| 2015-11-23T18:45:31.081+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.082+0000 d20012| 2015-11-23T18:45:31.081+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.083+0000 d20012| 2015-11-23T18:45:31.081+0000 I NETWORK [signalProcessingThread] closing listening socket: 15 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.083+0000 d20012| 2015-11-23T18:45:31.081+0000 I NETWORK [signalProcessingThread] closing listening socket: 16 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.083+0000 d20012| 2015-11-23T18:45:31.081+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20012.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.083+0000 d20012| 2015-11-23T18:45:31.081+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.084+0000 d20012| 2015-11-23T18:45:31.081+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.084+0000 d20012| 2015-11-23T18:45:31.081+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.084+0000 d20012| 2015-11-23T18:45:31.081+0000 I NETWORK [conn1] end connection 127.0.0.1:55048 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.084+0000 d20012| 2015-11-23T18:45:31.082+0000 I NETWORK [conn6] end connection 10.99.163.247:45897 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.100+0000 d20012| 2015-11-23T18:45:31.099+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.100+0000 d20012| 2015-11-23T18:45:31.099+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.822+0000 d20014| 2015-11-23T18:45:31.821+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.822+0000 d20014| 2015-11-23T18:45:31.822+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.823+0000 d20014| 2015-11-23T18:45:31.822+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.878+0000 2015-11-23T18:45:31.877+0000 I - [thread1] shell: stopped mongo program on port 20012 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.878+0000 ReplSetTest stop *** Mongod in port 20012 shutdown with code (0) *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.878+0000 ReplSetTest stopSet deleting all dbpaths [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.901+0000 ReplSetTest stopSet *** Shut down repl set - test worked **** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.901+0000 ReplSetTest stop *** Shutting down mongod in port 20013 *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.902+0000 2015-11-23T18:45:31.901+0000 I - [thread1] shell: stopped mongo program on port 20013 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.902+0000 ReplSetTest stop *** Mongod in port 20013 shutdown with code (-6) *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.902+0000 ReplSetTest stop *** Shutting down mongod in port 20014 *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.903+0000 d20014| 2015-11-23T18:45:31.901+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.903+0000 d20014| 2015-11-23T18:45:31.902+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:31.904+0000 d20014| 2015-11-23T18:45:31.903+0000 I REPL [signalProcessingThread] Stopping replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:32.842+0000 d20014| 2015-11-23T18:45:32.841+0000 I STORAGE [conn5] got request after shutdown() [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:32.842+0000 d20015| 2015-11-23T18:45:32.841+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20014; HostUnreachable End of file [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:32.847+0000 d20015| 2015-11-23T18:45:32.847+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:32.848+0000 d20015| 2015-11-23T18:45:32.848+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:32.848+0000 d20015| 2015-11-23T18:45:32.848+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:33.845+0000 d20015| 2015-11-23T18:45:33.845+0000 I NETWORK [initandlisten] connection accepted from 10.99.163.247:57143 #20 (6 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:34.849+0000 d20015| 2015-11-23T18:45:34.848+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:34.850+0000 d20015| 2015-11-23T18:45:34.849+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:34.850+0000 d20015| 2015-11-23T18:45:34.850+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20013; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.826+0000 d20014| 2015-11-23T18:45:35.826+0000 W EXECUTOR [rsBackgroundSync] killCursors command task failed: CallbackCanceled Callback canceled [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.827+0000 d20014| 2015-11-23T18:45:35.827+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.827+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.828+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [signalProcessingThread] closing listening socket: 21 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.828+0000 d20015| 2015-11-23T18:45:35.827+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20014; HostUnreachable Connection reset by peer [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.828+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [signalProcessingThread] closing listening socket: 22 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.829+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20014.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.829+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.829+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.829+0000 d20014| 2015-11-23T18:45:35.827+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.830+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [conn1] end connection 127.0.0.1:47948 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.831+0000 d20015| 2015-11-23T18:45:35.827+0000 I NETWORK [conn18] end connection 10.99.163.247:57110 (5 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.831+0000 d20014| 2015-11-23T18:45:35.827+0000 I NETWORK [conn8] end connection 10.99.163.247:37220 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.831+0000 d20015| 2015-11-23T18:45:35.828+0000 I NETWORK [ReplicaSetMonitorWatcher] Socket recv() errno:104 Connection reset by peer 10.99.163.247:20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.832+0000 d20015| 2015-11-23T18:45:35.828+0000 I NETWORK [ReplicaSetMonitorWatcher] SocketException: remote: (NONE):0 error: 9001 socket exception [RECV_ERROR] server [10.99.163.247:20014] [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.832+0000 d20015| 2015-11-23T18:45:35.828+0000 I REPL [ReplicationExecutor] Error in heartbeat request to ip-10-99-163-247:20014; HostUnreachable Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.832+0000 d20015| 2015-11-23T18:45:35.829+0000 W NETWORK [ReplicaSetMonitorWatcher] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.833+0000 d20015| 2015-11-23T18:45:35.830+0000 W NETWORK [ReplicaSetMonitorWatcher] Failed to connect to 10.99.163.247:20012, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.833+0000 d20015| 2015-11-23T18:45:35.830+0000 W NETWORK [ReplicaSetMonitorWatcher] Failed to connect to 10.99.163.247:20011, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.833+0000 d20015| 2015-11-23T18:45:35.831+0000 W NETWORK [ReplicaSetMonitorWatcher] Failed to connect to 10.99.163.247:20010, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.834+0000 d20015| 2015-11-23T18:45:35.831+0000 W NETWORK [ReplicaSetMonitorWatcher] No primary detected for set test-rs0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.834+0000 d20015| 2015-11-23T18:45:35.831+0000 I NETWORK [ReplicaSetMonitorWatcher] All nodes for set test-rs0 are down. This has happened for 1 checks in a row. Polling will stop after 29 more failed checks [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.897+0000 d20014| 2015-11-23T18:45:35.897+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.897+0000 d20014| 2015-11-23T18:45:35.897+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.904+0000 d20015| 2015-11-23T18:45:35.904+0000 I NETWORK [conn19] end connection 10.99.163.247:57111 (4 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.905+0000 2015-11-23T18:45:35.904+0000 I - [thread1] shell: stopped mongo program on port 20014 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.905+0000 ReplSetTest stop *** Mongod in port 20014 shutdown with code (0) *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.905+0000 ReplSetTest stop *** Shutting down mongod in port 20015 *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.906+0000 d20015| 2015-11-23T18:45:35.904+0000 I NETWORK [conn5] end connection 10.99.163.247:56542 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.906+0000 d20015| 2015-11-23T18:45:35.904+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.906+0000 d20015| 2015-11-23T18:45:35.904+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:35.907+0000 d20015| 2015-11-23T18:45:35.906+0000 I REPL [signalProcessingThread] Stopping replication applier threads [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:36.808+0000 2015-11-23T18:45:36.808+0000 d20015| 2015-11-23T18:45:36.808+0000 I STORAGE [conn6] got request after shutdown() [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:36.809+0000 I NETWORK [ReplicaSetMonitorWatcher] Detected bad connection created at 1448304273736894 microSec, clearing pool for ip-10-99-163-247:20015 of 0 connections [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:36.809+0000 2015-11-23T18:45:36.808+0000 I NETWORK [ReplicaSetMonitorWatcher] Socket closed remotely, no longer connected (idle 10 secs, remote host 10.99.163.247:20014) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:36.810+0000 2015-11-23T18:45:36.809+0000 W NETWORK [ReplicaSetMonitorWatcher] Failed to connect to 10.99.163.247:20014, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:36.810+0000 2015-11-23T18:45:36.809+0000 W NETWORK [ReplicaSetMonitorWatcher] Failed to connect to 10.99.163.247:20013, reason: errno:111 Connection refused [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:36.810+0000 2015-11-23T18:45:36.810+0000 W NETWORK [ReplicaSetMonitorWatcher] No primary detected for set test-rs1 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:36.811+0000 2015-11-23T18:45:36.810+0000 I NETWORK [ReplicaSetMonitorWatcher] All nodes for set test-rs1 are down. This has happened for 1 checks in a row. Polling will stop after 29 more failed checks [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.821+0000 d20015| 2015-11-23T18:45:37.819+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.821+0000 d20015| 2015-11-23T18:45:37.819+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.821+0000 d20015| 2015-11-23T18:45:37.819+0000 I NETWORK [signalProcessingThread] closing listening socket: 24 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.822+0000 d20015| 2015-11-23T18:45:37.819+0000 I NETWORK [signalProcessingThread] closing listening socket: 25 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.822+0000 d20015| 2015-11-23T18:45:37.819+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20015.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.822+0000 d20015| 2015-11-23T18:45:37.819+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.822+0000 d20015| 2015-11-23T18:45:37.819+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.823+0000 d20015| 2015-11-23T18:45:37.819+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.823+0000 d20015| 2015-11-23T18:45:37.820+0000 I NETWORK [conn20] end connection 10.99.163.247:57143 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.823+0000 c20018| 2015-11-23T18:45:37.820+0000 I NETWORK [conn52] end connection 10.99.163.247:37244 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.823+0000 d20015| 2015-11-23T18:45:37.820+0000 I NETWORK [conn1] end connection 127.0.0.1:35712 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.824+0000 c20017| 2015-11-23T18:45:37.820+0000 I NETWORK [conn52] end connection 10.99.163.247:56123 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.824+0000 c20017| 2015-11-23T18:45:37.820+0000 I NETWORK [conn53] end connection 10.99.163.247:56126 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.824+0000 c20016| 2015-11-23T18:45:37.820+0000 I NETWORK [conn58] end connection 10.99.163.247:48811 (3 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.824+0000 c20018| 2015-11-23T18:45:37.820+0000 I NETWORK [conn53] end connection 10.99.163.247:37247 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.825+0000 c20016| 2015-11-23T18:45:37.821+0000 I NETWORK [conn57] end connection 10.99.163.247:48808 (2 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.890+0000 d20015| 2015-11-23T18:45:37.889+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.890+0000 d20015| 2015-11-23T18:45:37.890+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.905+0000 2015-11-23T18:45:37.904+0000 I - [thread1] shell: stopped mongo program on port 20015 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.905+0000 ReplSetTest stop *** Mongod in port 20015 shutdown with code (0) *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.906+0000 ReplSetTest stopSet deleting all dbpaths [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.912+0000 ReplSetTest stopSet *** Shut down repl set - test worked **** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.913+0000 c20016| 2015-11-23T18:45:37.912+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.913+0000 c20016| 2015-11-23T18:45:37.912+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.914+0000 c20016| 2015-11-23T18:45:37.914+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.915+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.915+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [signalProcessingThread] closing listening socket: 32 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.915+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [signalProcessingThread] closing listening socket: 33 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.915+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20016.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.916+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.916+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.916+0000 c20016| 2015-11-23T18:45:37.914+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.916+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [conn2] end connection 10.99.163.247:48277 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.917+0000 c20016| 2015-11-23T18:45:37.914+0000 I NETWORK [conn1] end connection 127.0.0.1:49084 (0 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.964+0000 c20016| 2015-11-23T18:45:37.963+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:37.964+0000 c20016| 2015-11-23T18:45:37.963+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.913+0000 2015-11-23T18:45:38.912+0000 I - [thread1] shell: stopped mongo program on port 20016 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.913+0000 c20017| 2015-11-23T18:45:38.912+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.914+0000 c20017| 2015-11-23T18:45:38.912+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.915+0000 c20017| 2015-11-23T18:45:38.914+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.915+0000 c20017| 2015-11-23T18:45:38.914+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.915+0000 c20017| 2015-11-23T18:45:38.914+0000 I NETWORK [signalProcessingThread] closing listening socket: 35 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.916+0000 c20017| 2015-11-23T18:45:38.914+0000 I NETWORK [signalProcessingThread] closing listening socket: 36 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.916+0000 c20017| 2015-11-23T18:45:38.914+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20017.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.916+0000 c20017| 2015-11-23T18:45:38.914+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.916+0000 c20017| 2015-11-23T18:45:38.914+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.917+0000 c20017| 2015-11-23T18:45:38.915+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.917+0000 c20017| 2015-11-23T18:45:38.915+0000 I NETWORK [conn2] end connection 10.99.163.247:55592 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.917+0000 c20017| 2015-11-23T18:45:38.915+0000 I NETWORK [conn1] end connection 127.0.0.1:35479 (0 connections now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.968+0000 c20017| 2015-11-23T18:45:38.967+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:38.968+0000 c20017| 2015-11-23T18:45:38.968+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.913+0000 2015-11-23T18:45:39.912+0000 I - [thread1] shell: stopped mongo program on port 20017 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.914+0000 c20018| 2015-11-23T18:45:39.913+0000 I CONTROL [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.914+0000 c20018| 2015-11-23T18:45:39.913+0000 I FTDC [signalProcessingThread] Shuting down full-time diagnostic data capture [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.916+0000 c20018| 2015-11-23T18:45:39.915+0000 I CONTROL [signalProcessingThread] now exiting [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.916+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [signalProcessingThread] shutdown: going to close listening sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.916+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [signalProcessingThread] closing listening socket: 38 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.916+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [signalProcessingThread] closing listening socket: 39 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.917+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [signalProcessingThread] removing socket file: /tmp/mongodb-20018.sock [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.917+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [signalProcessingThread] shutdown: going to flush diaglog... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.917+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [signalProcessingThread] shutdown: going to close sockets... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.917+0000 c20018| 2015-11-23T18:45:39.915+0000 I STORAGE [signalProcessingThread] WiredTigerKVEngine shutting down [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.918+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [conn2] end connection 10.99.163.247:36713 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:39.918+0000 c20018| 2015-11-23T18:45:39.915+0000 I NETWORK [conn1] end connection 127.0.0.1:49909 (1 connection now open) [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.013+0000 c20018| 2015-11-23T18:45:40.013+0000 I STORAGE [signalProcessingThread] shutdown: removing fs lock... [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.013+0000 c20018| 2015-11-23T18:45:40.013+0000 I CONTROL [signalProcessingThread] dbexit: rc: 0 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.914+0000 2015-11-23T18:45:40.913+0000 I - [thread1] shell: stopped mongo program on port 20018 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.923+0000 *** ShardingTest test completed successfully in 97.136 seconds *** [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.924+0000 2015-11-23T18:45:40.923+0000 E QUERY [thread1] Error: 1 thread threw [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.924+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.924+0000 Foreground [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.924+0000 Error: command failed: { [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.924+0000 "code" : 6, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.924+0000 "ok" : 0, [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.925+0000 "errmsg" : "network error while attempting to run command 'findandmodify' on host 'ip-10-99-163-247:20013' " [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.925+0000 } : undefined [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.925+0000 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.925+0000 quietlyDoAssert@jstests/concurrency/fsm_libs/assert.js:53:15 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.925+0000 assert.commandWorked@src/mongo/shell/assert.js:266:5 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.926+0000 wrapAssertFn@jstests/concurrency/fsm_libs/assert.js:60:13 [js_test:fsm_all_sharded_replication_legacy_config_servers_with_balancer] 2015-11-23T18:45:40.926+0000 assertWithLevel/