Uploaded image for project: 'Core Server'
  1. Core Server
  2. SERVER-23627

Randomly locking in WiredTiger storage engine

    XMLWordPrintable

    Details

    • Type: Bug
    • Status: Closed
    • Priority: Critical - P2
    • Resolution: Gone away
    • Affects Version/s: 3.0.10
    • Fix Version/s: None
    • Component/s: Performance, WiredTiger
    • Labels:
    • Environment:
      production
    • Operating System:
      ALL
    • Steps To Reproduce:
      Hide

      Not able to reproduce in staging environment.

      Show
      Not able to reproduce in staging environment.

      Description

      Hi,

      last week we have upgraded from mongo 3.0.6 to 3.0.10. And after upgrading server we are getting random locking issues (even for 1 minute).
      In general these queries will take ~70ms, but during that time it start taking 70 seconds (for us its an outage).
      Tried switching primary to other secondary, but still facing the same issue.

      Below are the logs, when it start happening

      2016-04-09T08:15:02.106+0530 I NETWORK  [conn222315] end connection 10.14.4.20:51763 (287 connections now open)
      2016-04-09T08:15:02.107+0530 I NETWORK  [initandlisten] connection accepted from 10.14.4.20:51770 #222318 (288 connections now open)
      2016-04-09T08:15:03.223+0530 I NETWORK  [initandlisten] connection accepted from 10.14.1.108:40449 #222319 (289 connections now open)
      2016-04-09T08:15:03.223+0530 I NETWORK  [conn222319] end connection 10.14.1.108:40449 (288 connections now open)
      2016-04-09T08:15:03.223+0530 I NETWORK  [initandlisten] connection accepted from 10.14.1.108:40450 #222320 (289 connections now open)
      2016-04-09T08:15:03.224+0530 I NETWORK  [initandlisten] connection accepted from 10.14.1.108:40451 #222321 (290 connections now open)
      2016-04-09T08:15:03.225+0530 I NETWORK  [conn222321] end connection 10.14.1.108:40451 (289 connections now open)
      2016-04-09T08:15:03.225+0530 I NETWORK  [initandlisten] connection accepted from 10.14.1.108:40452 #222322 (290 connections now open)
      2016-04-09T08:15:03.232+0530 I NETWORK  [initandlisten] connection accepted from 10.14.1.108:40457 #222323 (291 connections now open)
      2016-04-09T08:15:03.340+0530 I COMMAND  [conn222323] command roster_production.$cmd command: count { count: "inventories", query: { current_location.fix_time: { $lte: new Date(1460166903224) }, status: { $in: [ "idle" ] } } } planSummary: COLLSCAN keyUpdates:0 writeConflicts:0 numYields:786 reslen:44 locks:{ Global: { acquireCount: { r: 1574 } }, Database: { acquireCount: { r: 787 } }, Collection: { acquireCount: { r: 787 } } } 107ms
      2016-04-09T08:15:03.449+0530 I COMMAND  [conn222323] command roster_production.$cmd command: count { count: "inventories", query: { current_location.fix_time: { $lte: new Date(1460166903224) }, status: { $in: [ "idle" ] } } } planSummary: COLLSCAN keyUpdates:0 writeConflicts:0 numYields:786 reslen:44 locks:{ Global: { acquireCount: { r: 1574 } }, Database: { acquireCount: { r: 787 } }, Collection: { acquireCount: { r: 787 } } } 107ms
      2016-04-09T08:15:03.710+0530 I COMMAND  [conn222323] command roster_production.$cmd command: count { count: "inventories", query: { current_location.fix_time: { $lte: new Date(1460166903224) }, status: { $in: [ "idle" ] } } } planSummary: COLLSCAN keyUpdates:0 writeConflicts:0 numYields:786 reslen:44 locks:{ Global: { acquireCount: { r: 1574 } }, Database: { acquireCount: { r: 787 } }, Collection: { acquireCount: { r: 787 } } } 107ms
      2016-04-09T08:15:06.767+0530 I NETWORK  [initandlisten] connection accepted from 127.0.0.1:50024 #222324 (292 connections now open)
      2016-04-09T08:15:13.233+0530 I NETWORK  [conn222320] end connection 10.14.1.108:40450 (291 connections now open)
      2016-04-09T08:15:14.758+0530 I NETWORK  [conn222317] end connection 10.14.1.207:46537 (290 connections now open)
      2016-04-09T08:15:14.758+0530 I NETWORK  [initandlisten] connection accepted from 10.14.1.207:46545 #222325 (291 connections now open)
      2016-04-09T08:15:23.438+0530 I COMMAND  [conn222184] command local.$cmd command: dbStats { dbStats: 1 } keyUpdates:0 writeConflicts:0 numYields:0 reslen:188 locks:{ Global: { acquireCount: { r: 2 } }, Database: { acquireCount: { R: 1 }, acquireWaitCount: { R: 1 }, timeAcquiringMicros: { R: 76628917 } } } 76629ms
      2016-04-09T08:15:23.438+0530 I WRITE    [conn222239] update roster_production.inventories query: { _id: ObjectId('56c6b540af1df3a27122eae6') } update: { $set: { last_request_device_time: new Date(1460169847843), updated_at: new Date(1460169847848), current_location._id: ObjectId('57086c773b6f82232e000000') } } nscanned:1 nscannedObjects:1 nMatched:1 nModified:1 keyUpdates:1 writeConflicts:0 numYields:1 locks:{ Global: { acquireCount: { r: 3, w: 3 } }, Database: { acquireCount: { w: 3 }, acquireWaitCount: { w: 1 }, timeAcquiringMicros: { w: 75590590 } }, Collection: { acquireCount: { w: 2 } }, oplog: { acquireCount: { w: 1 } } } 75590ms
      2016-04-09T08:15:23.438+0530 I WRITE    [conn222296] update roster_production.inventories query: { _id: ObjectId('56a3181aaf1df3a27122c482') } update: { $set: { last_request_device_time: new Date(1460169848325), updated_at: new Date(1460169848330), current_location._id: ObjectId('57086c78706d2756aa000000') } } nscanned:1 nscannedObjects:1 nMatched:1 nModified:1 keyUpdates:1 writeConflicts:0 numYields:1 locks:{ Global: { acquireCount: { r: 3, w: 3 } }, Database: { acquireCount: { w: 3 }, acquireWaitCount: { w: 1 }, timeAcquiringMicros: { w: 75106091 } }, Collection: { acquireCount: { w: 2 } }, oplog: { acquireCount: { w: 1 } } } 75106ms
      2016-04-09T08:15:23.438+0530 I NETWORK  [conn222184] end connection 127.0.0.1:50008 (290 connections now open)
      2016-04-09T08:15:23.438+0530 I WRITE    [conn222231] insert roster_production.legs query: { _id: ObjectId('57086c78706d27547b000004'), session_id: ObjectId('5708621d706d272159000850'), engaged_by: 114135077, coverage: 1471870, wait_time: 686.0, request_time: new Date(1460169847000), lazy: false, source: "fixed", previous_status: "idle", current_status: "engaged", next_status: "idle", inventory_id: ObjectId('56cedb47af1df3a27122ef0d'), updated_at: new Date(1460169848132), created_at: new Date(1460169848132), start: { _id: ObjectId('57086c78706d27547b000002'), time: new Date(1460168370000), location: { _id: ObjectId('57086c78706d27547b000005'), alt: null, accuracy: 20.0, fix_time: new Date(1460168367000), provider: "fused", bearing: 197.0 } }, end: { _id: ObjectId('57086c78706d27547b000003'), time: new Date(1460169846000), location: { _id: ObjectId('57086c78706d27547b000006'), alt: null, accuracy: 20.0, fix_time: new Date(1460169842000), altitude: 570.4, provider: "fused", bearing: 1.0 } } } ninserted:1 keyUpdates:0 writeConflicts:0 numYields:0 locks:{ Global: { acquireCount: { r: 2, w: 2 } }, Database: { acquireCount: { w: 2 }, acquireWaitCount: { w: 1 }, timeAcquiringMicros: { w: 75305189 } }, Collection: { acquireCount: { w: 1 } }, oplog: { acquireCount: { w: 1 } } } 75305ms
      

      Getting Lots of queries like above one.

        Attachments

          Activity

            People

            • Votes:
              0 Vote for this issue
              Watchers:
              8 Start watching this issue

              Dates

              • Created:
                Updated:
                Resolved: