Uploaded image for project: 'Core Server'
  1. Core Server
  2. SERVER-6829

splitting with few chunks will never succeed if many duplicates

    XMLWordPrintableJSON

Details

    • Icon: Bug Bug
    • Resolution: Done
    • Icon: Major - P3 Major - P3
    • None
    • 2.3.0
    • Sharding
    • None
    • Sharding
    • ALL

    Description

      If for some reason the data being inserted has a few duplicate shard keys (no so many as to cause problems) we will never split the initial chunk.

      For example, if we're inserting documents of size 300k and insert two documents in a row with the same shard key each time, we'll never split b/c for a split with only one chunk, the maxChunkSize is forced to 1024 bytes, and 1/2 of that is 500k.

      Would be better to specify a different parameter as a target chunk size and split between targetChunkSize and maxChunkSize wherever possible. Mongod could also be smarter here since it knows the actual size of the initial chunk.

      Attachments

        Activity

          People

            backlog-server-sharding [DO NOT USE] Backlog - Sharding Team
            greg_10gen Greg Studer
            Votes:
            0 Vote for this issue
            Watchers:
            1 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: