Priority: Major - P3
Affects Version/s: 2.0.0, 2.2.0
Fix Version/s: None
When running a map/reduce job on a collection containing a large number of entity for a single key in the map/reduce operation, the computation fails because of the size of the object fed into the reduce.
In our case the reduce accumulates the values encountered in an array. The finalize would after trim down these values and apply the final operations to the final object. This accumulation process cannot be filtered before and needs all the informations.
Avg Object Size: 82 Kb, Size of the database: 60 Mb, Storage: 80 Mb
Object Count: 801
Number of MapReduce keys: 1