Handle small measurement counts in buckets for ARHASH

XMLWordPrintableJSON

    • Type: Task
    • Resolution: Fixed
    • Priority: Major - P3
    • 5.0.0-rc0
    • Affects Version/s: None
    • Component/s: Querying
    • Fully Compatible
    • Query Execution 2021-03-22, Query Execution 2021-04-05, Query Execution 2021-04-19
    • 28
    • None
    • 3
    • None
    • None
    • None
    • None
    • None
    • None

      When testing SERVER-54221 we realized that if the gTimeseriesBucketMaxCount is significantly larger than the actual bucket counts we will exhaust the kMaxAttempt to get a non-duplicate document and fail the query.

      This behavior is not ideal. We should implement a fall-back mechanism when this is the case. Two ideas discussed is to figure out if we can cheaply compute the maximum bucket count, or use a trial stage to see if we are able to sample the collection, and if not fallback to topk sorting.

            Assignee:
            David Storch
            Reporter:
            Eric Cox (Inactive)
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

              Created:
              Updated:
              Resolved: