all definitions |
discriminator |
Item whose values will be used to segment objects into buckets (if applicable).
Usually required. |
matchingRule |
Matching rule to be applied when creating filters (if applicable).
Optional. |
numberOfBuckets |
Number of buckets to be created (if applicable).
Optional. |
numericSegmentation |
from |
Start of the processing space (inclusive).
If omitted, 0 is assumed. |
to |
End of the processing space (exclusive).
If not present, both bucketSize and numberOfBuckets must be defined and the end of processing space is determined as their product.
In the future we might implement dynamic determination of this value e.g. by counting objects to be processed. |
bucketSize |
Size of one bucket.
If not present it is computed as the total processing space divided by number of buckets (i.e. to and numberOfBuckets must be present). |
stringSegmentation |
boundary |
position |
Position(s) to which the boundary characters apply.
Should be specified, because
-
the ordering of boundary specifications is undefined,
-
multiple definitions of the same boundary characters is not possible.
|
characters |
Characters that make up the boundaries.
These characters must be sorted.
Reserved characters: '-', '$' (to be implemented later)
Escaping character: '\' |
depth |
If a value N greater than 1 is specified here, boundary values are repeated N times.
This means that if values of V1, V2, …, Vk are specified, the resulting sequence is V1, V2, …, Vk, V1, V2, …, Vk etc, with N repetitions - so N × k values in total. |
comparisonMethod |
-
interval (the default)
resulting in interval queries like item >= 'a' and item < 'b'
-
prefix
resulting in prefix queries like item starts with 'a' . *1
-
exactMatch
Use exact value matching. *1
*1 This is quite risky and should be used only when you are absolutely sure that boundary values cover all possible values of the discriminator.
|
oidSegmentation |
The same as stringSegmentation but providing defaults of discriminator = # and characters = 0-9a-f (repeated depth times, if needed). |
explicitSegmentation |
content |
Explicit content of work buckets to be used.
This is useful e.g. when dealing with filter-based buckets.
But any other bucket content (e.g. numeric intervals, string intervals, string prefixes) might be used here as well. |
implicitSegmentation |
(none) - see all definitions |
Implicit content of work buckets for given kind of activity to be used. |
Additional configuration for the Buckets |
allocation (experimental) |
bucketCreationBatch |
How many buckets are to be created at once |
workAllocationInitialDelay |
Size of random interval for the initial delay. |
workAllocationFreeBucketWaitInterval |
if specified, overrides the time used to wait for free bucket(s) reclamation.
This is applied when no free buckets are available but the work is not completely done. |
sampling (experimental) |
regular
|
interval |
Interval of buckets in the sample (i.e. N means that each N-th bucket is selected). |
sampleSize |
Number of buckets in the sample.
It is converted to an interval by dividing the total number of buckets (if known) by the sample size. |
random
|
probability |
Probability of including a bucket in the sample (a number between 0 and 1). |
sampleSize |
Approximate number of buckets in the sample.
It is converted to a probability by dividing the sample size by the total number of buckets (if known). |