You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Found that the spaces_reader returns too many records unless using 1 worker and 1 slicer, also reducing the interval to 1s independent of the number of workers or slicers returns close to the correct number of records but is still a bit too high.
Used a control group of data of 6.95M records in all the tests.
Tests were ran with elasticsearch-asset version 2.6.2, node-12 on dataeng3, teraslice version 0.76.1
workers
slicers
interval
docs returned (M)
20
10
auto
8.38
20
1
auto
7.81
1
1
auto
6.95
20
10
1s
6.97
20
10
1m
8.19
20
10
1hr
8.48
Ran a job with 20 workers and 10 slicers, interval auto that deduped the records and the count came to 6.95M, so it looks like it's picking up duplicate records.
The text was updated successfully, but these errors were encountered:
Found that the
spaces_reader
returns too many records unless using 1 worker and 1 slicer, also reducing the interval to 1s independent of the number of workers or slicers returns close to the correct number of records but is still a bit too high.Used a control group of data of 6.95M records in all the tests.
Tests were ran with elasticsearch-asset version 2.6.2, node-12 on dataeng3, teraslice version 0.76.1
Ran a job with 20 workers and 10 slicers, interval auto that deduped the records and the count came to 6.95M, so it looks like it's picking up duplicate records.
The text was updated successfully, but these errors were encountered: