You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For the Wuhan, China data source (seem to be getting all the legacy servers lately), pyesridump was only pulling 1001 records when there were 3429 in the full data set.
, it looks like if the source doesn't support returnCountOnly, the bounding box is recursively subdivided into four quadrants (Quadtree-style) with a stopping condition when there are < maxRecords in a given quadrant.
This should generally retrieve everything, except for the following test in _scrape_an_envelope:
iflen(features) ==max_records:
It appears the Wuhan source returns 1001 records where max_records is 1000, which executes the same code as if it had returned 999 results i.e. assumes the base case has been met and returns early. This could be fixed by changing the conditional to:
iflen(features) >=max_records
With the new OID enumeration from Faster method for object ID enumeration for sources that do not support pagination #33, it might make sense to use the quadrant-based method as a fallback only if the source supports neither returnCountOnly nor returnIdsOnly. Otherwise OID enumeration should be fewer queries. Does that make sense or are there some other edge cases to consider?
The text was updated successfully, but these errors were encountered:
For the Wuhan, China data source (seem to be getting all the legacy servers lately), pyesridump was only pulling 1001 records when there were 3429 in the full data set.
From this line:
pyesridump/esridump/dumper.py
Line 244 in 5823178
The text was updated successfully, but these errors were encountered: