> On 6 Oct 2021, at 11:37, Christiaan Hofman <[email protected]> wrote: > > > >> On 6 Oct 2021, at 11:29, mn <[email protected] <mailto:[email protected]>> wrote: >> >> >> On 03.10.21 16:58, Christiaan Hofman wrote: >> >>>>> >>>>>> For efficiency, we don’t fetch all the results at once. If you show >>>>>> the status bar, you may see the total number of available results. >>>>>> >>>>>> If you search repeatedly, further results will be fetched. >>>>> >>>>> OK. >>>>> Is there some way to fetch all results at once? >>>>> A setting or option to customize the numbers of results fetched? >>>>> If not, adding those would be most welcome. >>>>> >>>> >>>> No, there isn’t. >>>> >>> >>> BTW, this is not just our choice. It is also the policy for the server >>> for the web interface. And they threaten to block your IP address when >>> you don’t comply with their policy, so I don’t think it is a good idea >>> to ignore that. >>> >> >> On their website >> https://www.ncbi.nlm.nih.gov/books/NBK25497/#chapter2.Usage_Guidelines_and_Requiremen >> >> <https://www.ncbi.nlm.nih.gov/books/NBK25497/#chapter2.Usage_Guidelines_and_Requiremen> >> they give the following: >> >> > In order not to overload the E-utility servers, NCBI recommends that >> users post no more than three URL requests per second and limit large >> jobs to either weekends or between 9:00 PM and 5:00 AM Eastern time >> during weekdays. Failure to comply with this policy may result in an IP >> address being blocked from accessing NCBI. >> >> The mechanics behind the scene here elude me: BibDesk fetches 50 >> _results_ in one go, apparently fine with the '≤3 URL requests /s'? >> >> On the face of it, I'd conclude that smaller portions (like 20 results >> per 'search') would be fine in any case. >> But then 150 results at a time as well? >> >> The week day angle seems quite vague, but open to interpretation that >> larger requests on weekends will be possible/tolerated? >> >> But I suspect that 'URL requests' as limited on server and the 'BibDesk >> results fetcheing' don not even correspond in that matter? >> >> >> — Mike >> > > A comment in our code also mentions a limit on the number of results. Perhaps > they have changed that policy over time. But getting a larger number of > results can also slow down the search (a lot). Also, every fetch action is a > URL action. The first one is two, because we first need to get the number of > results. > > Christiaan
BTW, if you really get a very large number of results you probably did not target your search very well, and fetching a large number of results is really not that useful, just wasteful. Are you going to look for the result you need in 15000 items? Christiaan
_______________________________________________ Bibdesk-users mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/bibdesk-users
