> On 6 Oct 2021, at 11:37, Christiaan Hofman <[email protected]> wrote:
> 
> 
> 
>> On 6 Oct 2021, at 11:29, mn <[email protected] <mailto:[email protected]>> wrote:
>> 
>> 
>> On 03.10.21 16:58, Christiaan Hofman wrote:
>> 
>>>>> 
>>>>>> For efficiency, we don’t fetch all the results at once. If you show
>>>>>> the status bar, you may see the total number of available results.
>>>>>> 
>>>>>> If you search repeatedly, further results will be fetched.
>>>>> 
>>>>> OK.
>>>>> Is there some way to fetch all results at once?
>>>>> A setting or option to customize the numbers of results fetched?
>>>>> If not, adding those would be most welcome.
>>>>> 
>>>> 
>>>> No, there isn’t.
>>>> 
>>> 
>>> BTW, this is not just our choice. It is also the policy for the server
>>> for the web interface. And they threaten to block your IP address when
>>> you don’t comply with their policy, so I don’t think it is a good idea
>>> to ignore that.
>>> 
>> 
>> On their website
>> https://www.ncbi.nlm.nih.gov/books/NBK25497/#chapter2.Usage_Guidelines_and_Requiremen
>>  
>> <https://www.ncbi.nlm.nih.gov/books/NBK25497/#chapter2.Usage_Guidelines_and_Requiremen>
>> they give the following:
>> 
>> > In order not to overload the E-utility servers, NCBI recommends that
>> users post no more than three URL requests per second and limit large
>> jobs to either weekends or between 9:00 PM and 5:00 AM Eastern time
>> during weekdays. Failure to comply with this policy may result in an IP
>> address being blocked from accessing NCBI.
>> 
>> The mechanics behind the scene here elude me: BibDesk fetches 50
>> _results_ in one go, apparently fine with the '≤3 URL requests /s'?
>> 
>> On the face of it, I'd conclude that smaller portions (like 20 results
>> per 'search') would be fine in any case.
>> But then 150 results at a time as well?
>> 
>> The week day angle seems quite vague, but open to interpretation that
>> larger requests on weekends will be possible/tolerated?
>> 
>> But I suspect that 'URL requests' as limited on server and the 'BibDesk
>> results fetcheing' don not even correspond in that matter?
>> 
>> 
>> — Mike
>> 
> 
> A comment in our code also mentions a limit on the number of results. Perhaps 
> they have changed that policy over time. But getting a larger number of 
> results can also slow down the search (a lot). Also, every fetch action is a 
> URL action. The first one is two, because we first need to get the number of 
> results. 
> 
> Christiaan


BTW, if you really get a very large number of results you probably did not 
target your search very well, and fetching a large number of results is really 
not that useful, just wasteful. Are you going to look for the result you need 
in 15000 items?

Christiaan

_______________________________________________
Bibdesk-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/bibdesk-users

Reply via email to