-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't get over 10000 results from the API #310
Comments
I'm writing a crawler for data exchange and the same happens when page = 1001. |
In the case of bio.tools it would make sense to have page 1001 since there are 10,000+ tools and 10 tools per page. But I think this is a hard limit in the code/framework/server which gives the server error. I've tried the same on dev.bio.tools: |
It looks DJango limit: one should modify the limit:
|
Reassigning to you @piotrgithub1 - this looks like it's easy to fix |
Hi @jaanisoe Best regards, |
thanks @piotrgithub1 - we have already quite a few groups / dependencies (in USA, France, Spain ...) which are taking the whole bio.tools data, so there's a real use-case here. |
Made a temporary workaround for the bug. NOTE: leaving this open until we get a confirmation from @jaanisoe Best regards, |
Thanks a lot, Piotr !
I will try it right know.
The use case is the following :
- iterate over all entries and get its JSON serialization
- transform it into RDF (JSON-LD)
- populate a knowledge base for further semantic web querying.
Alban
… Le 16 févr. 2018 à 15:02, piotrgithub1 ***@***.***> a écrit :
Made a temporary workaround for the bug.
Long-term we need to take a closer look at how this is used by the dependant services and make a more use specialized functionality for getting larger amount of content out of the registry as tight coupling with ontologies is clearly insufficient.
NOTE: leaving this open until we get a confirmation from @jaanisoe <https://github.com/jaanisoe>
Best regards,
Piotr
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub <#310 (comment)>, or mute the thread <https://github.com/notifications/unsubscribe-auth/AKLGELJDYRjFcfeMFTsgxBfO8ENZrHOkks5tVYpggaJpZM4RWi3C>.
--
Alban Gaignard, PhD, CNRS
|
Yes, it's working now. Thanks! |
Awesome :) |
Hi all, I confirm,
I've been able to crawl all the 10059 entries !
Thanks again,
Alban
… Le 16 févr. 2018 à 15:09, Alban Gaignard ***@***.***> a écrit :
Thanks a lot, Piotr !
I will try it right know.
The use case is the following :
- iterate over all entries and get its JSON serialization
- transform it into RDF (JSON-LD)
- populate a knowledge base for further semantic web querying.
Alban
> Le 16 févr. 2018 à 15:02, piotrgithub1 ***@***.*** ***@***.***>> a écrit :
>
> Made a temporary workaround for the bug.
> Long-term we need to take a closer look at how this is used by the dependant services and make a more use specialized functionality for getting larger amount of content out of the registry as tight coupling with ontologies is clearly insufficient.
>
> NOTE: leaving this open until we get a confirmation from @jaanisoe <https://github.com/jaanisoe>
> Best regards,
> Piotr
>
> —
> You are receiving this because you commented.
> Reply to this email directly, view it on GitHub <#310 (comment)>, or mute the thread <https://github.com/notifications/unsubscribe-auth/AKLGELJDYRjFcfeMFTsgxBfO8ENZrHOkks5tVYpggaJpZM4RWi3C>.
>
--
Alban Gaignard, PhD, CNRS
--
Alban Gaignard, PhD, CNRS
|
We need to check this still works when we make the next big release |
See also #355 |
fix confirmed |
Or more precisely, the maximum working value of the
page
parameter is 1000:https://bio.tools/api/tool?page=1000
When the next
page
is tried (as listed by"next": "?page=1001"
), I get a "Server Error (500)":https://bio.tools/api/tool?page=1001
The text was updated successfully, but these errors were encountered: