Open SimonShapiro opened 5 years ago
I was reading the readme file and I noticed that there is a size limit "Its current function is to set a quota for disk usage of just 25 MB, which is what we can be sure the current prototype can tolerate under load."
I know that the full data set is around 100MB, could this be the reason for the sudden error 500 after running ok for some time?
which is what we can be sure the current prototype can tolerate under load
@kjetilk @megoth That claim does not have any grounding. 25MB has to do with disk space on our public servers, not with the server software. Can we fix that?
@kjetilk @megoth That claim does not have any grounding. 25MB has to do with disk space on our public servers, not with the server software. Can we fix that?
Well, it is more a heuristic, but I agree that "be sure" is a bit too strongly worded, so I changed the wording.
So, @SimonShapiro , it wouldn't be the reason, it is more that we don't know much about the scalability properties of the server, other than that it doesn't scale well. We have at least one more bug that relates to scalability, #849 , you could check if this is related.
It is still useful to get this kind of reports, to understand better where the limits are, but unless it turns out to be low hanging fruit, we are more likely to focus on rearchitecting the server to better scale.
If you want me to trap any specific log information, let me know how to do so and I will re-run it.
I prepare all 1.6m triples using the python Rdflib then I convert them to json-ld as part of the pipeline where they become around 150,000 json-ld objects on disk. Currently the json is served with any old httpbserver(msoft iis in production). I was looking to solid to move away from the json-ld strategy.
On Tue, 18 Dec 2018 at 21:49, Kjetil Kjernsmo notifications@github.com wrote:
@kjetilk https://github.com/kjetilk @megoth https://github.com/megoth That claim does not have any grounding. 25MB has to do with disk space on our public servers, not with the server software. Can we fix that?
Well, it is more a heuristic, but I agree that "be sure" is a bit too strongly worded, so I changed the wording.
So, @SimonShapiro https://github.com/SimonShapiro , it wouldn't be the reason, it is more that we don't know much about the scalability properties of the server, other than that it doesn't scale well. We have at least one more bug that relates to scalability, #849 https://github.com/solid/node-solid-server/issues/849 , you could check if this is related.
It is still useful to get this kind of reports, to understand better where the limits are, but unless it turns out to be low hanging fruit, we are more likely to focus on rearchitecting the server to better scale.
— You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub https://github.com/solid/node-solid-server/issues/1008#issuecomment-448383250, or mute the thread https://github.com/notifications/unsubscribe-auth/ABl65_hlcST0yVoeQrHF3ICs3HdI_B9Sks5u6WLYgaJpZM4ZBZmZ .
I have experienced a strange behaviour over the REST interface using node-fetch module.
While this is not the most efficient way of doing things, it has thrown up the following error 500.
I read in around 11 ttl files so that total about 1,6m triples. I now want to pivot them and create a file per subject. I process each of the 1,6m triples as set out below. Check whether the resource exists. If it does not, create it using PUT, else PATCH in the new triple.
For logging I write out a 'pulse' every 5,000 records.
The code below is an extract of the process.
Just after the 80,000th log entry and before the 85,000th log entry, the server starts producing error 500 and never recovers. As can be seen in the output below: