You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
At the moment, the SPARK node tries to retrieve all content of CID, regardless of the size. Some CIDs represent GBs of data.
IMO, this is a problem - we don't want Stations to use so much bandwidth.
It also creates a problem in spark-api, where we currently represent byte_length as a 32bit signed integer, which overflows at 2GB.
2023-09-05T15:54:19Z app[17814d5b527638] cdg [info]error: value "2753993443" is out of range for type integer
2023-09-05T15:54:19Z app[17814d5b527638] cdg [info] at /app/node_modules/pg-pool/index.js:45:11
2023-09-05T15:54:19Z app[17814d5b527638] cdg [info] at runMicrotasks (<anonymous>)
2023-09-05T15:54:19Z app[17814d5b527638] cdg [info] at processTicksAndRejections (node:internal/process/task_queues:96:5)
2023-09-05T15:54:19Z app[17814d5b527638] cdg [info] at async setRetrievalResult (file:///app/index.js:74:5)
2023-09-05T15:54:19Z app[17814d5b527638] cdg [info] at async handler (file:///app/index.js:12:5)
I am proposing to introduce a new retrieval error status - content too large.
The content you are editing has changed. Please copy your edits and refresh the page.
Spark clients should be allowed to abort retrieval if it is too large, without getting penalized. Then ideally they won't even report the result. However, Spark shouldn't have a problem with retrieval testing for large CIDs, a result is a result and is useful.
I think therefore we want the solution to be on the Station module side - it should abort the request - and for the Station module not to be penalized for not reporting in a large retrieval.
At the moment, the SPARK node tries to retrieve all content of CID, regardless of the size. Some CIDs represent GBs of data.
IMO, this is a problem - we don't want Stations to use so much bandwidth.
It also creates a problem in spark-api, where we currently represent
byte_length
as a 32bit signed integer, which overflows at 2GB.I am proposing to introduce a new retrieval error status - content too large.
Tasks
The text was updated successfully, but these errors were encountered: