Re: Fatal Error : Invalid Memory alloc request size 1236252631
От | Rob Sargent |
---|---|
Тема | Re: Fatal Error : Invalid Memory alloc request size 1236252631 |
Дата | |
Msg-id | 7adb97ac-5433-054d-32a0-a1046f7212da@gmail.com обсуждение исходный текст |
Ответ на | Re: Fatal Error : Invalid Memory alloc request size 1236252631 (Sai Teja <saitejasaichintalapudi@gmail.com>) |
Список | pgsql-general |
On 8/17/23 07:35, Sai Teja wrote: > Hi Team, > > Even I used postgreSQL Large Objects by referring this link to store > and retrieve large files (As bytea not working) > https://www.postgresql.org/docs/current/largeobjects.html > > But even now I am unable to fetch the data at once from large objects > > select lo_get(oid); > > Here I'm getting the same error message. > > But if I use select data from pg_large_object where loid = 49374 > Then I can fetch the data but in page wise (data splitting into rows > of each size 2KB) > > So, here how can I fetch the data at single step rather than page by > page without any error. > > And I'm just wondering how do many applications storing huge amount of > data in GBs? I know that there is 1GB limit for each field set by > postgreSQL. If so, how to deal with these kind of situations? Would > like to know about this to deal with real time scenarios. > > We need to store large content (huge volume of data) and retrieve it. > Currently It is not happening due to limit of field size set by > postgreSQL. > > Would request to share your insights and suggestions on this to help > me for resolving this issue. > > > My first attempt at handling large payload was to use Java Selector directly in my app. This worked but manually chunking the data was tricky. I switched to using Tomcat and it handles large http(s) payloads seamlessly.
В списке pgsql-general по дате отправления: