Thank you all for your response.
Show quoted text
1. What version of PG is it? ("SELECT VERSION();" should tell you.)
PG Version 16
2. Are you truncating the table after each test run, or deleting all records, or appending?
created new schema for each run.
3. Is the blob data stored in BYTEA column data, or are you using the (discouraged) "Large Objects"?
Blob storage
4. How are you loading the blob data?
used the Import data using a COPY statement, followed this doc
On Sun, Dec 15, 2024, 10:22 PM Rui DeSousa <rui.desousa@xxxxxxxxxx> wrote:
> On Dec 15, 2024, at 8:22 PM, Anex Hul <anexsql2014@xxxxxxxxx> wrote:
>
> Hello everyone,
>
> Testing 100 million records data import from Azure blob storage to Azure postgresql. I did run the test 5 times and the time it took keep increasing for each run.
> Is there know justification for this linear increment of the time it took for same size of data?
Check you I/O quotas; you might have hit quota limits and being throttled.