It looks like it'll take 16 hours to load everything into postgres.
Now just parsing the text and logging the rows (30,000,000+ rows) takes like 30 minutes, at the moment I'm ingesting the data into postgres which does the same thing excepts batches stuff for postgres to ingestion. I'm at 4hrs and 34m and I've only ingested 5'400'000+ rows, that's a lot but that's like 1/6 of the total rows.
I think it's time we look at an inter process communication approach. I guess my distributed system course is coming in handy for once.
Is this an issue on the postgres end?
I don't think so? I looked at the usage and it looks like it's barely even at capacity?
It looks like it'll take 16 hours to load everything into postgres.
Now just parsing the text and logging the rows (30,000,000+ rows) takes like 30 minutes, at the moment I'm ingesting the data into postgres which does the same thing excepts batches stuff for postgres to ingestion. I'm at 4hrs and 34m and I've only ingested 5'400'000+ rows, that's a lot but that's like 1/6 of the total rows.
I think it's time we look at an inter process communication approach. I guess my distributed system course is coming in handy for once.
Is this an issue on the postgres end?
I don't think so? I looked at the usage and it looks like it's barely even at capacity?