Problem Description:
I am using a multi-tenant database and need to synchronize data from multiple tables in Postgres to Elasticsearch. However, if I write all the content in the same file, it may exceed the file size limit. Therefore, I want to split it into many small files to avoid any issues.
Please help me.
Apologies for the delay. Can you share your schema either publicly or privately?
The schema can be split based on database/index pair.
How large are we talking about here?
PGSync version: 2.5.0
Postgres version: 15.1
Elasticsearch version: 8.7.0
Redis version: 5.0.7
Python version: Python3: 3.8.10
Problem Description: I am using a multi-tenant database and need to synchronize data from multiple tables in Postgres to Elasticsearch. However, if I write all the content in the same file, it may exceed the file size limit. Therefore, I want to split it into many small files to avoid any issues. Please help me.