One Star

How to efficiently load 30mil rows from Kafka into Postgresql w/TOS?

Hi Team,
I have a scenario where i want to load 30million rows of data using Kafka into PostgreSQL. My job is going to be quite straightforward with few filters and tmap. 
I would like to know:
1. What are the best configuration settings for TOS in such scenario?
2. how can i reduce cpu/memory consumptions?
3. what is the most efficient way to load huge data using TOD? Shall i do it in one go or in batches?
4. And also, what could be the average time to load such huge data using TOS?
Thanks in advance.
Rera
1 REPLY
Moderator

Re: How to efficiently load 30mil rows from Kafka into Postgresql w/TOS?

Hi Rera,
I have a scenario where i want to load 30million rows of data using Kafka into PostgreSQL. My job is going to be quite straightforward with few filters and tmap.

tMap is cache component consuming much memory. For a large set of data, try to store the data on disk instead of memory.
Here is an option "Use Batch Size" in tPostgresqloutput which is used to activate the batch mode for data processing.
What does your filter look like? What's your current row rate(rows/s)? Is it a normal speed?
Please take a look at document about:TalendHelpCenter:Exception outOfMemory
Best regards
Sabrina
--
Don't forget to give kudos when a reply is helpful and click Accept the solution when you think you're good with it.