I am creating a Job just reading from an oracle source (500,000 rows) and writing directly in a csv. It given me Timeout error after writng 50,000 lines, when i increased xms to 512m & xmx to 512g it gives below error message after reading all lines but it took around 2 hours to read all lnes. we have 48 gb RAM in the machine and no other users are there. when i am executing the DB source it's working fine and returning output in some seconds.
Any help will be appreciated.
Exception in thread "main" java.lang.OutOfMemoryError
at java.lang.AbstractStringBuilder.hugeCapacity(Unknown Source)
at java.lang.AbstractStringBuilder.newCapacity(Unknown Source)
at java.lang.AbstractStringBuilder.ensureCapacityInternal(Unknown Source)
at java.lang.AbstractStringBuilder.append(Unknown Source)
at java.lang.StringBuilder.append(Unknown Source)
You can increase the heap from the Run->Advanced settings tab
If you see my earlier mail, i already did the same changes but still it's throwing error after reading all data.
More important its taking so much time to read the data.
Talend named a Leader.
Kickstart your first data integration and ETL projects.
Learn how to do cool things with Context Variables
Find out how to migrate from one database to another using the Dynamic schema
Pick up some tips and tricks with Context Variables