Big data query and insert to Vertica

One Star

Big data query and insert to Vertica

Hello, 
I have a series of files that need to be loaded into a Vertica table from a Vertica Query and these row numbers range from 100 million to 300 million rows (possibly more) per file (10GB to 30GB). I have devised the attached mapping to do this.
The insert is a query with some joins to other tables. The tJava's you can see are really for logging and monitoring and merely show the time this stage of the job has run. The tMap is there to allow 4 millions rows to commit at a time.
The Question
There isn't much else to it really.  It takes about 45 minutes to load the data to the tVerticaOutputBulkExec which I am happy with.  I will also be happy if it takes less than 5 hours to load the data to the table but I was wondering if I could improve on the tVerticeBulkExec and insert the rows faster?
Highlighted
Employee

Re: Big data query and insert to Vertica

Are you using COPY under "Action on data" ?
How large is your 4 million rows in size on disk?
Take a look at Talend Vertica Tips & Techniques.

2019 GARNER MAGIC QUADRANT FOR DATA INTEGRATION TOOL

Talend named a Leader.

Get your copy

OPEN STUDIO FOR DATA INTEGRATION

Kickstart your first data integration and ETL projects.

Download now

What’s New for Talend Summer ’19

Watch the recorded webinar!

Watch Now

Best Practices for Using Context Variables with Talend – Part 4

Pick up some tips and tricks with Context Variables

Blog

How Media Organizations Achieved Success with Data Integration

Learn how media organizations have achieved success with Data Integration

Read

Why Companies Move to the Cloud: 7 Success Stories

Learn how and why companies are moving to the Cloud

Read Now