Hello, I have a series of files that need to be loaded into a Vertica table from a Vertica Query and these row numbers range from 100 million to 300 million rows (possibly more) per file (10GB to 30GB). I have devised the attached mapping to do this. The insert is a query with some joins to other tables. The tJava's you can see are really for logging and monitoring and merely show the time this stage of the job has run. The tMap is there to allow 4 millions rows to commit at a time. The Question There isn't much else to it really. It takes about 45 minutes to load the data to the tVerticaOutputBulkExec which I am happy with. I will also be happy if it takes less than 5 hours to load the data to the table but I was wondering if I could improve on the tVerticeBulkExec and insert the rows faster?