I got a requirement to ingest data from mssql to hortonworks hive database (the requirement is to create hdfs file and table dynamically - without defining any schema). So I am planned to use tSqoopImport component. Now I am using Java API mode and able to load the data into HDFS file. but the problem here is its working only for text and sequence files, not working for avro and parquet files.
now the questions are, 1) Is it possible to work with the other file formats also?
2) Is it possible to change the hdfs text files delimiter (by default its taking "," comma)?
3) Is it possible to load the hive tables using the same tsqoopimport component?
Please have a look at this reference about:TalendHelpCenter:Which big data formats are supported.
If you want to load data to Hive in bigdata spark job, please have a look at this example shared on talend help center.
Introduction to Talend Open Studio for Data Integration.
Practical steps to developing your data integration strategy.
Create systems and workflow to manage clean data ingestion and data transformation.