I got a requirement to ingest data from mssql to hortonworks hive database (the requirement is to create hdfs file and table dynamically - without defining any schema). So I am planned to use tSqoopImport component. Now I am using Java API mode and able to load the data into HDFS file. but the problem here is its working only for text and sequence files, not working for avro and parquet files.
now the questions are, 1) Is it possible to work with the other file formats also?
2) Is it possible to change the hdfs text files delimiter (by default its taking "," comma)?
3) Is it possible to load the hive tables using the same tsqoopimport component?
Please have a look at this reference about:TalendHelpCenter:Which big data formats are supported.
If you want to load data to Hive in bigdata spark job, please have a look at this example shared on talend help center.
Talend named a Leader.
Kickstart your first data integration and ETL projects.
Watch the recorded webinar!
Learn how to make your data more available, reduce costs and cut your build time
Read about OTTO's experiences with Big Data and Personalized Experiences
Take a look at this video about Talend Integration with Databricks