We are faced with reading a huge number of hive tables with different schemas. So we were hoping we can develop a single talend job to read these different tables. We recently found out that big data components do not support dynamic schema/type. Does anybody know of a possible solution or work around towards changing schemas with big data components?
Are you in Standard DI job or Bigdata batch job?
Here is a new feature jira issue about "Support Hive Metastore HA".
There's an article "How to operationalize meta-data in Talend with Dynamic Schemas" by Robert Griswold which mentions you can use a String for big data components since Dynamic Schema is not supported by Big Data components.
The blog however doesn't mention how this can be done, do you know a way?
Talend named a Leader.
Kickstart your first data integration and ETL projects.
Watch the recorded webinar!
Learn how to make your data more available, reduce costs and cut your build time
Read about OTTO's experiences with Big Data and Personalized Experiences
Take a look at this video about Talend Integration with Databricks