Minimum of 5 years hands-on experience with Data
Integration, ETL and/or Data Engineering
Minimum of 5 years hands-on experience with Big Data tools
and techniques
Minimum 3 years of hands-on experience with Talend used in
conjunction of Hadoop MapReduce/Spark/Hive.
Ability and willingness to be hands-on including use of SQL,
ETL tools, scripts, etc. to tune performance, data transformation routines,
etc. when necessary.
Creating data flow documentation and business process
modeling
Data integration design and development
Leveraging a Big Data capability and understanding the
schema-on-read approach and implications
Hands-on experience with Hadoop ecosystem (Cloudera
preferred) including MapReduce, Spark, Hive/Impala, HBase, Kafka
Extensive experience with building data pipelines and
implementing complex data processing using Talend in conjunction with the Big
Data platform
Hands-on experience with programming languages such as SQL,
python, Scala, Java, R, Unix shell scripts frequently used to build workloads
on Hadoop platforms.