360itprofessionals1
Big Data Engineer (W2 Candidates only)
Company
Role
Big Data Engineer (W2 Candidates only)
Location
Job type
Contract
Posted
105 months ago
Salary
Job description
- § Lead Big Data engineering team with specialization on data ingestion (from 100+ source systems in batch and near real-time), egestion and governance. § Participate in collaborative software and system design and development of the new NCR Data Lake on Hortonworks HDP and HDF distributions. § Manage own learning and contribute to technical skill building of the team. § Inspire and cultivate the engineering mindset and systems thinking. § Gain deep technical expertise in the data movement patterns, practices and tools. § Play active role in Big Data Communities of Practice. § Put the minimal system needed into production. Required Qualifications § Bachelor’s degree or higher in Computer Science or a related field. § Good understanding of distributed computing and big data architectures. § Passion for software engineering and craftsman-like coding prowess. § Proven experience in developing Big Data solutions in Hadoop Ecosystem using Apache NiFi, Kafka, Flume, Sqoop, Apache Atlas, Hive, HDFS, HBase and Spark (Hortonworks HDP and HDF preferred). § Experience with at least one of the leading CDC (Change Data Capture) tools like Informatica PowerCenter. § Development experience with at least one NoSQL database. HBase or Cassandra preferred. § Polyglot development (4-5 years+): Capable of developing in Java and Scala with good understanding of functional programming, SOLID principles and, concurrency models and modularization. § DevOps: Appreciates the CI and CD model and always builds to ease consumption and monitoring of the system. Experience with Maven (or Gradle or SBT) and Git preferred. § Experience in Agile development including Scrum and other lean techniques. § Should believe in You Build! You Ship! And You Run! Philosophy. § Personal qualities such as creativity, tenacity, curiosity, and passion for deep technical excellence. Desired Qualifications § Experience with Big Data migrations/transformations programs in the Data Warehousing and/or Business Intelligence areas. § Experience with ETL tools like Talend, Pentaho, Attunity etc. § Knowledge of Teradata, Netezza etc. § Good grounding in NoSQL data stores such as Cassandra, Neo4j etc. § Strong knowledge on computer algorithms. § Experience with workload orchestration and automation tools like Oozie, Control-M etc. § Experience in building self-contained applications using Docker, Vagrant. Chef.
Thanks and Regards,
Ankur Bhatia


