Data Engineer

Location: Gurugram/Remote

Job ID: DE2008

Job Type: Full time

Experience: 2-6 years

Why would you like to join us?

TransOrg Analytics, an award winning - Big Data and Predictive analytics- company, offers advanced analytics solutions to industry leaders and Fortune 500 companies across India, US, UK, Singapore and the Middle East. Our products - Clonizo' (customer cloning) have yielded significant incremental benefits to our clients. We have been recognized by the CIO Review magazine as the - Predictive Analytics Company of the Year- and by TiE for excellence in entrepreneurship.

Overview

This position is for a Data Engineer with an experience in SQL/NoSQL, Cloud, Hadoop and/or Spark platforms

Responsibilities

  • Design and implement data engineering projects.
  • Integrate multiple data sources to create data lake/data mart Perform data ingestion and ETL processes using SQL, Scoop, Spark or Hive
  • Knowledge of new components and various emerging technologies in on-premises and Cloud (AWS/Azure/Google)
  • Collaborate with various cross-functional teams: infrastructure, network and database
  • Work with various teams to setup and manage users, secure and govern platforms and data and maintain business continuity through contingency plans (data archiving etc.)
  • Monitor job performances, manage file system/disk-space, cluster & database connectivity, log files, manage backup/security and troubleshoot various user issues
  • Design, implement, test and document performance benchmarking strategy for platforms as well as for different use cases
  • Setup, administer, monitor, tune, optimize and govern large scale implementations
  • Implement machine learning models on real time input data stream
  • Drive customer communication during critical events and participate/lead various operational improvement initiatives

Education & Skills Summary

  • 2 - 6 years relevant experience in data engineering
  • Exposure to any or all latest data engineering ecosystem platforms such as AWS, Azure, GCP, Cloudera and Data bricks
  • Sound knowledge of Python/Scala/Java
  • Good knowledge of SQL / NoSQL databases and data warehouse concepts
  • Hands on experience of working on databases such as Sql Servers, PostgreSql, Cloud infrastructure, etc.
  • Excellent knowledge of data backup, recovery, security and integrity
  • Sound knowledge on Spark, HDFS/HIVE/HBASE, Shell Scripting, and Spark Streaming
  • Excellent communication skills
  • Must be proficient with data ingestion tools like Sqoop, flume, talend, and Kafka

To apply for any open positions please share your updated CV along with job id at: careers@transorg.com

To apply for any open positions please share your updated CV along with job id at: