What you will be doing:
- Develop and maintain automated real-time and batch data pipelines;
- Build automation around various data services;
- Contribute into a Data Profiling and Anomaly detection tools development;
- Work closely with Engineering Teams on their software and deploy requirements;
- Drive innovation into Data Platform architecture and software stack.
What we need to see:
- A Bachelor’s degree in Computer Science, Engineering, or a related field;
- 5+ years of data engineering experience;
- Strong Python programming skills;
- Understanding of the operations of relational DBMS and distributed systems;
- Proficiency in big data tools (e.g., Hadoop, Spark, Kafka) and data pipeline/workflow management tools;
- Solid SQL knowledge;
- Fluent English.
Ways to stand out from the crowd:
- Knowledge of columnar and NoSQL databases;
- Experience with cloud services like AWS (EC2, EMR, RDS, Redshift);
- PySpark/Scala Spark Experience and ability to build solutions with it;
- Clickhouse knowledge.
- Proficiency in working with GPU-accelerated databases (e.g., HeavyDB, Kinetica);
Job Category: python
Job Type: Full Time
Job Location: Chennai
Total Experience: 5+ Years