CTC- 15-50 LPA(Based on experience)
Location-Bangalore
We are seeking a skilled Data Engineer with strong expertise in Python, SQL, and modern big data technologies to design and build scalable, efficient data flows and pipelines in a cloud environment. The ideal candidate will have hands-on experience with real-time data ingestion using Kafka, proficiency in Spark for data processing, and familiarity with AWS cloud services such as Glue and Athena. This role demands strong programming skills, experience with ETL frameworks, version control proficiency, and a proactive, collaborative mindset for seamless delivery in Agile teams.
Key Responsibilities
- Develop, design, and optimize data processing pipelines leveraging Python, SQL, and Spark.
- Architect scalable and repeatable data flows and management solutions in a Cloud environment.
- Utilize Kafka platform for real-time, high-volume data ingestion and processing.
- Implement infrastructure-as-code practices for deployment consistency and reliability.
- Maintain and use version control systems such as GIT; develop and update technical documentation.
- Collaborate effectively with cross-functional teams, ensuring strong communication and agile delivery.
- Apply Agile, Scrum, or Kanban methodologies to manage product delivery efficiently.
Qualifications
- Bachelor’s or Master’s degree in Computer Science, Information Systems, or related field.
- Minimum 4 years of professional experience building data pipelines and managing big data technologies.
- Strong hands-on experience with ETL frameworks (e.g., Airflow, Jenkins) for production deployment.
- Deep understanding of data structures and willingness to adopt new technologies.
- Experience with AWS cloud tools, including Glue and Athena.
- Proficient in English, with excellent communication skills.
- Demonstrated ability to deliver data engineering solutions with a product-oriented mindset.