Software Engineer II - Data Engineer - Spark, Python, Databricks or AWS EMR
Join JPMorgan Chase & Co. in Bengaluru / Bangalore as a Software Engineer II - Data Engineer. This permanent role offers an exciting opportunity to grow your skills and advance your career in the Commercial & Investment Bank.
As an emerging member of our agile software engineering team, you will be instrumental in enhancing, designing, and delivering secure, stable, and scalable technology products. You'll execute software solutions through the design, development, and technical troubleshooting of multiple components within our technical landscape.
Job Responsibilities
- Design, develop, and maintain scalable data pipelines and ETL processes.
- Work with large datasets using Spark on Databricks or AWS EMR.
- Write efficient SQL queries for data extraction, transformation, and analysis.
- Collaborate with data scientists, analysts, and other engineering teams to deliver high-quality data solutions.
- Implement data processing workflows on AWS services such as S3, ECS, Lambda, EMR, and Glue.
- Develop and maintain Python scripts for data processing and automation.
- Ensure data quality, integrity, and security across all data engineering activities.
- Troubleshoot and resolve data-related issues in a timely manner.
Required Qualifications and Skills
- Formal training or certification on software engineering concepts and 2+ years applied experience.
- Proven expertise in Data Engineering with Spark.
- Hands-on experience with Databricks or AWS EMR.
- Strong knowledge of SQL and database concepts.
- Experience in ETL and data processing workflows.
- Proficiency in AWS services: S3, ECS, Lambda, EMR/Glue.
- Advanced skills in Python programming.
- Excellent problem-solving and analytical abilities.
- Bachelor's degree in Computer Science, Information Technology, or related field (or equivalent experience).
- Strong communication and collaboration skills.
- Ability to work independently and as part of a team.
Preferred Qualifications and Skills
- Experience with Infrastructure as Code (IaaC) using Terraform or CloudFormation.
- Familiarity with writing unit test cases for Python code.
- Knowledge of version control systems such as BitBucket or GitHub.
- Understanding of CI/CD pipelines and automation tools.
