
Java GCP Data Engineer
Full Job Description
Impetus Technologies, a distinguished digital engineering firm, is seeking a talented Java GCP Data Engineer for its Chennai, India location. This permanent role offers an exciting opportunity to contribute to cutting-edge data solutions. We are looking for professionals with 1-4 years of hands-on experience in developing robust, high-performance data systems, leveraging strong Java programming skills and deep expertise in Google Cloud Platform (GCP) services.
Key Responsibilities:
- Design, develop, and maintain scalable ETL/ELT data pipelines on GCP using Java-based big data frameworks.
- Build and optimize batch and streaming data processing solutions utilizing Dataproc, Dataflow (with Apache Beam in Java), and Spark.
- Implement best practices for high-quality, efficient, and maintainable code.
- Conduct unit and integration testing, and effectively troubleshoot complex data pipeline issues.
- Collaborate with cross-functional teams to define data requirements and deliver effective solutions.
- Provide accurate development effort estimations and contribute to sprint planning.
- Participate in code reviews and provide mentorship to junior engineers.
- Architect cost-optimized and performance-efficient solutions leveraging GCP-native services.
Essential Skills & Experience:
- Proficiency in Java for building distributed data processing applications.
- Hands-on experience with Big Data technologies: Apache Spark (Java/Scala APIs), Hadoop, Hive.
- Experience with Spark DataFrames and Spark SQL using Java or Scala (PySpark knowledge is a plus).
- Strong understanding of data structures, algorithms, and object-oriented programming in Java.
- Solid knowledge of SQL, data modeling, and data warehousing principles.
- Experience with Linux/Unix environments and scripting (Bash or similar).
- Proven analytical and problem-solving skills for debugging and optimizing data pipelines.
- Ability to design and implement scalable, fault-tolerant data processing systems.
Preferred Qualifications:
- Hands-on experience with GCP services including BigQuery, Dataflow, Dataproc, Cloud Storage, Pub/Sub, and IAM.
- Familiarity with workflow orchestration tools like Airflow or Cloud Composer.
- Experience in cloud migration projects, particularly from on-premise Hadoop to GCP.
- Exposure to streaming data pipelines using Pub/Sub and Dataflow.
- Understanding of CI/CD pipelines and DevOps practices in a cloud context.
For immediate consideration, please forward your resume along with your notice period, current CTC, and expected CTC to [HIDDEN TEXT].
Company
Impetus Technologies
Impetus Technologies is a leading digital engineering company dedicated to empowering enterprises in their transformation journeys. We specialize in unraveling complex challenges in analytics, AI, and...