Resume and JobRESUME AND JOB
Capgemini logo

GCP + Data Engineer

Capgemini

Software and Technology Jobs

GCP + Data Engineer

full-timePosted: Nov 18, 2025

Job Description

GCP + Data Engineer

📋 Job Overview

The GCP + Data Engineer role at Sogeti involves designing and developing robust data pipelines using the Hadoop ecosystem and GCP tools. The position requires implementing Spark/Scala/PySpark jobs for data transformation and aggregation, ensuring code quality through reviews and testing, and collaborating with business analysts for seamless production deployments. This role suits experienced professionals passionate about big data technologies and continuous innovation.

📍 Location: Bangalore

💼 Experience Level: Executives

🏢 Business Unit: Other Delivery units

🎯 Key Responsibilities

  • Design and develop data pipelines using Hadoop ecosystem and GCP tools
  • Implement Spark/Scala/PySpark jobs for data transformation and aggregation
  • Ensure code quality through peer reviews and unit testing
  • Collaborate with Business Analysts and manage smooth production deployments

✅ Required Qualifications

  • Experienced Data Engineer with 4+ years in designing and developing data pipelines using Hadoop ecosystem, Spark, Scala, and PySpark
  • Skilled in data transformation, aggregation, and distributed processing
  • Strong knowledge of Hive, Impala, and data warehousing methodologies
  • Proficient in CI/CD tools (Git, Jenkins), Agile/DevOps practices, and Cloudera distribution
  • Hands-on experience with Kafka, Spark Streaming, and job scheduling tools like Autosys
  • Adept at troubleshooting and production deployments

⭐ Preferred Qualifications

  • Expertise in GCP
  • Professional Data Engineer certification
  • Familiarity with Python/Java for pipeline development

🛠️ Required Skills

  • Hadoop ecosystem
  • GCP tools
  • Spark
  • Scala
  • PySpark
  • Hive
  • Impala
  • Data warehousing methodologies
  • CI/CD tools
  • Git
  • Jenkins
  • Agile practices
  • DevOps practices
  • Cloudera distribution
  • Kafka
  • Spark Streaming
  • Autosys
  • Troubleshooting
  • Production deployments
  • Python
  • Java
  • Distributed processing
  • Data transformation
  • Aggregation
  • Peer reviews
  • Unit testing
  • Collaboration with Business Analysts

Locations

  • Bangalore, India

Salary

Estimated Salary Rangemedium confidence

2,500,000 - 4,200,000 INR / yearly

Source: ai estimated

* This is an estimated range based on market data and may vary based on experience and qualifications.

Skills Required

  • Hadoop ecosystemintermediate
  • GCP toolsintermediate
  • Sparkintermediate
  • Scalaintermediate
  • PySparkintermediate
  • Hiveintermediate
  • Impalaintermediate
  • Data warehousing methodologiesintermediate
  • CI/CD toolsintermediate
  • Gitintermediate
  • Jenkinsintermediate
  • Agile practicesintermediate
  • DevOps practicesintermediate
  • Cloudera distributionintermediate
  • Kafkaintermediate
  • Spark Streamingintermediate
  • Autosysintermediate
  • Troubleshootingintermediate
  • Production deploymentsintermediate
  • Pythonintermediate
  • Javaintermediate
  • Distributed processingintermediate
  • Data transformationintermediate
  • Aggregationintermediate
  • Peer reviewsintermediate
  • Unit testingintermediate
  • Collaboration with Business Analystsintermediate

Required Qualifications

  • Experienced Data Engineer with 4+ years in designing and developing data pipelines using Hadoop ecosystem, Spark, Scala, and PySpark (experience)
  • Skilled in data transformation, aggregation, and distributed processing (experience)
  • Strong knowledge of Hive, Impala, and data warehousing methodologies (experience)
  • Proficient in CI/CD tools (Git, Jenkins), Agile/DevOps practices, and Cloudera distribution (experience)
  • Hands-on experience with Kafka, Spark Streaming, and job scheduling tools like Autosys (experience)
  • Adept at troubleshooting and production deployments (experience)

Preferred Qualifications

  • Expertise in GCP (experience)
  • Professional Data Engineer certification (experience)
  • Familiarity with Python/Java for pipeline development (experience)

Responsibilities

  • Design and develop data pipelines using Hadoop ecosystem and GCP tools
  • Implement Spark/Scala/PySpark jobs for data transformation and aggregation
  • Ensure code quality through peer reviews and unit testing
  • Collaborate with Business Analysts and manage smooth production deployments

Target Your Resume for "GCP + Data Engineer" , Capgemini

Get personalized recommendations to optimize your resume specifically for GCP + Data Engineer. Takes only 15 seconds!

AI-powered keyword optimization
Skills matching & gap analysis
Experience alignment suggestions

Check Your ATS Score for "GCP + Data Engineer" , Capgemini

Find out how well your resume matches this job's requirements. Get comprehensive analysis including ATS compatibility, keyword matching, skill gaps, and personalized recommendations.

ATS compatibility check
Keyword optimization analysis
Skill matching & gap identification
Format & readability score

Tags & Categories

Other Delivery unitsSoftware EngineeringExecutivesOther Delivery units

Answer 10 quick questions to check your fit for GCP + Data Engineer @ Capgemini.

Quiz Challenge
10 Questions
~2 Minutes
Instant Score

Related Books and Jobs

No related jobs found at the moment.

Capgemini logo

GCP + Data Engineer

Capgemini

Software and Technology Jobs

GCP + Data Engineer

full-timePosted: Nov 18, 2025

Job Description

GCP + Data Engineer

📋 Job Overview

The GCP + Data Engineer role at Sogeti involves designing and developing robust data pipelines using the Hadoop ecosystem and GCP tools. The position requires implementing Spark/Scala/PySpark jobs for data transformation and aggregation, ensuring code quality through reviews and testing, and collaborating with business analysts for seamless production deployments. This role suits experienced professionals passionate about big data technologies and continuous innovation.

📍 Location: Bangalore

💼 Experience Level: Executives

🏢 Business Unit: Other Delivery units

🎯 Key Responsibilities

  • Design and develop data pipelines using Hadoop ecosystem and GCP tools
  • Implement Spark/Scala/PySpark jobs for data transformation and aggregation
  • Ensure code quality through peer reviews and unit testing
  • Collaborate with Business Analysts and manage smooth production deployments

✅ Required Qualifications

  • Experienced Data Engineer with 4+ years in designing and developing data pipelines using Hadoop ecosystem, Spark, Scala, and PySpark
  • Skilled in data transformation, aggregation, and distributed processing
  • Strong knowledge of Hive, Impala, and data warehousing methodologies
  • Proficient in CI/CD tools (Git, Jenkins), Agile/DevOps practices, and Cloudera distribution
  • Hands-on experience with Kafka, Spark Streaming, and job scheduling tools like Autosys
  • Adept at troubleshooting and production deployments

⭐ Preferred Qualifications

  • Expertise in GCP
  • Professional Data Engineer certification
  • Familiarity with Python/Java for pipeline development

🛠️ Required Skills

  • Hadoop ecosystem
  • GCP tools
  • Spark
  • Scala
  • PySpark
  • Hive
  • Impala
  • Data warehousing methodologies
  • CI/CD tools
  • Git
  • Jenkins
  • Agile practices
  • DevOps practices
  • Cloudera distribution
  • Kafka
  • Spark Streaming
  • Autosys
  • Troubleshooting
  • Production deployments
  • Python
  • Java
  • Distributed processing
  • Data transformation
  • Aggregation
  • Peer reviews
  • Unit testing
  • Collaboration with Business Analysts

Locations

  • Bangalore, India

Salary

Estimated Salary Rangemedium confidence

2,500,000 - 4,200,000 INR / yearly

Source: ai estimated

* This is an estimated range based on market data and may vary based on experience and qualifications.

Skills Required

  • Hadoop ecosystemintermediate
  • GCP toolsintermediate
  • Sparkintermediate
  • Scalaintermediate
  • PySparkintermediate
  • Hiveintermediate
  • Impalaintermediate
  • Data warehousing methodologiesintermediate
  • CI/CD toolsintermediate
  • Gitintermediate
  • Jenkinsintermediate
  • Agile practicesintermediate
  • DevOps practicesintermediate
  • Cloudera distributionintermediate
  • Kafkaintermediate
  • Spark Streamingintermediate
  • Autosysintermediate
  • Troubleshootingintermediate
  • Production deploymentsintermediate
  • Pythonintermediate
  • Javaintermediate
  • Distributed processingintermediate
  • Data transformationintermediate
  • Aggregationintermediate
  • Peer reviewsintermediate
  • Unit testingintermediate
  • Collaboration with Business Analystsintermediate

Required Qualifications

  • Experienced Data Engineer with 4+ years in designing and developing data pipelines using Hadoop ecosystem, Spark, Scala, and PySpark (experience)
  • Skilled in data transformation, aggregation, and distributed processing (experience)
  • Strong knowledge of Hive, Impala, and data warehousing methodologies (experience)
  • Proficient in CI/CD tools (Git, Jenkins), Agile/DevOps practices, and Cloudera distribution (experience)
  • Hands-on experience with Kafka, Spark Streaming, and job scheduling tools like Autosys (experience)
  • Adept at troubleshooting and production deployments (experience)

Preferred Qualifications

  • Expertise in GCP (experience)
  • Professional Data Engineer certification (experience)
  • Familiarity with Python/Java for pipeline development (experience)

Responsibilities

  • Design and develop data pipelines using Hadoop ecosystem and GCP tools
  • Implement Spark/Scala/PySpark jobs for data transformation and aggregation
  • Ensure code quality through peer reviews and unit testing
  • Collaborate with Business Analysts and manage smooth production deployments

Target Your Resume for "GCP + Data Engineer" , Capgemini

Get personalized recommendations to optimize your resume specifically for GCP + Data Engineer. Takes only 15 seconds!

AI-powered keyword optimization
Skills matching & gap analysis
Experience alignment suggestions

Check Your ATS Score for "GCP + Data Engineer" , Capgemini

Find out how well your resume matches this job's requirements. Get comprehensive analysis including ATS compatibility, keyword matching, skill gaps, and personalized recommendations.

ATS compatibility check
Keyword optimization analysis
Skill matching & gap identification
Format & readability score

Tags & Categories

Other Delivery unitsSoftware EngineeringExecutivesOther Delivery units

Answer 10 quick questions to check your fit for GCP + Data Engineer @ Capgemini.

Quiz Challenge
10 Questions
~2 Minutes
Instant Score

Related Books and Jobs

No related jobs found at the moment.