Principal Consultant-Senior Data Engineer


Job Details

With a startup spirit and 115,000 + curious and courageous minds, we have the expertise to go deep with the world s biggest brands and we have fun doing it! We dream in digital, dare in reality, and reinvent the ways companies work to make an impact far bigger than just our bottom line. We re harnessing the power of technology and humanity to create meaningful transformation that moves us forward in our pursuit of a world that works better for people.

Now, we re calling upon the thinkers and doers, those with a natural curiosity and a hunger to keep learning, keep growing. People who thrive on fearlessly experimenting, seizing opportunities, and pushing boundaries to turn our vision into reality. And as you help us create a better world, we will help you build your own intellectual firepower.

Welcome to the relentless pursuit of better.

Inviting applications for the role of Principal Consultant-Senior Data Engineer

Include optimizing data pipelines, ensuring data integrity and consistency, enhancing system resiliency where applicable, maintaining and improving data security, proactive alerting, and monitoring for data pipelines, and automating repetitive data-oriented tasks.

Responsibilities

  • Automate data tasks on GCP.
  • Work with data domain owners, data scientists and other stakeholders to that data is consumed effectively on GCP.
  • Design, build, secure and maintain data infrastructure, including data pipelines, databases, data warehouses, and data processing platforms on GCP.
  • Measure and monitor the quality of data on GCP data platforms.
  • Implement robust monitoring and alerting systems to proactively identify and resolve issues in data systems. Respond to incidents promptly to minimize downtime and data loss.
  • Develop automation scripts and tools to streamline data operations and make them scalable to ensure accommodate growing data volumes and user traffic.
  • Optimize data systems to ensure efficient data processing, reduce latency, and improve overall system performance.
  • Collaborate with data and infrastructure teams to forecast data growth and plan for future capacity requirements.
  • Ensure data security and compliance with data protection regulations. Implement best practices for data access controls and encryption.
  • Collaborate with data engineers, data scientists, and software engineers to understand data requirements, troubleshoot issues, and support data-driven initiatives.
  • Continuously assess and improve data infrastructure and data processes to enhance reliability, efficiency, and performance.
  • Maintain clear and up-to-date documentation related to data systems, configurations, and standard operating procedures.

Qualifications we seek in you!

Minimum Qualifications / Skills

  • Bachelor s or master s degree in computer science, Software Engineering, Data Science or related field, or equivalent practical experience

Preferred Qualifications/ Skills

  • Proficiency in data technologies, such as relational databases, data warehousing, big data platforms (e.g., Hadoop, Spark), data streaming (e.g., Kafka), and cloud services (e.g., AWS, GCP, Azure).
  • Strong programming skills in languages like Python (numpy, pandas, pyspark), Java (Core Java, Spark with Java, functional interface, lambda, java collections), or Scala, with experience in automation and scripting.
  • Experience with containerization and orchestration tools like Docker and Kubernetes is a plus.
  • Experience with data governance(data plex), data security, and compliance best practices on GCP.
  • Solid understanding of software development methodologies and best practices, including version control (e.g., Git) and CI/CD pipelines.
  • Strong background in cloud computing and data-Intensive applications and services, with a focus on Google Cloud Platform.
  • Experience with data quality assurance and testing on GCP.
  • Proficiency with GCP data services (BigQuery; Dataflow; Data Fusion; Dataproc; Cloud Composer; Pub/Sub; Google Cloud Storage).
  • Strong understanding of logging and monitoring using tools such as Cloud Logging, ELK Stack, AppDynamics, New Relic, Splunk, etc.
  • Knowledge of AI and ML tools is a plus.
  • Google Associate Cloud Engineer or Data Engineer certification is a plus.
  • Experience in data engineering or data science on GCP.

The approximate annual base compensation range for this position is [$65,000 to $75,000]. The actual offer, reflecting the total compensation package plus benefits, will be determined by a number of factors which include but are not limited to the applicant s experience, knowledge, skills, and abilities; geographic location; and internal equity.

Genpact is an Equal Opportunity Employer and considers applicants for all positions without regard to race, color, religion or belief, sex, age, national origin, citizenship status, marital status, military/veteran status, genetic information, sexual orientation, gender identity, physical or mental disability or any other characteristic protected by applicable laws. Genpact is committed to creating a dynamic work environment that values diversity and inclusion, respect and integrity, customer focus, and innovation. For more information, visit www.genpact.com . Follow us on Twitter, Facebook, LinkedIn, and YouTube.

Furthermore, please do note that Genpact does not charge fees to process job applications and applicants are not required to pay to participate in our hiring process in any other way. Examples of such scams include purchasing a 'starter kit,' paying to apply, or purchasing equipment or training.





 GENPACT

 06/01/2024

 Chicago,IL