Your browser does not support javascript! Please enable it, otherwise web will not work for you.

GCP Datalake - Senior Engineer @ Iris Software

Home > Software Development






Iris Software  GCP Datalake - Senior Engineer

Job Description

1. Data Engineering:

Cloud Platform Expertise: Strong knowledge of GCP (Google Cloud Platform) services, including Composer, Data Fusion, BigQuery, Cloud Storage, and Dataflow.

Data Pipeline Development: Experience in building and maintaining data pipelines using technologies like Apache Airflow, Python, SQL, and potentially other ETL tools.

Data Modeling Schema Design: Ability to design and implement efficient data models and schemas for data ingestion, storage, and processing.

Data Quality Validation: Expertise in ensuring data quality, validating data integrity, and implementing data quality checks within the pipeline.

Troubleshooting Debugging: Proficiency in identifying and resolving issues within the pipeline, including performance bottlenecks, data inconsistencies, and error handling.

CI/CD Automation: Experience with continuous integration and continuous delivery (CI/CD) pipelines for automating data pipeline deployments and updates.

2. Data Integration Connectivity:

API Integration: Expertise in integrating with various APIs (e.g., Salesforce, Workday, Peoplesoft, Siplast, etc.) and understanding API security, authentication, and data formats.

Data Transformation Manipulation: Skill in transforming data using different methods (e.g., data cleaning, aggregation, filtering, etc.) and applying appropriate data transformation techniques.

Database Technologies: Experience with relational databases

3. Application Platform Support:

Troubleshooting Problem Solving: Ability to identify and resolve issues related to application performance, data integration, and platform stability.

Monitoring Alerting: Experience in setting up monitoring systems to track pipeline performance, data quality, and potential failures.

Incident Management: Skill in handling and managing incidents, communicating updates to stakeholders, and following incident resolution procedures.

Documentation Communication: Clear communication skills and ability to document procedures, best practices, and technical documentation for the team and stakeholders.

4. Domain Expertise:

Business Understanding: A deep understanding of the business processes and data requirements related to the specific data pipelines.

Data Understanding: Knowledge of the data structures, formats, and relationships within the various source systems.

Technical Collaboration: Ability to effectively collaborate with other teams (e.g., business analysts, data analysts, application developers) to understand requirements and troubleshoot issues.

Prior experience in SAP, Salesforce etc esp finance domain

5. Additional Skills:

Version Control (Git): Understanding and experience with version control systems like Git for managing code and pipeline changes.

Cloud Security: Knowledge of cloud security best practices and implementing security measures for data pipelines and cloud infrastructure.

Mandatory Competencies
Cloud - GCP - Cloud Build, Google Cloud Deploy, Dataflow, Cloud Run
Cloud - GCP - Cloud Data Fusion, Dataproc, BigQuery, Cloud Composer, Cloud Bigtable
Cloud - Azure - Azure Data Factory (ADF), Azure Databricks, Azure Data Lake Storage, Event Hubs, HDInsight
Development Tools and Management - Development Tools and Management - CI/CD
Programming Language - Python - Apache Airflow
Database - Database Programming - SQL
DevOps/Configuration Mgmt - DevOps/Configuration Mgmt - Git
Enterprise Applications - ERP - PeopleSoft
Beh - Communication and collaboration

Job Classification

Industry: IT Services & Consulting
Functional Area / Department: Engineering - Software & QA
Role Category: Software Development
Role: Data Platform Engineer
Employement Type: Full time

Contact Details:

Company: Iris Software
Location(s): Noida, Gurugram

+ View Contactajax loader


Keyskills:   gcp python cloud security sap google cloud platform airflow databricks logistics sql alerting salesforce data quality git data modeling data cleaning devops xml debugging bigquery etl data integration finance data lake azure

 Fraud Alert to job seekers!

₹ Not Disclosed

Similar positions

Hcltech - Java Full Stack (f2f Hiring Drive-18 Apr)

  • HCLTech
  • 8 - 13 years
  • Hyderabad
  • 14 hours ago
₹ Not Disclosed

Data Engineer-Data Platforms-Snowflake

  • IBM
  • 5 - 8 years
  • Bengaluru
  • 15 hours ago
₹ Not Disclosed

Custom Software Engineer

  • Accenture
  • 7 - 12 years
  • Noida, Gurugram
  • 17 hours ago
₹ Not Disclosed

Data Platform Engineer

  • Accenture
  • 2 - 5 years
  • Bengaluru
  • 18 hours ago
₹ Not Disclosed

Iris Software

Iris Software Inc.