Your browser does not support javascript! Please enable it, otherwise web will not work for you.

Lead Analyst - Azure Data Brick With Etl And Power Bi @ CGI

Home > Business Intelligence & Analytics

CGI  Lead Analyst - Azure Data Brick With Etl And Power Bi

Job Description

The ETL Lead Developer will be responsible for designing, developing, and implementing robust and scalable data architectures and ETL solutions. This role requires deep expertise in Databricks for data processing and Lakehouse architecture, Kimball dimensional modelling for data warehousing, and Power BI and Microsoft Fabric for data visualization and analytics platform implementation. As a Lead ETL Developer, you will lead the design and implementation of scalable, secure, and high-performance data solutions using the Databricks Lakehouse Platform. You will work with ETL Architects to define architectural standards, guide engineering teams, and collaborate with stakeholders to align data strategies with business goals. Your role will focus on leveraging Databricks' unified analytics capabilities to build enterprise-grade data platforms that support advanced analytics, machine learning, and real-time data processing.

Your future duties and responsibilities:

Develop and own the end-to-end architecture of data platforms built on Databricks, including ingestion, transformation, storage, and consumption layers.
. Develop and maintain data models, data flow diagrams, and other architectural documentation based on Kimball dimensional modelling principles.
. Design and implement end-to-end data architectures leveraging Microsoft Fabric's capabilities ( Embedded Power BI integration/ implementation).
. Design and implement Lakehouse architectures using Delta Lake, Unity Catalog, and structured streaming.
. Proven expertise in implementing data governance using Unity Catalog, including fine-grained access control, column-level lineage, data classification, audit logging, and centralized metadata management across workspaces and cloud environments
. Develop scalable ETL/ELT pipelines using Apache Spark, PySpark, and Databricks Workflows.
. Development of integration of Databricks with enterprise systems such as data catalogs, data quality frameworks, ML platforms, and BI tools.
. Design and development of high-performance reporting models and paginated reports, configurable inquiries and interactive dashboards using Power BI.
. Guide and implement teams in implementing CI/CD pipelines, version control, and automated testing for Databricks notebooks and jobs.
. Provide technical leadership in performance tuning, cost optimization, and cluster configuration.
. Participate in architectural reviews, code audits, and mentoring sessions to ensure adherence to standards and scalability.
. Provide technical leadership and guidance to data engineers and developers.
. Collaborate closely with clients, business stakeholders, and internal teams to translate business requirements into technical solutions.
. Stay current with Databricks innovations and advocate for adoption of new features and capabilities.Required qualifications to be successful in this role:

Education Qualification: Bachelors degree in computer science or related field or higher with minimum 5 years of relevant experience.Must to have Skills-
6+ of experience in data architecture and engineering, with 5+ years in Databricks and Apache Spark. Strong proficiency in SQL & DAX
. Experience of migrating Snowflake and other custom EDW/ ETL solutions to Databricks.
. Experience of migrating different reporting solutions like Cognos, SAP BO etc to Power BI and Databricks.
. Deep knowledge and practical experience with Kimball dimensional modelling and data warehousing concepts and temporal data structures.
. Expertise in designing and deploying ETL/ELT pipelines for large-scale data integration.
. Proficiency in Power BI for paginated report and dashboard development, including DAX.
. Deep expertise in Delta Lake, structured streaming, PySpark, and SQL.
. Strong understanding of Lakehouse architecture, data mesh, and modern data stack principles.
. Experience with Unity Catalog, Databricks Repos, Jobs API, and Workflows.
. Proven ability to design and implement secure, governed, and highly available data platforms.
. Familiarity with cloud platforms (Azure, AWS, GCP) and their integration with Databricks.
. Experience with CI/CD, DevOps, and infrastructure-as-code tools (Terraform, GitHub Actions, Azure DevOps).
. Knowledge of machine learning lifecycle, MLflow, and model deployment strategies.
. An understanding of E-R data models (conceptual, logical, and physical).
. Understanding of advanced data warehouse concepts is required
. Strong analytical skills, including a thorough understanding of how to interpret customer business requirements and translate them into technical designs and solutions.
. Strong communication skills both verbal and written. Capable of collaborating effectively across a variety of IT and Business groups, across regions, roles and able to interact effectively with all levels.
. Strong problem-solving skills. Ability to identify where focus is needed and bring clarity to business objectives, requirements, and priorities.Must to Have :

  • Azure Databricks, Databricks Lakehouse Architecture
  • ETL ELT, Data Architecture
  • Apache Spark PySpark, Delta Lake, Delta Live Tables (DLT)
  • Unity Catalog, Medallion Architecture
  • Dimensional Modeling (Star & Snowflake), Kimball, Data Vault
  • Slowly Changing Dimensions (SCD Types 1, 2, 3)
  • Data Governance, RBAC, Data Lineage, Metadata Management
  • CI/CD & DevOps (Azure DevOps, GitHub Actions, Terraform)
  • SQL, Power BI, Self-Service Analytics, Semantic Model, Paginated Reports
  • Data Quality (Great Expectations), Performance Tuning, Cost Optimization
  • Cloud Platforms (Azure, AWS, GCP), Azure Data Factory, Synapse, Event Hubs
  • Nice-to-Have Skills:

  • Streaming frameworks (Kafka, Event Hubs), workspace automation
  • Advanced data modeling for Finance, Performance Budgeting, HRM systems
  • Subject-area models for financial reporting, workforce analytics, payroll insights
  • Delta Change Data Feed (CDF) and real-time data marts
  • Certifications: o Databricks Certified Data Engineer Associate Professionalo Databricks Certified Associate Developer for Apache Sparko Azure Power BI certificationsCGI is an equal opportunity employer. In addition, CGI is committed to providing accommodation for people with disabilities in accordance with provincial legislation. Please let us know if you require reasonable accommodation due to a disability during any aspect of the recruitment process and we will work with you to address your needs
  • Skills:

    • Azure DevOps
    • English

    Job Classification

    Industry: IT Services & Consulting
    Functional Area / Department: Data Science & Analytics
    Role Category: Business Intelligence & Analytics
    Role: BI Developer
    Employement Type: Full time

    Contact Details:

    Company: CGI
    Location(s): Bengaluru

    + View Contactajax loader


    Keyskills:   Azure Data Brick data warehousing advocate etl developer sql spark gcp devops apache spark etl azure architecture github data engineer power bi databricks machine learning budgeting r kafka terraform payroll aws infrastructure as code finance

     Fraud Alert to job seekers!

    ₹ Not Disclosed

    Similar positions

    Data Science Practitioner

    • Accenture
    • 1 - 3 years
    • Bengaluru
    • 1 day ago
    ₹ Not Disclosed

    Business Analyst

    • Sapwood Ventures
    • 2 - 6 years
    • Chennai
    • 1 day ago
    ₹ Not Disclosed

    Customer Success Research Analyst

    • Infobeans
    • 7 - 12 years
    • Indore
    • 1 day ago
    ₹ Not Disclosed

    Reporting Analyst

    • IntouchCX
    • 1 - 4 years
    • Hyderabad
    • 1 day ago
    ₹ Not Disclosed

    CGI

    Mphasis applies next-generation technology to help enterprises transform businesses globally. Customer centricity is foundational to Mphasis and is reflected in the Mphasis Front2Back™ Transformation approach. Front2Back™ uses the exponential power of cloud and cognitive to provide hyper-persona...