Gopal G.
Data Engineer
Gopal is a Data Engineer with over eight years of experience in regulated sectors like automotive, technology, and energy. He excels in GCP, Azure, AWS, and Snowflake, with expertise in full life cycle development, data modeling, database architecture, and performance optimization.
His proudest achievements include creating and optimizing ETL/ELT pipelines across multi-cloud environments. Gopal's Google Cloud, AWS, Microsoft Azure, and Snowflake certifications highlight his commitment to continuous learning and professional excellence.
He holds a Master's degree in Computer Engineering.
Main expertise
- Fact Data Modeling 8 years
- ETL 8 years
- Unix shell 7 years
Other skills
- Pandas 4 years
- MySQL 4 years
- Apache ZooKeeper 4 years
Selected experience
Employment
Data Engineer
Nissan Motor Corporation - 1 year
-
Designing and implementing efficient and scalable data pipelines on Google Cloud Platform (GCP) to collect, process, and transform raw data into usable formats for analysis and consumption;
-
Leading and managing offshore teams to successfully implement various data engineering tasks, ensuring alignment with project goals and maintaining high-quality standards through regular communication, clear documentation, and effective task delegation;
-
Overseeing governance and compliance of data stored in Big Query, ensuring adherence to UK and EU GDPR regulations;
-
Conducting Data Privacy Impact Assessments (DPIA) for various projects at Nissan UK Limited and implementing necessary measures to mitigate or reduce risks;
-
Building and maintaining data warehouses, data lakes, and data lake houses on GCP using services like Big Query, Google Cloud Storage (GCS), and Bigtable;
-
Integrating data from various sources into GCP using services like Cloud Storage, Cloud Pub/Sub, and Cloud SQL;
-
Implementing proper data governance and security measures using GCP Identity and Access Management (IAM) and Data Loss Prevention (DLP) for compliance;
-
Building data pipelines using Google Dataflow to handle large volumes of data efficiently;
-
Implementing ETL/ELT processes to extract data from various sources and load them into data warehouses or data lakes;
-
Developing streaming pipelines for real-time data ingestion utilizing Kafka and Kafka Connect;
-
Implementing Python-based transformations and Big Query procedures, orchestrating their execution seamlessly with Google Cloud Composer;
-
Engineering transformations using Apache Beam, optimized for peak performance on Google DataProc clusters.
Technologies:
- Technologies:
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- AWS S3
- Data Analytics
- Looker
- Snowflake
- BigQuery
- Pandas
- MySQL
- Data Modeling
- Database testing
- Apache ZooKeeper
- AWS Athena
- Redshift
- Python
- SQL
- Apache Kafka
- Apache Airflow
- Apache Spark
- Hadoop
- Google Cloud
- Data Engineering
-
Lead Data Engineer
Technovert - 2 years 7 months
-
Developing ETL processes using Python and SQL to transform raw data into usable formats and load them into Big Query for analysis;
-
Building and architecting multiple data pipelines, managing end-to-end ETL and ELT processes for data ingestion and transformation in GCP, and coordinating tasks among the team;
-
Designing and implementing data pipelines using GCP services such as Dataflow, Dataproc, and Pub/Sub;
-
Migrating Oracle DSR to Big Query using Data Proc, Python, Airflow, and Looker;
-
Designing and developing a Python ingestion framework to load data from various source systems, including AR modules, inventory modules, files, and web services, into Big Query;
-
Developing pipelines to load data from customer-placed manual files in Google Drive to GCS and subsequently to Big Query using Big Query stored procedures;
-
Participating in code reviews and contributing to the development of best practices for data engineering on GCP;
-
Implementing data security and access controls using GCP's Identity and Access Management (IAM) and Cloud Security Command Centre.
Technologies:
- Technologies:
- Databricks
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- AWS S3
- Oracle
- Salesforce
- Data Analytics
- Microsoft Power BI
- Snowflake
- BigQuery
- Pandas
- MySQL
- Data Modeling
- Database testing
- Apache ZooKeeper
- Azure
- Azure Data Factory
- Azure Synapse
- Python
- SQL
- Apache Kafka
- Apache Airflow
- Apache Spark
- Hadoop
- Google Cloud
- Data Engineering
-
Data Engineer
Accenture - 1 year 8 months
-
Designing and implementing Snowflake data warehouses, developing schemas, tables, and views optimized for performance and data accessibility;
-
Extracting data from Oracle databases, transforming it into CSV files, and loading these files into a Snowflake data warehouse stage hosted on AWS S3, ensuring secure and efficient data transfer and storage;
-
Creating and utilizing virtual warehouses in Snowflake according to business requirements, effectively tracking credit usage to enhance business insights and resource allocation;
-
Designing and configuring Snow pipe pipelines for seamless and near-real-time data loading, reducing manual intervention, and enhancing data freshness;
-
Parsing XML data and organizing it into structured Snowflake tables for efficient data storage and seamless data analysis;
-
Designing and implementing JSON data ingestion pipelines, leveraging Snowflake's capabilities to handle nested and complex JSON structures;
-
Designing and deploying Amazon Redshift clusters, optimizing schema design, distribution keys, and sort keys for optimal query performance;
-
Leveraging AWS Lambda functions and Step Functions to orchestrate ETL workflows, ensuring data accuracy and timely processing;
-
Creating and maintaining data visualizations and reports using Amazon Quick Sight to facilitate data analysis and insights.
Technologies:
- Technologies:
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- Oracle
- Data Analytics
- Tableau
- Data Modeling
- Database testing
- Python
- SQL
- Data Engineering
-
BI Consultant, General Electric
Tech Mahindra - 2 years 7 months
-
Designing and implementing Teradata packages to facilitate seamless data extraction, transformation, and loading (ETL) operations from diverse sources into data warehouses;
-
Developing interactive and dynamic reports using SSRS, providing stakeholders with timely and insightful data visualizations for informed decision-making;
-
Conducting rigorous data validation and quality checks to ensure the integrity and accuracy of processed data;
-
Optimizing ETL performance by employing advanced techniques, resulting in a 25% reduction in processing time;
-
Developing the ingestion strategy for loading data from multiple source systems to the operational layer in the data warehouse using Python, SQL, and stored procedures;
-
Understanding and developing design documents as deliverables for the project;
-
Implementing SCD Type 1 and Type 2 functionality and developing custom scripts in Teradata for integration and functionality development for different modules like Primavera P6 and Oracle Project module;
-
Managing and troubleshooting issues as a DWH analyst to ensure the smooth flow of business operations;
-
Preparing unit test cases and performing end-to-end integration testing;
-
Actively participating in design discussions and reviewing solutions;
-
Involving in peer review discussions on development before moving to higher environments;
-
Loading data from multiple files to a single target table using ODI variables;
-
Configuring and developing ETL mappings to load data from XML and complex (unstructured/semi-structured) files;
-
Utilizing Power BI to design and develop insightful visualizations and interactive dashboards, enabling data-driven decision-making for stakeholders and enhancing overall data engineering solutions.
Technologies:
- Technologies:
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- Oracle
- Data Analytics
- Tableau
- Data Modeling
- SQL
- Data Engineering
-
Education
MSc.Computer Software Engineering
University of West London · 2022 - 2023
MSc.Electronics and Communications
Jawaharlal university of Hyderabad · 2012 - 2016
Find your next developer within days, not months
In a short 25-minute call, we would like to:
- Understand your development needs
- Explain our process to match you with qualified, vetted developers from our network
- You are presented the right candidates 2 days in average after we talk