Gopal G.
Data Engineer
Gopal er en dataingeniør med over åtte års erfaring i regulerte sektorer som bilindustri, teknologi og energi. Han er enestående innen GCP, Azure, AWS og Snowflake, med ekspertise i full livssyklusutvikling, datamodellering, databasearkitektur og ytelsesoptimalisering.
Hans stolteste prestasjoner inkluderer å lage og optimalisere ETL/ELT-pipeliner på tvers av multisky-miljøer. Gopals Google Cloud, AWS, Microsoft Azure og Snowflake sertifiseringer fremhever hans forpliktelse til kontinuerlig læring og profesjonell dyktighet.
Han har en mastergrad i datateknikk.
Hovedekspertise
- Fact Data Modeling 8 år
- ETL 8 år
- Unix shell 7 år
Andre kunnskaper
- Pandas 4 år
- MySQL 4 år
- Apache ZooKeeper 4 år
Utvalgt opplevelse
Arbeidserfaring
Data Engineer
Nissan Motor Corporation - 1 year
-
Designing and implementing efficient and scalable data pipelines on Google Cloud Platform (GCP) to collect, process, and transform raw data into usable formats for analysis and consumption;
-
Leading and managing offshore teams to successfully implement various data engineering tasks, ensuring alignment with project goals and maintaining high-quality standards through regular communication, clear documentation, and effective task delegation;
-
Overseeing governance and compliance of data stored in Big Query, ensuring adherence to UK and EU GDPR regulations;
-
Conducting Data Privacy Impact Assessments (DPIA) for various projects at Nissan UK Limited and implementing necessary measures to mitigate or reduce risks;
-
Building and maintaining data warehouses, data lakes, and data lake houses on GCP using services like Big Query, Google Cloud Storage (GCS), and Bigtable;
-
Integrating data from various sources into GCP using services like Cloud Storage, Cloud Pub/Sub, and Cloud SQL;
-
Implementing proper data governance and security measures using GCP Identity and Access Management (IAM) and Data Loss Prevention (DLP) for compliance;
-
Building data pipelines using Google Dataflow to handle large volumes of data efficiently;
-
Implementing ETL/ELT processes to extract data from various sources and load them into data warehouses or data lakes;
-
Developing streaming pipelines for real-time data ingestion utilizing Kafka and Kafka Connect;
-
Implementing Python-based transformations and Big Query procedures, orchestrating their execution seamlessly with Google Cloud Composer;
-
Engineering transformations using Apache Beam, optimized for peak performance on Google DataProc clusters.
Teknologier:
- Teknologier:
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- AWS S3
- Data Analytics
- Looker
- Snowflake
- BigQuery
- Pandas
- MySQL
- Data Modeling
- Database testing
- Apache ZooKeeper
- AWS Athena
- Redshift
- Python
- SQL
- Apache Kafka
- Apache Airflow
- Apache Spark
- Hadoop
- Google Cloud
- Data Engineering
-
Lead Data Engineer
Technovert - 2 years 7 months
-
Developing ETL processes using Python and SQL to transform raw data into usable formats and load them into Big Query for analysis;
-
Building and architecting multiple data pipelines, managing end-to-end ETL and ELT processes for data ingestion and transformation in GCP, and coordinating tasks among the team;
-
Designing and implementing data pipelines using GCP services such as Dataflow, Dataproc, and Pub/Sub;
-
Migrating Oracle DSR to Big Query using Data Proc, Python, Airflow, and Looker;
-
Designing and developing a Python ingestion framework to load data from various source systems, including AR modules, inventory modules, files, and web services, into Big Query;
-
Developing pipelines to load data from customer-placed manual files in Google Drive to GCS and subsequently to Big Query using Big Query stored procedures;
-
Participating in code reviews and contributing to the development of best practices for data engineering on GCP;
-
Implementing data security and access controls using GCP's Identity and Access Management (IAM) and Cloud Security Command Centre.
Teknologier:
- Teknologier:
- Databricks
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- AWS S3
- Oracle
- Salesforce
- Data Analytics
- Microsoft Power BI
- Snowflake
- BigQuery
- Pandas
- MySQL
- Data Modeling
- Database testing
- Apache ZooKeeper
- Azure
- Azure Data Factory
- Azure Synapse
- Python
- SQL
- Apache Kafka
- Apache Airflow
- Apache Spark
- Hadoop
- Google Cloud
- Data Engineering
-
Data Engineer
Accenture - 1 year 8 months
-
Designing and implementing Snowflake data warehouses, developing schemas, tables, and views optimized for performance and data accessibility;
-
Extracting data from Oracle databases, transforming it into CSV files, and loading these files into a Snowflake data warehouse stage hosted on AWS S3, ensuring secure and efficient data transfer and storage;
-
Creating and utilizing virtual warehouses in Snowflake according to business requirements, effectively tracking credit usage to enhance business insights and resource allocation;
-
Designing and configuring Snow pipe pipelines for seamless and near-real-time data loading, reducing manual intervention, and enhancing data freshness;
-
Parsing XML data and organizing it into structured Snowflake tables for efficient data storage and seamless data analysis;
-
Designing and implementing JSON data ingestion pipelines, leveraging Snowflake's capabilities to handle nested and complex JSON structures;
-
Designing and deploying Amazon Redshift clusters, optimizing schema design, distribution keys, and sort keys for optimal query performance;
-
Leveraging AWS Lambda functions and Step Functions to orchestrate ETL workflows, ensuring data accuracy and timely processing;
-
Creating and maintaining data visualizations and reports using Amazon Quick Sight to facilitate data analysis and insights.
Teknologier:
- Teknologier:
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- Oracle
- Data Analytics
- Tableau
- Data Modeling
- Database testing
- Python
- SQL
- Data Engineering
-
BI Consultant, General Electric
Tech Mahindra - 2 years 7 months
-
Designing and implementing Teradata packages to facilitate seamless data extraction, transformation, and loading (ETL) operations from diverse sources into data warehouses;
-
Developing interactive and dynamic reports using SSRS, providing stakeholders with timely and insightful data visualizations for informed decision-making;
-
Conducting rigorous data validation and quality checks to ensure the integrity and accuracy of processed data;
-
Optimizing ETL performance by employing advanced techniques, resulting in a 25% reduction in processing time;
-
Developing the ingestion strategy for loading data from multiple source systems to the operational layer in the data warehouse using Python, SQL, and stored procedures;
-
Understanding and developing design documents as deliverables for the project;
-
Implementing SCD Type 1 and Type 2 functionality and developing custom scripts in Teradata for integration and functionality development for different modules like Primavera P6 and Oracle Project module;
-
Managing and troubleshooting issues as a DWH analyst to ensure the smooth flow of business operations;
-
Preparing unit test cases and performing end-to-end integration testing;
-
Actively participating in design discussions and reviewing solutions;
-
Involving in peer review discussions on development before moving to higher environments;
-
Loading data from multiple files to a single target table using ODI variables;
-
Configuring and developing ETL mappings to load data from XML and complex (unstructured/semi-structured) files;
-
Utilizing Power BI to design and develop insightful visualizations and interactive dashboards, enabling data-driven decision-making for stakeholders and enhancing overall data engineering solutions.
Teknologier:
- Teknologier:
- Fact Data Modeling
- ETL
- Unix shell
- Performance Testing
- Unit Testing
- Oracle
- Data Analytics
- Tableau
- Data Modeling
- SQL
- Data Engineering
-
Utdannelse
MSc.Computer Software Engineering
University of West London · 2022 - 2023
MSc.Electronics and Communications
Jawaharlal university of Hyderabad · 2012 - 2016
Finn din neste utvikler innen dager, ikke måneder
I løpet av en kort 25-minutters samtale ønsker vi å:
- Forstå dine utviklingsbehov
- Forklare prosessen vår der vi matcher deg med kvalifiserte, evaluerte utviklere fra vårt nettverk
- Dele de neste stegene for å finne riktig match, ofte på mindre enn en uke