Default profile banner
AA

Aman Agrawal

@amanagrawal

Data Engineer at Hashedin By Deloitte

Gorakhpur, India

linkedin.com/in/aman-agrawal-bbb8b2199

Hashedin By DeloitteAbes Engineering College

Aman Agrawal is a Data Engineer with 2 years of experience specializing in building robust data pipelines. He is proficient in Python, Pyspark, and SQL, with extensive experience utilizing AWS services for client infrastructure modernization. His expertise includes orchestrating ETL processes using AWS Glue, managing data migration from legacy systems like DB2, and handling real-time data streams via Kafka.

Experience

Data Engineer

Hashedin By Deloitte

•Invalid Date - Present•Gorakhpur, India

Received Applause and Spot Award for top impactor. Client Infrastructure Modernization to AWS. Provisioning all AWS resources using Cloudformation templates using Bitbucket and Bamboo data pipeline. Converting Cobol DB2 batch scripts business logics into data transformation rules in Pyspark running on AWS Glue ETL. Created one time data migration and validation scripts in Pyspark to migrate DB2 legacy tables to AWS Aurora Cluster. Orchestration of Glue Jobs using AWS Glue workflow, Triggers, AWS Step functions, Control-M etc. Configure FTP on EC2 instances. Load real-time streams into Aurora using AWS Kinesis and Kafka. Creation of JDBC connections and Glue network connector through cloudformation template that assists with accessing data stores in AWS Glue Studio and other AWS services. Develop and maintain data pipelines to extract data from a variety of sources(DB2, Kafka etc.) and populate target database. Implementing data quality checks and perform unit test of pyspark scripts using python libraries like boto3, moto, mock, awsglue utils etc running over a docker container in python venv and generating sonar coverage report.

Software Engineer Intern

Hashedin By Deloitte

•Invalid Date - Invalid Date

Create a dynamic shopping cart template with HTML/CSS/TypeScript/Angular. Create API's for shopping cart application with Spring Boot MVC and enter table records in postgresql database running on docker container. Used pyspark script to fetch real-time stock index data from kafka server/broker and create a data pipeline to ETL into postgresql database.

Education

Abes Engineering College

B. Tech

Electronics and Communications (ECE)

Invalid Date - Invalid Date

Licenses & Certifications

AWS Certified Cloud Practitioner

AWS

Issued: Invalid Date• No expiration

Python Programming and Data Science Certifications

NPTEL

• No expiration

Skills

Python
Pyspark
SQL
AWS
ETL
Bamboo
Pandas
Opencv
Django
Docker
Data Science
Tensorflow/Keras
Apache Airflow
Git
JavaScript
Angular
Data Visualization
Apache Hadoop
Data Warehousing
HTML/CSS
Splunk
Kafka