I am Ankit Pushpam

my-porfolio-headshot

Name: Ankit Pushpam

Profile: Data Engineer

Email: pushpam.ankit@gmail.com

Phone: (650)-229-3946

Skill

AWS Cloud 90%
Python 85%
Pyspark 75%
Advanced SQL 90%
About me

A motivated data professional around 3 years as Data Engineer and 5 years as a Data Analyst, skilled in scalable data solutions, data modeling, and cloud technologies. Excels in cross-functional collaboration and problem-solving with a passion for leveraging data to solve complex business challenges.

Currently at Outcomes, I design and implement data solutions for data-driven decision-making. I build a robust ETL pipeline using AWS Glue, Python, PySpark, and Advanced SQL, ensuring data availability for analysis.

I was a Data Engineer at Capital One via Iquest Solutions, responsible for building robust data pipelines. I extracted customer account data from the data lake, transformed it as needed, and loaded it onto Snowflake data warehouse using technology like AWS, Python, Pyspark. As a Masters student at California State University, I worked on Big Data, NLP, and Time-series projects using Hadoop, Spark, and AWS. Previously, I worked at Micro Focus as a Business Planning Analyst.

Portfolio

Explore my data engineering expertise in ETL pipelines, leveraging Airflow, AWS, Docker, Kafka, Python, and Spark, feeding data into AWS Redshift and Snowflake.

paxos algorithm nodes

Twitter Airflow Data Pipeline

Data Pipeline, Airflow
virtual memory manager in C image

Kafka, Spark, and Airflow Data Generation, Streaming, and Processing Project

ETL pipeline
Augmented reality space portal

Bay Area Housing Analysis

Web Scraping

Skills

Explore the skills section for a breakdown of my technical proficiencies. This segment highlights the concrete abilities and knowledge I've cultivated throughout my journey, shaping my capacity to contribute effectively in diverse projects.

Language


Python
Pyspark
Java
Advanced SQL

Cloud, Big Data and Streaming

AWS Certified
Spark
Hadoop
Kafka
Kensis

Databases and Data Warehouse

PostgreSQL
MS SQL Server
AWS Redshift
Snowflake

Blogs

Welcome to my Data Engineering blog, where I share expertise in Python, Spark, Airflow, Kafka, Docker, Terraform, and AWS. Explore insights, tutorials, and discussions on building robust data pipelines.

AWS Cloud

Setting Up LocalStack and DynamoDB Tables with Docker and Terraform

This guide demonstrates setting up LocalStack and provisioning DynamoDB tables via Terraform for local development. Additionally, it explores visualizing and interacting with these tables using NoSQL Workbench, enhancing DynamoDB-dependent app development.

ETL

Mastering Serverless ETL: A Step-by-Step Guide with AWS Glue

this guide demonstrates the creation of a serverless ETL pipeline with AWS Glue, providing a foundational understanding of the process.

Python and SQL

Generating Mock Data for Complex Relational Tables

This article offers a practical solution to generate mock data for complex relational databases using Python, SQLite, SQLAlchemy, and Faker libraries.

Resume

Browse through my professional journey, accomplishments, and skills in the resume section. Here, you'll find a comprehensive overview of my experiences, education, and expertise that define my career path.