Demonstrated leadership in Data related Technical Programs. 17 years of Proficiency in data management, analytics and visualization, data flow, data integrity, data automation and data science.
Currently working at Amazon Web Services as a Senior Data Architect.
In my role at AWS, I was responsible for leading a team of engineers in the design, implementation, and maintenance of a large-scale data lake. I have a deep understanding of the full data engineering lifecycle, from data collection to data analysis and visualization. My core skills are with Python, Pyspark, Hadoop, Hive, cloud infrastructure, batch and streaming data. I am also proficient in a variety of data engineering tools and technologies, including AWS, Redshift, Snowflake, Postgres, MySQL, CloudFormation, CDK, Terraform, and CI/CD tools. Here are some of the highlights from my previous role:
Developed and popularized reusable solutions that reduced data lake build costs by 80%. This was achieved by creating a library of reusable code and templates that could be used to quickly and easily build data lakes. The library was made available to all engineers, which saved them time and effort in building their own data lakes.
Developed reusable assets and published guidance in aws-samples (Github) and AWS documents. This helped to ensure that other engineers could easily find and use the reusable assets that I had developed. The aws-samples repository is a great resource for finding reusable assets, and the AWS documentation is a great resource for learning about how to use AWS services.