Upload

Henry A. Python Engineer

Data Analyst (DA), Data Extraction and ETL, Data Science, Business Intelligence (BI), Data Visualization

Summary

- 9+ years experience working with Python;
- 3+ years of experience with AI/ML projects;
- 5 years of experience with PowerBI and 4 years of experience with Tableau and other visualization tools like Spotfire;
- Good skills with Vaex, and Dask;
- Was a part of a financial debt management project;
- 6 years of experience working with MySQL, SQL, and PostgreSQL;
- 8 years of experience with various data sets (ETL, Data Engineer, Data Quality Engineer);
- 3 years of experience with Amazon Web Services (AWS), Google Cloud Platform (GCP);
- Proven commercial experience in HR and related Talent projects;
- Extensive experience with Spark and Hadoop;
- Background with TensorFlow, Scikit-learn and PyTorch;
- Data Analytics/Engineering with Cloud Service Providers (AWS, GCP);
- Deep abilities working with Kubernetes (K8s);
- Hands-on scripting experience with Python; Microsoft Power BI, Tableau, Sisense, CI/CD principles, Data Validation, Data QA, SQL, Pipelines, ETL, and Automated web scraping;
- Pet web3 projects (solidity, wallet integration);
- Upper-intermediate English.

Work Experience  

Python, AI/ML Engineer, Talents HR Platform

May 2023 - Current

Overview: 

  • Enhanced Customer Engagement: Leveraged AI technologies such as ChatGPT, Dall-E 3, Stable Diffusion, and AWS Bedrock to create personalized marketing campaigns, resulting in a 25% increase in customer engagement. Optimized backend performance for scalability and integrated robust security features (JWT, OAuth2, RBAC), improving security compliance by 30%.

  • Microservice for job resumes (profile) and Job description parser and scraper functionality that includes integration with LinkedIn, a popular workable, glassdoor-like platform, Google Docs, PDF & Word parsers. Used R, Shiny (RStudio) with Python for predictive analytics.
  • AI-Driven Face Matching Model: Developed a model using Deep Face and Face Recognition Dlib to recognize lost children, achieving an 85% match rate. Applied predictive algorithms like KNN, SVM, Linear Regression, and G-Boost to various datasets, handling large-scale data from Kaggle for diverse applications.

  • Efficient Data Management: Managed extensive data on the front end with PostgreSQL and Databricks. Wrote efficient Python code using FastAPI and Asyncio, ensuring seamless integration of AI models, and reduced bug rates by 15% through regular code reviews and best engineering practices.

  • System Reliability and Scalability: Improved system architecture and coding practices, leading to a 20% increase in system reliability. Developed scalable AI solutions with AWS Lambda and GCP, enabling the digital marketing platform to handle varying workloads efficiently.

  • Real-Time Personalized Experiences: Designed and developed backend infrastructure for AI-driven digital marketing, empowering marketers to create highly targeted and dynamic campaigns, significantly enhancing customer engagement and the overall performance of the digital marketing platform.

  • Use Case: Develop and train deep learning models with PyTorch to forecast future sales trends and customer demand.

Responsibilities

Technologies: AWS, Restful API, Python, Pytest, Allure, R, Shiny, JavaScript, Docker, Kubernetes, ChatGPT, Dall-E 3, Stable Diffusion, AWS Bedrock, Face Recognition Dlib, KNN (K-Nearest Neighbors), SVM (Support Vector Machine), Linear Regression, G-Boost (Gradient Boosting), JWT (JSON Web Token), OAuth2 Authentication, PyTorch, RBAC (Role-Based Access Control), Scikit-learn, FastAPI, Asyncio, PostgreSQL, Databricks, LinkedIn API, Glassdoor-like platform integration, Google Docs API, PDF parser.

Power BI/AI Engineer

January 2023 - May 2023

Overview: Startup that revolutionizes the home equity market in the US. Our team is working on providing outstanding BI services with accessible data to decision-makers as well as streamlining the current services and their effectiveness.

Responsibilities

  • Design and develop Tableau dashboards;
  • Utilized Spotfire to design and implement interactive dashboards that provided real-time insights into key business metrics;
  • Produce well-designed, efficient code by using the best software development practices;
  • Perform code reviews for compliance with the best engineering practices, coding standards, and quality criteria set for the projects;
  • Use TensorFlow to automate and optimize code reviews, ensuring compliance with best engineering practices through AI-driven code quality assessments;
  • Provide suggestions to improve the architecture, and coding practices, build/verification toolset, and solve customer problems.

Technologies: Tableau, SQL, Snowflake, TensorFlow.

Data/AI Engineer, Data management platform / Amazon E-Commerce Aggregator

2020-August 2022

Overview: A next-generation consumer goods company reimagining how the world’s most-loved products become accessible to everyone. We use a deep understanding of rankings, ratings, and reviews to identify and acquire quality brands and use world-class expertise and data science to make their products better or create new ones to meet changing customer demand.

Responsibilities:

  • Use Sisense to build dashboards for tracking updates to selected amazon store brands for determined time periods. I used the interactive SQL palette for querying the tables to filter the needed information (columns) to be displayed in the dashboard. This dashboard provides the data engineering manager with the necessary information to make decisions on store brands.
  • Create and support ELT data pipelines built on Snowflake and DBT while ensuring high-quality data
  • Develop and deploy data warehousing models, and support existing processes/ETLs (extract/transform/load), and functions (in Python/SQL/DBT) in a cloud data warehouse environment using Snowflake, AWS services 
  • SQL statements and developing in Python
  • Design and develop data pipelines (DAGs). Automation tests. 

Technologies: Sisense, AirFlow, Snowflake, NoSQL, Python3, AWS S3, Hadoop, Spark, GitLab CI/CD, Kubernetes, LDAP.

AI Engineer, AI Project

September 2020 – July 2021
Responsibilities:

  • Develop automation workflows with RPA (UiPath).
  • Set up and manage web-based cloud services on AWS EC2.
  • Utilize TensorFlow to build and deploy models that predict key business metric

Python Instructor, NDA

March 2020 – March 2021
Responsibilities:

  • Teach Python programming to students. 
  • Develop a curriculum to be used for teaching python programming and data analysis. 

Python Developer, IoT-leveraged agricultural tech company 

A project on monitoring and reporting sample data from agricultural plants on a field of land.

May 2020 – August 2020
Responsibilities:

  • Hands-on setting up, maintaining, and deploying services to AWS EC2. 
  • Automated web scraping of data from webpages using Selenium. 
  • Carried out multi-processing and parallelizing of code with PySpark.
  • Used Spark for 2 cases of data processing in an ELT phase:

1. Data was collected from drones and other specialized bots were used to physically survey the land area and take samples from the soil and air for properties such as soil pH, moisture content, specific gravity, etc for different types of crops planted on the field. This data was received gotten in real-time, and placed on a queue to be loaded into AWS DynamoDB.   The transformation involves converting some data properties from the queue such as temperature from degree celsius to the kelvin scale, moisture content from cubic centimeters to cubic meters, etc. The transformed data is then loaded into AWS s3.

2. Process large batch data averaging 10 million rows with spark: There were cases where I had to transform data on a different database containing historical data to consolidate the currently maintained tables in another database. The historical data contains millions of rows of IoT-generated values. To optimize speed and memory usage for transformation, I used python’s implementation of Spark (Pyspark) to carry out the same transformation technique on the batch data to backfill the current table in the database. 

 IT Analyst, FieldworkAfrica 

July 2019 – 2021
Responsibilities:

  • Developed data visualizations on PowerBI and Tableau to track areas of high and low drink consumption to establish which areas are potentially viable to push a new drink to.
  • Provided daily and historical data reports and visualizations to the technical director. Daily and historical reports included tracking the coverage of data collection in geographical areas, and providing updates on data quality checks and target data samples.
  • Developed and maintained cloud services on the Google cloud platform. 
  • Developed questionnaire scripts on ODK for market research. 
  • Led a data collection team of 10 people. 
  • Performed data analysis using data tools, visualizations, and dashboards.
  • Used PowerBI and Tableau to provide daily and historical data reports and

    visualizations to the technical director. Daily and historical reports included tracking the

    coverage of data collection in geographical areas, providing updates on data quality checks and

    target data samples.

Python Developer,  NDA

Jan 2019 – April 2019
Responsibilities:

  • Working on websites back-end with flask and Django. 
  • Maintaining SQL databases for proper scaling. 
  • Ensuring proper test units are integrated to promote clean codes. 

Data Science Trainee, DATA SCIENCE

2017
Responsibilities:

  • Implemented optimization algorithms. 
  • Carried out analytics with Microsoft Azure for prediction models. 
  • Generated various visualization models for data analytics with Power Bi and Seaborn. 

Campus Ambassador, NDA 

July 2016 - December 2016
Responsibilities:

  • Promoted the ScholarX mobile App on designated campuses and social platforms for the company achieving 1000 downloads on the Google Play Store.

Engineering Intern, NDA

April 2015 - July 2015
Responsibilities:

  • Assisted in a supervisory management role and design engineering in various structural steel processes.

Tableau Experience Highlights:

1. Real Estate Project

  • Developed and maintained predictive algorithms for US house prices using machine learning techniques such as regression and classification
  • Created interactive data visualizations for real estate agents and investors using Tableau
  • Analyzed a variety of data points on comparables for single-family homes and condos, including location, property age, and amenities
  • Assessed factors like ARV (After Repair Value), square footage, year built, number of beds and baths, garages, and local market conditions
  • Developed user-friendly dashboards to display real-time market trends and property values, enabling investors to make informed decisions quickly
  • Collaborated with a team of data scientists and engineers to continuously improve algorithms and visualizations

2. Tableau Specialist in Market Research Project:

  • Utilized Tableau for comprehensive daily and historical data reporting and visualization to support decision-making processes
  • Provided data insights and visualizations to the technical director, enabling a better understanding of market dynamics and trends
  • Created a range of custom dashboards for daily and historical reports that covered various aspects such as sales, customer demographics, and product performance
  • Monitored and analyzed data collection coverage in target geographical areas to ensure accurate representation of the market
  • Conducted regular data quality checks, including data validation and cleaning, to maintain high data accuracy and reliability
  • Collaborated with data engineers and analysts to optimize data collection methods and improve overall data quality
 

Education

  • College of Technology
    2018 - 2019
    Higher National Diploma (HND), Mechanical Engineering 
  • College of Technology
    2013 - 2016
    National Diploma (ND), Mechanical Engineering 

Certifications 

  • Python Developer Certificate (Sensegrass)
    2020
  • Google Scholarship Android
    Basics - 2018
  • Certificate of Completion (DSN 2nd Data Science Boot Camp)
    2017
  • Certificate of Proficiency in Human Resources and Skill Acquisition
    2014
  • Certificate of Participation ACM (Association for Computing Machinery)
    2017
  • Big Data Foundations (Level 1)
    2017
  • Data Science Foundations (Level 1)
    2016