Hire PySpark Developer

PySpark
Upstaff is the best deep-vetting talent platform to match you with top PySpark developers for hire. Scale your engineering team with the push of a button
PySpark
Show Rates Hide Rates
Grid Layout Row Layout
AWS big data services 5yr.
Microsoft Azure 3yr.
Python
Kafka
ETL
C#
C++
Scala
Big Data Fundamentals via PySpark
Deep Learning in Python
Keras
Linear Classifiers in Python
Pandas
PySpark
TensorFlow
Theano
.NET
.NET Core
.NET Framework
Apache Airflow
Apache Hive
Apache Oozie 4
Apache Spark
Apache Spark 2
Data Analysis
Apache Hadoop
Apache Hive
Apache Spark
Apache Spark 2
AWS Database
dbt
HDP
Microsoft SQL Server
pgSQL
PostgreSQL
Snowflake
SQL
AWS ML (Amazon Machine learning services)
Keras
Machine Learning
OpenCV
TensorFlow
Theano
AWS
GCP (Google Cloud Platform)
AWS Database
AWS ML (Amazon Machine learning services)
AWS Quicksight
AWS Storage
GCP AI
GCP Big Data services
Apache Kafka 2
Apache Oozie 4
Kubernetes
OpenZeppelin
Qt Framework
YARN 3
SPLL
Superset
...

- Data Engineer with a Ph.D. degree in Measurement methods, Master of industrial automation - 16+ years experience with data-driven projects - Strong background in statistics, machine learning, AI, and predictive modeling of big data sets. - AWS Certified Data Analytics. AWS Certified Cloud Practitioner. Microsoft Azure services. - Experience in ETL operations and data curation - PostgreSQL, SQL, Microsoft SQL, MySQL, Snowflake - Big Data Fundamentals via PySpark, Google Cloud, AWS. - Python, Scala, C#, C++ - Skills and knowledge to design and build analytics reports, from data preparation to visualization in BI systems.

Show more
Seniority Expert (10+ years)
Location Ukraine
Azure 5yr.
Python 4yr.
SQL 5yr.
Cloudera 2yr.
PySpark
Apache Airflow
Apache Spark
AWS Athena
Databricks
Data modeling Kimbal
Microsoft Azure Synapse Analytics
Microsoft Power BI
Tableau
Apache Spark
AWS ElasticSearch
AWS Redshift
dbt
HDFS
Microsoft Azure SQL Server
NoSQL
Oracle Database
Snowflake
Spark SQL
SSAS
SSIS
SSRS
AWS
GCP (Google Cloud Platform)
AWS ElasticSearch
AWS EMR
AWS Glue
AWS Glue Studio
AWS Redshift
AWS S3
Azure HDInsight
Azure Key Vault
Databricks
Microsoft Azure SQL Server
Microsoft Azure Synapse Analytics
Grafana
Inmon
Kafka
...

- 12+ years experience working in the IT industry; - 12+ years experience in Data Engineering with Oracle Databases, Data Warehouse, Big Data, and Batch/Real time streaming systems; - Good skills working with Microsoft Azure, AWS, and GCP; - Deep abilities working with Big Data/Cloudera/Hadoop, Ecosystem/Data Warehouse, ETL, CI/CD; - Good experience working with Power BI, and Tableau; - 4+ years experience working with Python; - Strong skills with SQL, NoSQL, Spark SQL; - Good abilities working with Snowflake and DBT; - Strong abilities with Apache Kafka, Apache Spark/PySpark, and Apache Airflow; - Upper-Intermediate English.

Show more
Seniority Senior (5-10 years)
Location Norway
Python 9yr.
SQL 6yr.
Microsoft Power BI 5yr.
Reltio
Databricks
Tableau 5yr.
NoSQL 5yr.
REST 5yr.
GCP (Google Cloud Platform) 4yr.
Data Testing 3yr.
AWS 3yr.
Data Testing 3yr.
R 2yr.
Shiny 2yr.
Spotfire 1yr.
JavaScript
Dask
Django Channels
Pandas
PySpark
Python Pickle
PyTorch
Scrapy
TensorFlow
Apache Airflow
Apache Spark
Data Mining
Data Modelling
Data Scraping
ETL
Reltio Data Loader
Reltio Integration Hub (RIH)
Sisense
Apache Spark
Aurora
AWS DynamoDB
AWS ElasticSearch
Microsoft SQL Server
MySQL
PostgreSQL
RDBMS
SQLAlchemy
Machine Learning
PyTorch
Spacy
TensorFlow
AWS Bedrock
AWS CloudWatch
AWS DynamoDB
AWS ElasticSearch
AWS Fargate
AWS Lambda
AWS S3
AWS SQS
API
GraphQL
RESTful API
Selenium
Unit Testing
Git
Linux
Pipeline
RPA (Robotic Process Automation)
RStudio
BIGData
Cronjob
MDM
Mendix
Parallelization
Reltio APIs
Reltio match rules
Reltio survivorship rules
Reltio workflows
Vaex
...

- 8 years experience with various data disciplines: Data Engineer, Data Quality Engineer, Data Analyst, Data Management, ETL Engineer - Extensive hands-on expertise with Reltio MDM, including configuration, workflows, match rules, survivorship rules, troubleshooting, and integration using APIs and connectors (Databricks, Reltio Integration Hub), Data Modeling, Data Integration, Data Analyses, Data Validation, and Data Cleansing) - 8+ years with Python for data applications, including hands-on scripting experience - Data QA, SQL, Pipelines, ETL, Automated web scraping. - Data Analytics/Engineering with Cloud Service Providers (AWS, GCP) - Extensive experience with Spark and Hadoop, Databricks - 6 years of experience working with MySQL, SQL, and PostgreSQL; - 5 years of experience with Amazon Web Services (AWS), Google Cloud Platform (GCP) including Data Analytics/Engineering services, Kubernetes (K8s) - 5 years of experience with PowerBI - 4 years of experience with Tableau and other visualization tools like Spotfire and Sisense; - 3+ years of experience with AI/ML projects, background with TensorFlow, Scikit-learn and PyTorch; - Upper-intermediate to advanced English, - Henry is comfortable and has proven track record working with North American timezones (4hour+ overlap)

Show more
Seniority Senior (5-10 years)
Location Nigeria
SQL
ETL
Microsoft Power BI
DAX Studio
Git
Python
PySpark
Apache Airflow
Business Analysis
Data Analysis
Data Analysis Expressions (DAX)
Tableau
Spark SQL
Cloud Functions
Google Data Studio
Analytical Skills
DFF (Feature Driven Development)
UML
Docker
Excel
Usability tests
Data Structures
Mathematics
Unreal Engine
...

- 3+ years of experience as a BI Engineer; - Strong abilities in Power BI, SSIS, Tableau, and Google Data Studio; - Deep skills in developing and optimizing ETL processes within business intelligence; - Experience with SQL, Python; - Familiar with Docker, Apache Airflow, and PySpark; - Good knowledge of data warehousing and business intelligence principles.

Show more
Seniority Middle (3-5 years)
Location Czech Republic
Python
C++
Java
Matplotlib
NLTK
NumPy
Pandas
PySpark
PyTorch
Scikit-learn
SciPy
TensorFlow
Apache Spark
Databricks
Jupyter Notebook
MapReduce
Apache Hadoop
Apache Spark
Google BigQuery
Greenplum
MongoDB
MySQL
NoSQL
PostgreSQL
SQL
AWS SageMaker (Amazon SageMaker)
NumPy
OpenCV
PyTorch
Scikit-learn
TensorFlow
AWS
IBM Spectrum LSF
Slurm
AWS Batch
AWS Lambda
AWS S3
AWS SageMaker (Amazon SageMaker)
Databricks
Google BigQuery
Docker
Git
Linux
PyCharm
Shell Scripts
Multi-threading
YAML
...

- 2+ years of experience with Python as a Data Engineer and Deep/Machine Learning Intern - Experience with Data Vault modeling and AWS cloud services (S3, Lambda, and Batch) - Cloud Services: Sagemaker, Google BigQuery, Google Data Studio, MS Azure Databricks, IBM Spectrum LSF, Slurm - Data Science Frameworks: PyTorch, TensorFlow, PySpark, NumPy, SciPy, scikit-learn, Pandas, Matplotlib, NLTK, OpenCV - Proficient in SQL, Python, Linux, Git, and Bash scripting. - Had experience leading a BI development team and served as a Scrum Master. - Native English - Native German

Show more
Seniority Middle (3-5 years)
Location Hannover, Germany
Python 10yr.
SQL 10yr.
Talend ETL 10yr.
AWS EC2
Apache Airflow
PHP
Apex DataLoader
Salesforce
Jasperreports
Schedulers
Pandas
PySpark
AWS Athena
Jupyter Notebook
Presto S3
AWS ElasticSearch
AWS Redshift
Microsoft SQL Server
MySQL
Oracle Database
PostgreSQL
SQLAlchemy
Teradata
AWS Boto3
AWS Cloudformation
AWS ElasticSearch
AWS EMR
AWS Lambda
AWS Redshift
AWS S3
Google Data Studio
Datadog
JMeter
New Relic
Salesforce
C/C++/C#
Cloud IAM
MWAA
RDS
RMAN
Unix\Linux
...

- Senior Data Engineer with 10+ of experience specializing in designing, optimizing, and maintaining data infrastructures, data flow automation, and algorithm development. - Has expertise in Python, SQL/NoSQL, ETL processes, PySpark, Apache Airflow, and an array of AWS services, complemented by a strong foundation in database systems and cloud-based solutions. Proven capability in handling large-scale data analytics and processing with a focus on performance and cost efficiency in cloud environments. Proficient in developing robust ETL pipelines, performing data migrations, and optimizing complex queries and storage procedures, leveraging extensive experience across multiple industries and platforms. - Start: ASAP - English: Upper-Intermediate

Show more
Seniority Expert (10+ years)
Location Limassol, Spain
Python 8yr.
AWS
R 1yr.
C++
Beautiful Soup
Keras
NLTK
NumPy
Pandas
PySpark
PyTorch
TensorFlow
Apache Airflow
Apache Spark
AWS Athena
Microsoft Power BI
Apache Spark
AWS ElasticSearch
AWS Redshift
Clickhouse
SQL
AWS SageMaker (Amazon SageMaker)
Keras
Kubeflow
Mlflow
NumPy
OpenCV
PyTorch
Spacy
TensorFlow
AWS EC2
AWS ECR
AWS ElasticSearch
AWS EMR
AWS Redshift
AWS S3
AWS SageMaker (Amazon SageMaker)
AWS Timestream (Amazon Time Series Database)
Eclipse
Grafana
Kafka
MQQT
Kubernetes
OpenAPI
ArcGIS
Autogen
BERT
GPT
Guroby
ONNX
Open Street Map
Rasa NLU
...

- 10+ years experience working in the IT industry; - 8+ years experience working with Python; - Strong skills with SQL; - Good abilities working with R and C++; - Deep knowledge of AWS; - Experience working with Kubernetes (K8s), and Grafana; - Strong abilities with Apache Kafka, Apache Spark/PySpark, and Apache Airflow; - Experience working with Amazon S3, Athena, EMR, Redshift; - Specialised in Data Science and Data Analysis; - Work experience as a team leader; - Upper-Intermediate English.

Show more
Seniority Expert (10+ years)
Location Poland
Python 7yr.
SQL 7yr.
JMeter 7yr.
PySpark 7yr.
Apache Airflow 4yr.
NumPy 7yr.
Pandas 7yr.
NumPy 7yr.
Matplotlib 6yr.
Scikit-learn 3yr.
MongoDB 3yr.
PostgreSQL 3yr.
Redis 3yr.
Scikit-learn 3yr.
DB MS SQL 3yr.
GCP Cloudbuild 3yr.
Plotly 2yr.
PyTorch 2yr.
TensorFlow 2yr.
Neo4j 2yr.
Oracle Database 2yr.
PyTorch 2yr.
TensorFlow 2yr.
Azure 2yr.
Azure 2yr.
Kubernetes 2yr.
SciPy 1yr.
Huggingface 1yr.
Looker Studio
...

Data Engineer with 7 years of expertise in data analytics/science, ETL, and cloud technologies, blending deep healthcare and pharma industry knowledge. Proficient in Python, SQL, and a suite of data engineering tools including, Apache Spark, Airflow, and BI tools such as Power BI. Implemented real-time data streaming using Kafka, and has experience with multiple cloud services from AWS, Azure, and GCP. Key achievements include optimizing SQL database performance, automating data quality checks, and uncovering new drug candidates through computational data discovery—demonstrating a strong fusion of domain knowledge and technical acumen.

Show more
Seniority Senior (5-10 years)
Location Poland
Python 6yr.
SQL 6yr.
JMeter 6yr.
PySpark 6yr.
Apache Airflow 5yr.
Apex DataLoader
Spring Data
Apache Spark Streaming
Data Analysis
Data Quality
Data visualization
ELT
Apache Spark Streaming
Data Warehousing
AWS ML (Amazon Machine learning services)
AWS Kinesis
AWS ML (Amazon Machine learning services)
AWS Security Groups
Microsoft Azure API
CI/CD
Microsoft Azure API
3D Modelling
Cloud solutions
Code Review
Data backups
Illustration and icons
Lakehouse
Privacy Compliance & Data Governance
...

A seasoned Data Engineer with over 6 years of experience in the field of software and big data engineering. Holds a strong academic background in Computer Science and Software Engineering, certified as a Google Cloud Professional Data Engineer. Demonstrates deep expertise in high-load system design, performance optimizations, and domain-specific solutions for Healthcare, Fintech, and E-commerce. Proficient in Python and SQL, with significant exposure to data engineering tools such as Apache Hadoop, Apache Spark, and Apache Airflow, and cloud technologies from AWS and GCP. Adept at working with various databases and message brokers, excelling in data modeling, BI, and data visualization using tools like Looker, Power BI, and Tableau. Enhanced system efficiencies through SQL and data pipeline optimizations, driving significant improvements in processing speed and system performance. A collaborative engineer with a strong grasp of DevOps practices, committed to best-in-class data governance and security standards.

Show more
Seniority Senior (5-10 years)
Location Poland

Talk to Our Expert

Our journey starts with a 30-min discovery call to explore your project challenges, technical needs and team diversity.
Manager
Maria Lapko
Global Partnership Manager
Trusted by People
Trusted by Businesses
Accenture
SpiralScout
Valtech
Unisoft
Diceus
Ciklum
Infopulse
Adidas
Proxet
Accenture
SpiralScout
Valtech
Unisoft
Diceus
Ciklum
Infopulse
Adidas
Proxet

Want to hire PySpark developer? Then you should know!

Share this article

How and where is PySpark used?

  • Real-time Data Processing: Streaming Analytics
  • Machine Learning: Predictive Analytics
  • Data Warehousing: ETL Processes
  • Graph Processing: Social Network Analysis
  • Natural Language Processing: Sentiment Analysis
  • Image Processing: Object Recognition

Compare Junior, Middle, Senior, and Expert/Team Lead PySpark Developer roles

Seniority NameYears of experienceResponsibilities and activitiesAverage salary (USD/year)
Junior1-2 yearsResponsibilities & Activities:

  • Assist in data processing
  • Develop simple PySpark scripts
50,000
Middle3-5 yearsResponsibilities & Activities:

  • Optimize PySpark jobs
  • Debug complex issues
70,000
Senior6-8 yearsResponsibilities & Activities:

  • Design scalable PySpark solutions
  • Lead project implementations
90,000
Expert/Team Lead9+ yearsResponsibilities & Activities:

  • Architect PySpark frameworks
  • Mentor junior developers
120,000

Quick Facts about PySpark.

  • PySpark was unleashed in 2013, born from the fiery depths of Apache Spark.
  • From data processing to machine learning, PySpark is the darling of big data projects.
  • To dance with PySpark, one must wield the mighty Python and grasp the Spark framework.
  • Hadoop, with its distributed computing prowess, is the popular companion of PySpark.
  • Did you know? PySpark can make your data woes disappear faster than you can say “Big Data Magic!”

TOP PySpark Related Technologies

  1. Apache Spark
  2. (UC Berkeley, 2014)

  3. Hadoop
  4. (Apache Software Foundation, 2006)

  5. Python
  6. (Guido van Rossum, 1991)

  7. Scala
  8. (Martin Odersky, 2003)

What are top PySpark instruments and tools?

  • PyCharm: A powerful IDE by JetBrains, released in 2010
  • Databricks: Collaborative Apache Spark-based analytics service, released in 2013
  • Apache Zeppelin: Interactive data analytics environment, released in 2013
  • Jupyter Notebook: Open-source web application for interactive coding, released in 2015
  • Apache Spark: Unified analytics engine for big data processing, released in 2014
Table of Contents

Talk to Our Expert

Our journey starts with a 30-min discovery call to explore your project challenges, technical needs and team diversity.
Manager
Maria Lapko
Global Partnership Manager

Hire PySpark Developer as Effortless as Calling a Taxi

Hire PySpark Developer

FAQs on PySpark Development

What is a PySpark Developer? Arrow

A PySpark Developer is a specialist in the PySpark framework/language, focusing on developing applications or systems that require expertise in this particular technology.

Why should I hire a PySpark Developer through Upstaff.com? Arrow

Hiring through Upstaff.com gives you access to a curated pool of pre-screened PySpark Developers, ensuring you find the right talent quickly and efficiently.

How do I know if a PySpark Developer is right for my project? Arrow

If your project involves developing applications or systems that rely heavily on PySpark, then hiring a PySpark Developer would be essential.

How does the hiring process work on Upstaff.com? Arrow

Post Your Job: Provide details about your project.
Review Candidates: Access profiles of qualified PySpark Developers.
Interview: Evaluate candidates through interviews.
Hire: Choose the best fit for your project.

What is the cost of hiring a PySpark Developer? Arrow

The cost depends on factors like experience and project scope, but Upstaff.com offers competitive rates and flexible pricing options.

Can I hire PySpark Developers on a part-time or project-based basis? Arrow

Yes, Upstaff.com allows you to hire PySpark Developers on both a part-time and project-based basis, depending on your needs.

What are the qualifications of PySpark Developers on Upstaff.com? Arrow

All developers undergo a strict vetting process to ensure they meet our high standards of expertise and professionalism.

How do I manage a PySpark Developer once hired? Arrow

Upstaff.com offers tools and resources to help you manage your developer effectively, including communication platforms and project tracking tools.

What support does Upstaff.com offer during the hiring process? Arrow

Upstaff.com provides ongoing support, including help with onboarding, and expert advice to ensure you make the right hire.

Can I replace a PySpark Developer if they are not meeting expectations? Arrow

Yes, Upstaff.com allows you to replace a developer if they are not meeting your expectations, ensuring you get the right fit for your project.