Hire PySpark Developer

PySpark
Upstaff is the best deep-vetting talent platform to match you with top PySpark developers for hire. Scale your engineering team with the push of a button
PySpark

Meet Our Devs

Show Rates Hide Rates
Grid Layout Row Layout
AWS big data services 5yr.
Microsoft Azure 3yr.
Python
Kafka
ETL
AWS ML (Amazon Machine learning services)
Keras
Machine Learning
OpenCV
TensorFlow
Theano
C#
C++
Scala
Apache Spark
Big Data Fundamentals via PySpark
Deep Learning in Python
Linear Classifiers in Python
Pandas
PySpark
.NET
.NET Core
.NET Framework
Apache Airflow
Apache Hive
Apache Oozie 4
Apache Spark 2
Data Analysis
Apache Hadoop
AWS Database
dbt
HDP
Microsoft SQL Server
pgSQL
PostgreSQL
Snowflake
SQL
AWS
GCP
AWS Quicksight
AWS Storage
GCP AI
GCP Big Data services
Apache Kafka 2
Kubernetes
OpenZeppelin
Qt Framework
YARN 3
SPLL
Superset
...

- Data Engineer with a Ph.D. degree in Measurement methods, Master of industrial automation - 16+ years experience with data-driven projects - Strong background in statistics, machine learning, AI, and predictive modeling of big data sets. - AWS Certified Data Analytics. AWS Certified Cloud Practitioner. Microsoft Azure services. - Experience in ETL operations and data curation - PostgreSQL, SQL, Microsoft SQL, MySQL, Snowflake - Big Data Fundamentals via PySpark, Google Cloud, AWS. - Python, Scala, C#, C++ - Skills and knowledge to design and build analytics reports, from data preparation to visualization in BI systems.

Show more
Seniority Expert (10+ years)
Location Ukraine
Azure 5yr.
Python 4yr.
SQL 5yr.
Cloudera 2yr.
Apache Spark
JSON
PySpark
XML
Apache Airflow
AWS Athena
Databricks
Data modeling Kimbal
Microsoft Azure Synapse Analytics
Power BI
Tableau
AWS ElasticSearch
AWS Redshift
dbt
HDFS
Microsoft Azure SQL Server
NoSQL
Oracle Database
Snowflake
Spark SQL
SSAS
SSIS
SSRS
AWS
GCP
AWS EMR
AWS Glue
AWS Glue Studio
AWS S3
Azure HDInsight
Azure Key Vault
API
Grafana
Inmon
REST
Kafka
databases
...

- 12+ years experience working in the IT industry; - 12+ years experience in Data Engineering with Oracle Databases, Data Warehouse, Big Data, and Batch/Real time streaming systems; - Good skills working with Microsoft Azure, AWS, and GCP; - Deep abilities working with Big Data/Cloudera/Hadoop, Ecosystem/Data Warehouse, ETL, CI/CD; - Good experience working with Power BI, and Tableau; - 4+ years experience working with Python; - Strong skills with SQL, NoSQL, Spark SQL; - Good abilities working with Snowflake and DBT; - Strong abilities with Apache Kafka, Apache Spark/PySpark, and Apache Airflow; - Upper-Intermediate English.

Show more
Seniority Senior (5-10 years)
Location Norway
Python 9yr.
SQL 6yr.
Power BI 5yr.
Reltio
Databricks
Tableau 5yr.
NoSQL 5yr.
REST 5yr.
GCP 4yr.
Data Testing 3yr.
AWS 3yr.
R 2yr.
Shiny 2yr.
Spotfire 1yr.
JavaScript
Machine Learning
PyTorch
Spacy
TensorFlow
Apache Spark
Dask
Django Channels
Pandas
PySpark
Python Pickle
Scrapy
Apache Airflow
Data Mining
Data Modelling
Data Scraping
ETL
Reltio Data Loader
Reltio Integration Hub (RIH)
Sisense
Aurora
AWS DynamoDB
AWS ElasticSearch
Microsoft SQL Server
MySQL
PostgreSQL
RDBMS
SQLAlchemy
AWS Bedrock
AWS CloudWatch
AWS Fargate
AWS Lambda
AWS S3
AWS SQS
API
GraphQL
RESTful API
Selenium
Unit Testing
Git
Linux
Pipeline
RPA (Robotic Process Automation)
RStudio
BIGData
Cronjob
MDM
Mendix
Parallelization
Reltio APIs
Reltio match rules
Reltio survivorship rules
Reltio workflows
Vaex
...

- 8 years experience with various data disciplines: Data Engineer, Data Quality Engineer, Data Analyst, Data Management, ETL Engineer - Extensive hands-on expertise with Reltio MDM, including configuration, workflows, match rules, survivorship rules, troubleshooting, and integration using APIs and connectors (Databricks, Reltio Integration Hub), Data Modeling, Data Integration, Data Analyses, Data Validation, and Data Cleansing) - Data QA, SQL, Pipelines, ETL, Automated web scraping. - Data Analytics/Engineering with Cloud Service Providers (AWS, GCP) - Extensive experience with Spark and Hadoop, Databricks - 6 years of experience working with MySQL, SQL, and PostgreSQL; - 5 years of experience with Amazon Web Services (AWS), Google Cloud Platform (GCP) including Data Analytics/Engineering services, Kubernetes (K8s) - 5 years of experience with PowerBI - 4 years of experience with Tableau and other visualization tools like Spotfire and Sisense; - 3+ years of experience with AI/ML projects, background with TensorFlow, Scikit-learn and PyTorch; - Upper-intermediate to advanced English, - Henry is comfortable and has proven track record working with North American timezones (4hour+ overlap)

Show more
Seniority Senior (5-10 years)
Location Nigeria
SQL
ETL
Power BI
DAX Studio
Git
Python
PySpark
Apache Airflow
Business Analysis
Data Analysis
Data Analysis Expressions (DAX)
Tableau
Spark SQL
Cloud Functions
Google Data Studio
Docker
Excel
FDD
UML
Usability tests
Data Structures
Mathematics
Unreal Engine
...

- 3+ years of experience as a BI Engineer; - Strong abilities in Power BI, SSIS, Tableau, and Google Data Studio; - Deep skills in developing and optimizing ETL processes within business intelligence; - Experience with SQL, Python; - Familiar with Docker, Apache Airflow, and PySpark; - Good knowledge of data warehousing and business intelligence principles.

Show more
Seniority Middle (3-5 years)
Location Czech Republic
Python
Data Science
SQL
AWS ML (Amazon Machine learning services)
PySpark
Presto S3
PostgreSQL
AWS EC2
AWS Lambda
GCE
Docker
Kafka
NFT
Django REST framework
Flink
General network knowledge
NLP
RDS
...

Expertise in Python, SQL, and ML web applications with a focus on practical applications in fields like AI and computer vision. Proven track record in AI with projects at DeepLogic AI, including creating an ML Web app for classifying blastocyst embryos and implementing LLM for in-house document processing. Further experience includes feature extraction via PySpark at Tata Digital, and exploiting BERT for NLP tasks at Concentrix. Academically grounded with a Master's in CS from Columbia University and a BTech in Math/CS from Delhi University. Holds AWS accreditations, demonstrating proficiency in cloud architectures, EC2, S3, and RDS. Skilled in end-to-end software development, including full-stack abilities shown at Datacultr with Django and PostgreSQL.

Show more
Seniority Intern
Location New York, United States
Python
AWS SageMaker (Amazon SageMaker)
NumPy
OpenCV
PyTorch
Scikit-learn
TensorFlow
C++
Java
Apache Spark
Matplotlib
NLTK
Pandas
PySpark
SciPy
Databricks
Jupyter Notebook
MapReduce
Apache Hadoop
Google BigQuery
Greenplum
MongoDB
MySQL
NoSQL
PostgreSQL
SQL
AWS
IBM Spectrum LSF
Slurm
AWS Batch
AWS Lambda
AWS S3
Docker
Git
Linux
PyCharm
Shell Scripts
Multi-threading
YAML
...

- 2+ years of experience with Python as a Data Engineer and Deep/Machine Learning Intern - Experience with Data Vault modeling and AWS cloud services (S3, Lambda, and Batch) - Cloud Services: Sagemaker, Google BigQuery, Google Data Studio, MS Azure Databricks, IBM Spectrum LSF, Slurm - Data Science Frameworks: PyTorch, TensorFlow, PySpark, NumPy, SciPy, scikit-learn, Pandas, Matplotlib, NLTK, OpenCV - Proficient in SQL, Python, Linux, Git, and Bash scripting. - Had experience leading a BI development team and served as a Scrum Master. - Native English - Native German

Show more
Seniority Middle (3-5 years)
Location Hannover, Germany
Python 10yr.
SQL 10yr.
Talend ETL 10yr.
AWS EC2
Apache Airflow
PHP
Apex DataLoader
Salesforce
C/C++/C#
Jasperreports
Schedulers
Pandas
PySpark
AWS Athena
Jupyter Notebook
Presto S3
AWS ElasticSearch
AWS Redshift
Microsoft SQL Server
MySQL
Oracle Database
PostgreSQL
SQLAlchemy
Teradata
AWS Boto3
AWS Cloudformation
AWS EMR
AWS Lambda
AWS S3
Google Data Studio
Datadog
JMeter
New Relic
Cloud IAM
MWAA
RDS
RMAN
Unix\Linux
...

- Senior Data Engineer with 10+ of experience specializing in designing, optimizing, and maintaining data infrastructures, data flow automation, and algorithm development. - Has expertise in Python, SQL/NoSQL, ETL processes, PySpark, Apache Airflow, and an array of AWS services, complemented by a strong foundation in database systems and cloud-based solutions. Proven capability in handling large-scale data analytics and processing with a focus on performance and cost efficiency in cloud environments. Proficient in developing robust ETL pipelines, performing data migrations, and optimizing complex queries and storage procedures, leveraging extensive experience across multiple industries and platforms. - Start: ASAP - English: Upper-Intermediate

Show more
Seniority Expert (10+ years)
Location Limassol, Spain
Python 8yr.
AWS
R 1yr.
AWS SageMaker (Amazon SageMaker)
BERT
Keras
Kubeflow
Mlflow
NumPy
OpenCV
PyTorch
Spacy
TensorFlow
C++
Apache Spark
Beautiful Soup
NLTK
Pandas
PySpark
Apache Airflow
AWS Athena
Power BI
AWS ElasticSearch
AWS Redshift
Clickhouse
SQL
AWS EC2
AWS ECR
AWS EMR
AWS S3
AWS Timestream (Amazon Time Series Database)
Eclipse
Grafana
Kafka
MQQT
Kubernetes
OpenAPI
ArcGIS
GPT
Guroby
ONNX
Open Street Map
Rasa NLU
...

- 10+ years experience working in the IT industry; - 8+ years experience working with Python; - Strong skills with SQL; - Good abilities working with R and C++; - Deep knowledge of AWS; - Experience working with Kubernetes (K8s), and Grafana; - Strong abilities with Apache Kafka, Apache Spark/PySpark, and Apache Airflow; - Experience working with Amazon S3, Athena, EMR, Redshift; - Specialised in Data Science and Data Analysis; - Work experience as a team leader; - Upper-Intermediate English.

Show more
Seniority Expert (10+ years)
Location Poland

Let’s set up a call to address your requirements and set up an account.

Talk to Our Expert

Our journey starts with a 30-min discovery call to explore your project challenges, technical needs and team diversity.
Manager
Maria Lapko
Global Partnership Manager
Trusted by People
Trusted by Businesses
Accenture
SpiralScout
Valtech
Unisoft
Diceus
Ciklum
Infopulse
Adidas
Proxet
Accenture
SpiralScout
Valtech
Unisoft
Diceus
Ciklum
Infopulse
Adidas
Proxet

Want to hire PySpark developer? Then you should know!

Share this article

How and where is PySpark used?

  • Real-time Data Processing: Streaming Analytics
  • Machine Learning: Predictive Analytics
  • Data Warehousing: ETL Processes
  • Graph Processing: Social Network Analysis
  • Natural Language Processing: Sentiment Analysis
  • Image Processing: Object Recognition

Compare Junior, Middle, Senior, and Expert/Team Lead PySpark Developer roles

Seniority NameYears of experienceResponsibilities and activitiesAverage salary (USD/year)
Junior1-2 yearsResponsibilities & Activities:

  • Assist in data processing
  • Develop simple PySpark scripts
50,000
Middle3-5 yearsResponsibilities & Activities:

  • Optimize PySpark jobs
  • Debug complex issues
70,000
Senior6-8 yearsResponsibilities & Activities:

  • Design scalable PySpark solutions
  • Lead project implementations
90,000
Expert/Team Lead9+ yearsResponsibilities & Activities:

  • Architect PySpark frameworks
  • Mentor junior developers
120,000

Quick Facts about PySpark.

  • PySpark was unleashed in 2013, born from the fiery depths of Apache Spark.
  • From data processing to machine learning, PySpark is the darling of big data projects.
  • To dance with PySpark, one must wield the mighty Python and grasp the Spark framework.
  • Hadoop, with its distributed computing prowess, is the popular companion of PySpark.
  • Did you know? PySpark can make your data woes disappear faster than you can say “Big Data Magic!”

TOP PySpark Related Technologies

  1. Apache Spark
  2. (UC Berkeley, 2014)

  3. Hadoop
  4. (Apache Software Foundation, 2006)

  5. Python
  6. (Guido van Rossum, 1991)

  7. Scala
  8. (Martin Odersky, 2003)

What are top PySpark instruments and tools?

  • PyCharm: A powerful IDE by JetBrains, released in 2010
  • Databricks: Collaborative Apache Spark-based analytics service, released in 2013
  • Apache Zeppelin: Interactive data analytics environment, released in 2013
  • Jupyter Notebook: Open-source web application for interactive coding, released in 2015
  • Apache Spark: Unified analytics engine for big data processing, released in 2014
Table of Contents

Talk to Our Expert

Our journey starts with a 30-min discovery call to explore your project challenges, technical needs and team diversity.
Manager
Maria Lapko
Global Partnership Manager

Hire PySpark Developer as Effortless as Calling a Taxi

Hire PySpark Developer

FAQs on PySpark Development

What is a PySpark Developer? Arrow

A PySpark Developer is a specialist in the PySpark framework/language, focusing on developing applications or systems that require expertise in this particular technology.

Why should I hire a PySpark Developer through Upstaff.com? Arrow

Hiring through Upstaff.com gives you access to a curated pool of pre-screened PySpark Developers, ensuring you find the right talent quickly and efficiently.

How do I know if a PySpark Developer is right for my project? Arrow

If your project involves developing applications or systems that rely heavily on PySpark, then hiring a PySpark Developer would be essential.

How does the hiring process work on Upstaff.com? Arrow

Post Your Job: Provide details about your project.
Review Candidates: Access profiles of qualified PySpark Developers.
Interview: Evaluate candidates through interviews.
Hire: Choose the best fit for your project.

What is the cost of hiring a PySpark Developer? Arrow

The cost depends on factors like experience and project scope, but Upstaff.com offers competitive rates and flexible pricing options.

Can I hire PySpark Developers on a part-time or project-based basis? Arrow

Yes, Upstaff.com allows you to hire PySpark Developers on both a part-time and project-based basis, depending on your needs.

What are the qualifications of PySpark Developers on Upstaff.com? Arrow

All developers undergo a strict vetting process to ensure they meet our high standards of expertise and professionalism.

How do I manage a PySpark Developer once hired? Arrow

Upstaff.com offers tools and resources to help you manage your developer effectively, including communication platforms and project tracking tools.

What support does Upstaff.com offer during the hiring process? Arrow

Upstaff.com provides ongoing support, including help with onboarding, and expert advice to ensure you make the right hire.

Can I replace a PySpark Developer if they are not meeting expectations? Arrow

Yes, Upstaff.com allows you to replace a developer if they are not meeting your expectations, ensuring you get the right fit for your project.