HimalayasHimalayas logo
IC

PySpark Developer

Inizio Partners Corp
United States only

Stay safe on Himalayas

Never send money to companies. Jobs on Himalayas will never require payment from applicants.

Job Summary:

We are seeking a highly skilled and experienced Python and PySpark Developer to join our team. The ideal candidate will be responsible for designing, developing, and optimizing big data pipelines and solutions using Python, PySpark, and distributed computing frameworks. This role involves working closely with data engineers, data scientists, and business stakeholders to process, analyze, and derive insights from large-scale datasets.

Key Responsibilities:

Data Engineering & Development:

  • Design and implement scalable data pipelines using PySpark and other big data frameworks.
  • Develop reusable and efficient code for data extraction, transformation, and loading (ETL).
  • Optimize data workflows for performance and cost efficiency.

Data Analysis & Processing:

  • Process and analyze structured and unstructured datasets.
  • Build and maintain data lakes, data warehouses, and other storage solutions.

Collaboration & Problem Solving:

  • Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions.
  • Troubleshoot and resolve performance bottlenecks in big data pipelines.

Code Quality & Documentation:

  • Write clean, maintainable, and well-documented code.
  • Ensure compliance with data governance and security policies.

Required Skills & Qualifications:

Programming Skills:

  • Proficient in Python with experience in data processing libraries like Pandas and NumPy.
  • Strong experience with PySpark and Apache Spark.

Big Data & Cloud:

  • Hands-on experience with big data platforms such as Hadoop, Databricks, or similar.
  • Familiarity with cloud services like AWS (EMR, S3), Azure (Data Lake, Synapse), or Google Cloud (BigQuery, Dataflow).

Database Expertise:

  • Strong knowledge of SQL and NoSQL databases.
  • Experience working with relational databases like PostgreSQL, MySQL, or Oracle.

Data Workflow Tools:

  • Experience with workflow orchestration tools like Apache Airflow or similar.

Problem Solving & Communication:

  • Ability to solve complex data engineering problems efficiently.
  • Strong communication skills to work effectively in a collaborative environment.

Preferred Qualifications:

  • Knowledge of data Lakehouse architectures and frameworks.
  • Familiarity with machine learning pipelines and integration.
  • Experience in CI/CD tools and DevOps practices for data workflows.
  • Certification in Spark, Python, or cloud platforms is a plus.

Education:

  • Bachelors or Masters degree in Computer Science, Data Engineering, or a related field.

About the job

Apply before

Posted on

Job type

Full Time

Experience level

Education

Bachelor degree
Postgraduate degree

Location requirements

Hiring timezones

United States +/- 0 hours

About Inizio Partners Corp

Learn more about Inizio Partners Corp and their company culture.

View company profile
Claim this profileIC

Inizio Partners Corp

View company profile

Similar remote jobs

Here are other jobs you might want to apply for.

View all remote jobs

77 remote jobs at Inizio Partners Corp

Explore the variety of open remote roles at Inizio Partners Corp, offering flexible work options across multiple disciplines and skill levels.

View all jobs at Inizio Partners Corp

Find your dream job

Sign up now and join over 100,000 remote workers who receive personalized job alerts, curated job matches, and more for free!

Sign up
Himalayas profile for an example user named Frankie Sullivan