Senior Data Engineer (PySpark, Hadoop, Spark)

Posted:
4/3/2025, 3:51:38 PM

Location(s):
पिंपरी चिंचवड, महाराष्ट्र, India ⋅ Maharashtra, India ⋅ Pune, Maharashtra, India ⋅ महाराष्ट्र, India

Experience Level(s):
Senior

Field(s):
Data & Analytics

    Software Requirements:

    • PySpark
    • Hadoop
    • Spark
    • Python
    • Unix
    • HDFS
    • PyArrow
    • SQL

    Overall Responsibilities:

    • Develop and optimize data processing workflows utilizing PySpark.
    • Implement and manage data pipelines within Hadoop and Spark ecosystems.
    • Collaborate with data scientists, analysts, and stakeholders to understand data requirements and deliver effective solutions.
    • Ensure data quality and integrity across all processes and workflows.
    • Monitor and resolve performance issues in big data applications.
    • Stay updated with the latest trends and advancements in big data technologies.

    Technical Skills:

    Category-wise:

    PySpark and Spark:

    • Proficient in PySpark for data processing.
    • Experience in Spark development, including performance optimization techniques.

    Hadoop and HDFS:

    • Competent with Hadoop ecosystems.
    • Skilled in handling HDFS for data storage and retrieval.

    Programming and Libraries:

    • Strong programming skills in Python.
    • Familiarity with supported libraries such as PyArrow.

    Unix:

    • Basic experience with Unix systems for data processing tasks.

    Databases (SQL):

    • Working knowledge of SQL for database management and querying.

    Experience:

    • Over 7 years of experience in big data environments.
    • Extensive exposure to PySpark and Spark development.
    • Significant experience in data engineering and managing large datasets.

    Day-to-Day Activities:

    • Develop, test, and deploy data processing pipelines using PySpark.
    • Perform data extraction, transformation, and loading (ETL) activities.
    • Collaborate with cross-functional teams to gather requirements and deliver solutions.
    • Optimize and troubleshoot existing data workflows and pipelines.
    • Document processes, workflows, and data models.
    • Continuously monitor and improve data processing performance.

    Qualifications:

    • Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field.
    • Certifications in big data technologies are a plus.

    Soft Skills:

    • Excellent problem-solving and analytical skills.
    • Strong communication skills to collaborate with team members and stakeholders.
    • Ability to work independently and as part of a team.
    • Attention to detail and a proactive approach to identifying and solving issues.
    • Adaptability to evolving technologies and methodologies.

    S​YNECHRON’S DIVERSITY & INCLUSION STATEMENT
     

    Diversity & Inclusion are fundamental to our culture, and Synechron is proud to be an equal opportunity workplace and is an affirmative action employer. Our Diversity, Equity, and Inclusion (DEI) initiative ‘Same Difference’ is committed to fostering an inclusive culture – promoting equality, diversity and an environment that is respectful to all. We strongly believe that a diverse workforce helps build stronger, successful businesses as a global company. We encourage applicants from across diverse backgrounds, race, ethnicities, religion, age, marital status, gender, sexual orientations, or disabilities to apply. We empower our global workforce by offering flexible workplace arrangements, mentoring, internal mobility, learning and development programs, and more.


    All employment decisions at Synechron are based on business needs, job requirements and individual qualifications, without regard to the applicant’s gender, gender identity, sexual orientation, race, ethnicity, disabled or veteran status, or any other characteristic protected by law.

    Candidate Application Notice