Pyspark Developer


Company Name


Job Description

  • Key Responsibilities:

    Big Data Development & Analytics:

    • Design and develop scalable big data solutions using Python and PySpark
    • Build distributed data processing applications on Apache Spark framework
    • Implement ETL/ELT pipelines for large-scale data transformation and processing
    • Develop real-time and batch data processing systems

    Project Lifecycle Support:

    • Actively support consulting teams across all project phases including problem definition, effort estimation, diagnosis, solution generation, design, and deployment
    • Participate in data architecture design and technical solution planning
    • Collaborate with cross-functional teams to deliver comprehensive data solutions

    Research & Solution Development:

    • Explore alternative data processing solutions through comprehensive research including literature surveys, public domain information analysis, and vendor evaluations
    • Build Proof of Concepts (POCs) for big data technologies and frameworks
    • Evaluate and recommend data processing tools and methodologies

    Requirements Analysis & Design:

    • Create technical requirement specifications from business needs
    • Define data processing workflows and detailed technical designs based on functional requirements
    • Design data models and schemas for optimal performance and scalability

    Issue Resolution & Optimization:

    • Configure solution requirements on various big data platforms and products
    • Diagnose root causes of performance issues and data processing bottlenecks
    • Identify and implement solution alternatives for complex data challenges
    • Optimize Spark jobs for better performance and resource utilization

    Quality & Innovation:

    • Contribute to unit-level and organizational data engineering initiatives
    • Provide high-quality, value-adding data solutions to customers
    • Support clients in their digital transformation journey through advanced analytics and big data capabilities

    Mandatory Technical Skills:

    Core Technologies:

    • Python - Strong proficiency in Python programming for data processing and analysis
    • PySpark - Extensive experience with PySpark for distributed data processing and Apache Spark ecosystem

Job Details

Experience : 3 To 8
Number Of Vacancies : 1000
Job Type : Permanent
Industry Type : IT/Software
Salary : 3 Lac - 9 Lac P.A

Education Summary

UG : BE/B.Tech,B.Sc,BCA PG : M.Sc,MCA,M.Tech

Contact Details

Contact Person : NA
Contact Number : 7338856831
e-mailId : askus@infosys.com
Address : Infosys Ltd. Plot No. 1, Rajiv Gandhi Infotech Park, Hinjawadi, Taluka Mulshi, Pune.

Back

Office Location

Central Jakarta No 1234, Jakarta, Indonesia