Python Developer with Data Processing

at R4 IT Solutions Inc
Published February 2, 2024
Location Philadelphia, PA
Category Default  
Job Type Full-time  


Role : Python Developer

Location: Philadelphia PA

Contract: Long term

Job Description


Python ETL Developer with strong Data Processing

Bachelor s degree preferably with Computer Science background.

At least 5+ years of experience implementing complex ETL pipelines preferably with Spark toolset.

At least 5+ years of experience with Java particularly within the data space

Technical expertise regarding data models, database design development, data mining and segmentation techniques

Good experience writing complex SQL and ETL processes

Excellent coding and design skills, particularly either in Scala or Python.

Strong practical working experience with Unix scripting in at least one of Python, Perl, Shell (either bash or zsh).

Experience in AWS technologies such as EC2, Redshift, Cloud formation, EMR, AWS S3, AWS Analytics required.

Experience designing and implementingdata pipelines in a onprem/cloud environment is required.

Experience building/implementing data pipelines usingDatabricks/On prem or similar cloud database.

Expert level knowledge of using SQL to write complex, highly optimized queries across large volumes of data.

Hands-on object-oriented programming experience using Python is required.

Professional workexperience building real-time data streams using Spark and Experience in Spark.

Knowledge or experience in architectural best practices in building data lakes

Develop and work with APIs

Develop and maintain scalable data pipelines and build out new API integrations to support continuing increases in data volume and complexity.

Collaborate with analytics and business teams to improve data models that feed business intelligence tools, increase data accessibility, and foster data-driven decision making across the organization.

Implement processes and systems to monitor data quality, to ensure production data accuracy, and ensure key stakeholder and business process access.

Write unit/integration tests, contribute to engineering wiki, and documents.

Perform data analysis required to troubleshoot data related issues and assist in the resolution of data issues.

Experience developing data integrations and data quality framework based on established requirements.

Experience with CI/CD processes and tools (e.g., concourse, Jenkins).

Experience with test driven development writing unit tests, test coverage using PyTest, PyUnit, pytest-cov libraries.

Experience working in an Agile environment.

Good understanding & usage of algorithms and data structures

Good Experience building reusable frameworks.

Experience working in an Agile Team environment.

AWS certification is preferable: AWS Developer/Architect/DevOps/Big Data

Excellent communication skills both verbal and written

Python Developer with Data Processing