Data Analyst with ETL PYTHON and JAVA

at DCM Infotech Limited
Published February 2, 2024
Location Philadelphia, PA
Category Default  
Job Type Full-time  

Description

Can we get someone strong in Python backed developer profile who is into data processing and not data analytics
URGENT - Look for ex comcast

Python Developer ( Not FULL STACK DEVELOPERS, Need candidates expert on DATA)

Responsibilities:
• Hands-on building of ETL pipelines using our internal framework written in Python
• Hands-on solutioning of real time REST APIs or other solutions for streaming data from Graph
• Modify existing application code or interfaces or build new application components from detailed requirements.
• nalysis of requirements, support of the design, development of the code, testing, debugging, deployment, and maintenance of those programs and interfaces. Documentation of the work is essential
• Participation in most aspects of programming and application development, including file design, update, storage, and retrieval
• Enhance processes to resolve operational problems and add new functions taking into consideration schedule, resource constraints, process complexity, dependencies, assumptions and application structure
• bility to maintain the developed solution on an on-going basis is essential
• bility to follow the existing development methodology and coding standards, and ensure compliance with the internal and external regulatory requirements
• Develop and implement databases, data collection systems, data analytics and other strategies that optimize statistical efficiency and quality
• cquire data from primary or secondary data sources and maintain databases/data systems
• Work with management to prioritize business and information needs
• Locate and define new process improvement opportunities
• Document design and data flow for existing and new applications being built.
• Co-ordinate with multiple different teams QA, Operations and other development team within the organization.
• Testing methods, including unit and integration testing (JUnit, JMock)
• bility to integrate with large teams, demonstrating strong verbal and written communication skills
• Utilization of software configuration management tools
• Code deployment and code versioning tools
• Excellent Communication Skills

Qualifications:

Bachelor's degree preferably with Computer Science background.

At least 5+ years of experience implementing complex ETL pipelines preferably with Spark toolset.

At least 5+ years of experience with Java particularly within the data space

Technical expertise regarding data models, database design development, data mining and segmentation techniques

Good experience writing complex SQL and ETL processes

Excellent coding and design skills, particularly in Java/Scala and Python and or Java.

Experience in UNIX (AIX) , UNIX Scripting

Experience in JAVA, J2EE, EJB, JavaBeans, JavaScript, servlets, JDBC, struts, Hibernate

Experience working with large data volumes, including processing, transforming and transporting large-scale data

Experience in AWS technologies such as EC2, Redshift, Cloud formation, EMR, AWS S3, AWS Analytics required.

Big data related AWS technologies like HIVE, Presto, Hadoop required.

AWS certification is preferable: AWS Developer/Architect/DevOps/Big Data

Excellent working knowledge of Apache Hadoop, Apache Spark, Kafka, Scala, Python etc.

Strong analytical skills with the ability to collect, organize, analyze, and disseminate significant amounts of information with attention to detail and accuracy

Good understanding & usage of algorithms and data structures

Good Experience building reusable frameworks.

Experience working in an Agile Team environment.

AWS certification is preferable: AWS Developer/Architect/DevOps/Big Data

Excellent communication skills both verbal and written