New York, NJ, USA
21 days ago
Associate- BIM (9720)
Position Summary

To be a driven business analyst who can work on complex Analytical problems and help the customer in better business decision making especially in the area of pharma/life sciences (domain).

Job Responsibilities

Your Responsibilities includes but not limited.

•Engage with Client to participate in requirement gathering, Status update on work, UAT and be the key partner in the overall engagement

• Participates in ETL Design using any python framework of new or changing mappings and workflows with the team and prepares technical specifications

• Crafts ETL Mappings, Mapplets, Workflows, Worklets using Informatica PowerCenter

• Write complex SQL queries with performance tuning and optimization

• Should be able to handle task independently and lead the team

• Responsible for unit testing, integration testing and UAT as and when required

• Good communication Skills

Education BE/B.TechMaster of Computer ApplicationWork Experience

Must have Skills:
• Advanced knowledge of PySpark ,python, pandas, numpy frameworks.
• Minimum 3 years of extensive experience in design, build and deployment of Spark/Pyspark - for data integration.
• Deep experience in developing data processing tasks using pySpark such as reading data from external sources, merge data, perform data enrichment and load in to target data destinations
• Create Spark jobs for data transformation and aggregation
• Spark query tuning and performance optimization - Good understanding of different file formats (ORC, Parquet, AVRO) to optimize queries/processing and compression techniques.
• Deep understanding of distributed systems (e.g. CAP theorem, partitioning, replication, consistency, and consensus)
• Experience in Modular Programming & Robust programming methodologies
• ETL knowledge and have done ETL development using any python framework
• Advanced SQL knowledge
• Ability to perform multiple task in continually changing environment
• Worked with Redshift/Synapse/Snowflake in the past Preferable.
• Good understanding and experience in the SDLC phases like the Requirements Specification, Analysis, Design, Implementation, Testing, Deployment and Maintenance
Optional:
• Used any of python framework Airflow/Luigi/Dagster/Apache Beam
• Pyspark
• Data warehouse
• Excellent troubleshooting skills
• Experience in different Operating Systems/ platforms like Windows, Linux and UNIX
• Work in an Agile environment, adopt to obstacle culture and team first attitude
• Lead/Participate in end-to-end software life cycle which includes Design, Plan, Develop, Automate, and Perform Various Tests
• Document software design, flows and use cases

Behavioural Competencies OwnershipTeamwork & LeadershipCultural FitMotivation to Learn and GrowTechnical Competencies Problem SolvingLifescience KnowledgeCommunicationPySparkPythonSQLAWS RDS
Confirm your E-mail: Send Email