Plano, TX, United States
24 hours ago
Lead Software Engineer

We have an opportunity to impact your career and provide an adventure where you can push the limits of what's possible.

As a Lead Big Data Software Engineer at JPMorgan Chase within the Consumer & Community Banking, you are an integral part of an agile team that works to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. As a core technical contributor, you are responsible for conducting critical technology solutions across multiple technical areas within various business functions in support of the firm’s business objectives.

Job responsibilities:

Promote to become data domain expert and drive to know everything about the data on the platform Create Functional and Technical Specifications, Epics and User Stories, Process Flows, Data Analysis, Mapping Documents, Implementation Plan, Agile artifacts Migrate from Hadoop to AWS using Pipelines, EMR Develop, enhance and test new/existing interfaces. The candidate will be part of existing agile team and will work on developing, enhancing ETL pipelines, design solutions, Identify, analyze, and interpret trends or patterns in complex data sets and transforming existing ETL logic into Hadoop Platform Innovate new ways of managing, transforming and validating data Establish and enforce guidelines to ensure consistency, quality and completeness of data assets Apply quality assurance best practices to all work products Gain experience of working in a development teams, using agile techniques and Object-Oriented development and scripting languages, is preferred. Adds to team culture of diversity, equity, inclusion, and respect

Required qualifications, capabilities, and skills:

Formal training or certification on software engineering concepts and 5+ years of applied experience Knowledge of application, data, and infrastructure architecture disciplines Strong experience with documentation and structuring information in tools like Confluence and Jira Experience in SparkSQL,Imapala and Bigdata technologies Familiar with Data Science concepts and applying them to analyze large volumes of data
Comfortable with data concepts; Oracle, Java, Python, Spark, Kafka, HDFS, AirFlow, Elastic Search. Working proficiency in SDLC CI/CD Execution (GitHub, Jenkins, SNOR, Spinnaker, AIM etc.) Minimum 2+  Experience with AWS Services like Lambda EC2 Experience in real time streaming data Strong Experience with UNIX shell scripting is must Experience with relational database environment (Oracle, Teradata, SQL Server, etc.) leveraging databases, tables/views, stored procedures, agent jobs, etc. with strong analytical skills with the ability to collect, organize, analyze, and disseminate significant amounts of information with attention to detail and accuracy Handle Dev Ops effort in terms of CICD, Scanning, Code, Performance testing and Test overaged  Preferred qualifications, capabilities, and skills: Experience in Athena, EMR, Redshift, Glue, Kinesis, Aurora,RDS,S3 Knowledge in one or more modern programming languages like Java or Python is a plus AWS Cloud Practitioner certification a plus for applicant and expected upon joining team Experience working on AdTech/MarTech platforms is a plus    

 

Confirm your E-mail: Send Email