Data Engineer III - Databricks and Python

Overview
Data Engineer III - Databricks and Python at JPMorganChaseAbout Us
J.P. Morgan is a global leader in financial services, providing strategic advice and products to the world’s most prominent corporations, governments, wealthy individuals and institutional investors. Our first-class business in a first-class way approach to serving clients drives everything we do. We strive to build trusted, long-term partnerships to help our clients achieve their business objectives. We are an equal opportunity employer and value diversity and inclusion. We do not discriminate on the basis of any protected attribute and provide reasonable accommodations as needed.Job Description
As a Data Engineer - Databricks in our Digital Intelligence team, you will design, develop, and maintain scalable data processing solutions using Databricks, Python, and AWS. You will collaborate with cross-functional teams to deliver high-quality data solutions that support business objectives.Responsibilities
Execute software solutions, design, development, and technical troubleshooting with the ability to think beyond routine or conventional approaches to build solutions or break down technical problems.Create secure and high-quality production code and maintain algorithms that run synchronously with appropriate systems.Produce architecture and design artifacts for complex applications while ensuring design constraints are met by software code development.Gather, analyze, synthesize, and develop visualizations and reporting from large, diverse data sets to improve software applications and systems.Proactively identify hidden problems and patterns in data to drive improvements in coding hygiene and system architecture.Contribute to software engineering communities of practice and events that explore new and emerging technologies.Provide guidance to the immediate team of software engineers on daily tasks and activities.Set overall guidance and expectations for team output, practices, and collaboration.Anticipate dependencies with other teams to deliver products in line with business requirements.Manage stakeholder relationships and the team’s work in accordance with compliance standards and service level agreements.Required Qualifications, Capabilities, And Skills
Bachelor’s degree in Computer Science, Information Technology, or a related field.Formal training or certification on software engineering concepts and applied experience.Hands-on experience in data mapping, data architecture, and data modeling on Databricks.Extensive experience in AWS, design, implementation, and maintenance of data pipelines using Python, PySpark on Databricks.Proficient in Python and PySpark, able to write and execute complex queries to perform curation and build views required by end users.Strong understanding of front-end and back-end technologies, with a focus on creating seamless user experiences.Extensive experience in Databricks data engineering, data warehousing concepts, ETL processes (Job Runs, Data Ingestion, Delta Live Tables, Spark Streaming).Experience in standing up and maintaining EC2/ECS instances, S3, Glue, and Lambda services.Experience in building Notebooks with complex code structures and debugging failed jobs.Proven experience in performance tuning to optimize job execution and prevent bottlenecks.Proven ability to deliver high-quality features into production systems in a rapid, iterative development environment.Preferred Qualifications, Capabilities, And Skills
Experience with machine learning and data science workflows.Familiarity with data visualization tools and techniques.Knowledge of data governance and security best practices.Experience in carrying out data analysis to support business insights.Employment ..... full job details .....