Job Description
ESSENTIAL DUTIES AND RESPONSIBILITIES include the following. Other duties and special projects may be assigned.
- Design, create and maintain data pipelines will be the primary responsibility of the Data Engineer.
- Support reporting and general analytics needs of the department.
- Drive automation through effective metadata management.
- Assist with renovating the data management infrastructure to drive automation in data integration and management.
- Utilize modern data preparation, integration and AI-enabled metadata management tools and techniques.
- Track data consumption patterns.
- Perform intelligent sampling and caching.
- Monitor schema changes.
- Recommend and automate integration flows.
- SENIOR LEVEL RESPONSIBILITIES * Work with data science teams and with business (data) analysts to refine their data requirements for various data and analytics initiatives. * Propose appropriate (and innovative) data ingestion, preparation, integration and operationalization techniques. * Train counterparts such as data scientists, data analysts, LOB users or any data consumers in data pipelining and preparation techniques. * Ensure that data users and consumers use the data provisioned to them responsibly through data governance and compliance initiatives. Participate in vetting and promoting content created in the business and by data scientists to the curated data catalog for governed reuse. * Become a data and analytics evangelist by promoting the available data and analytics capabilities and expertise to business unit leaders and educating them in leveraging these capabilities in achieving their business goals.
- Adheres to and complies with applicable, federal and state laws, regulations and guidance, including those related to anti-money laundering (i.e. Bank Secrecy Act, US PATRIOT Act, etc.).
- Adheres to Bank policies and procedures and completes required training.
- Identifies and reports suspicious activity.
EDUCATION
Bachelor's Degreein computer science, statistics, applied mathematics, data management, information systems, information science or a related quantitative field required Master's DegreeAn advanced degree in computer science preferred PhDstatistics, applied mathematics preferred information science (MIS), data management, information systems, information science (post-graduation diploma or related) or a related quantitative field or equivalent work experience preferred combination of IT skills, data governance skills, analytics skills and banking domain knowledge with a technical or computer science degree preferred
EXPERIENCE
- 6 years of work experience in data management disciplines including data integration, modeling, optimization and data quality, and/or other areas directly relevant to data engineering responsibilities and tasksrequired
- 6 years of experience working in cross-functional teams and collaborating with business stakeholders in the banking business domain, in support of a departmental and/or multi-departmental data management and analytics initiativerequired
- Strong experience with advanced analytics tools for object-oriented/object function scripting using languages such as R, Python, Java, and Scalarequired
- Strong experience with popular database programming languages including SQL and PL/SQL for relational databases and certifications on upcoming NoSQL/Hadoop oriented databases like MongoDB and Cassandra for nonrelational databasesrequired
- Strong experience in working with large, heterogeneous datasets in building and optimizing data pipelines, pipeline architectures and integrated datasets using traditional data integration technologies These should include ETL/ELT, data replication/CDC, message-oriented data movement, API design and access and upcoming data ingestion and integration technologies such as stream data integration, CEP and data virtualizationrequired
- Strong experience in working with SQL on Hadoop tools and technologies including HIVE, Impala, Presto and others from an open source perspective and Hortonworks Data Flow (HDF), Dremio, Informatica, Talend among others from a commercial vendor perspectiverequired
- Strong experience in working with and optimizing existing ETL processes and data integration and data preparation flows and helping to move them in productionrequired
- Strong experience in working with both open-source and commercial message queuing technologies (Kafka, JMS, Azure Service Bus, Amazon Simple queuing Service), stream data integration technologies such as Apache Nifi, Apache Beam, Apache Kafka Streams, Amazon Kinesis, others and stream analytics technologies (Apache Kafka, KSQL, Apache Spark)required
- Basic experience working with popular data discovery, analytics and BI software tools like Tableau, and OBI for semantic-layer-based data discoveryrequired
- Strong experience in working with data science teams in refining and optimizing data science and machine learning models and algorithmsrequired
- Basic experience in working with data governance teams and specifically business data stewards and the CISO in moving data pipelines into production with appropriate data quality, governance and security standards and certificationrequired
KNOWLEDGE, SKILLS AND ABILITIES
- Strong ability to design, build and manage data pipelines for data structures encompassing data transformation, data models, schemas, metadata and workload management. The ability to work with both IT and business in integrating analytics and data science output into business processes and workflows.
- Demonstrated ability to work across multiple deployment environments including cloud, on-premises and hybrid, multiple operating systems and through containerization techniques such as Docker, Kubernetes, AWS Elastic Container Service and others.
- Proficiency in agile methodologies and the capability of applying DevOps and increasingly DataOps principles to data pipelines to improve the communication, integration, reuse and automation of data flows between data managers and consumers across an organization
- Deep domain knowledge or previous experience working in the banking business would be a plus.
- Strong experience in working with SQL on Hadoop tools and technologies including HIVE, Impala, Presto and others from an open source perspective and Hortonworks Data Flow (HDF), Dremio, Informatica, Talend among others from a commercial vendor perspective.
ADDITIONAL KNOWLEDGE, SKILLS AND ABILITIES
* Strong ability to design, build and manage data pipelines for data structures encompassing data transformation, data models, schemas, metadata and workload management. The ability to work with both IT and business in integrating analytics and data science output into business processes and workflows.
* Demonstrated ability to work across multiple deployment environments including cloud, on-premises and hybrid, multiple operating systems and through containerization techniques such as Docker, Kubernetes, AWS Elastic Container Service and others.
* Proficiency in agile methodologies and the capability of applying DevOps and increasingly DataOps principles to data pipelines to improve the communication, integration, reuse and automation of data flows between data managers and consumers across an organization
* Deep domain knowledge or previous experience working in the banking business would be a plus.
* Strong experience in working with SQL on Hadoop tools and technologies including HIVE, Impala, Presto and others from an open source perspective and Hortonworks Data Flow (HDF), Dremio, Informatica, Talend among others from a commercial vendor perspective.
Date Posted
09/17/2022
Views
6
Similar Jobs
Data Analyst I - Empath Health
Views in the last 30 days - 16
Suncoast Hospice is seeking a Data Analyst I to join the Empath team The position requires 3 years of SQL experience reporting tool experience and pro...
View DetailsTerrain Engineer 4 - Cole Engineering Services
Views in the last 30 days - 8
The job requires a senior engineer to work on a team developing nextgeneration collective training software The engineer will support the integration ...
View DetailsSupply Chain Engineer - C.H. Robinson
Views in the last 30 days - 2
The job posting is for a Supply Chain Engineer at CH Robinson requiring a minimum of 2 years of professional work experience advanced proficiency in M...
View DetailsMetaPulse Data Manager - KnowBe4
Views in the last 30 days - 6
KnowBe4 is a company that provides security awareness training and simulated phishing platform They have been ranked as a best place to work and have ...
View DetailsMetaPulse QA Engineer - KnowBe4
Views in the last 30 days - 6
KnowBe4 is a company that provides security awareness training and simulated phishing platform They have been ranked as a best place to work and have ...
View DetailsSoftware Developer - AACSB
Views in the last 30 days - 8
AACSB is a leading organization in business education providing quality assurance thought leadership and learning opportunities The company values qua...
View Details