Jr. Data Engineer
Charles Schwab
Lone Tree, CO

About $130,000 - $170,000 a year

Skills
Your Opportunity

Do you want to be part of a Data Solutions Delivery team managing over 150+ terabytes of data and building the next generation analytics platform for a leading financial firm with over $3.2 trillion in assets under management? At Schwab, the Global Data Technology (GDT) organization governs the strategy and implementation of the enterprise data warehouse and emerging data platforms. We help Marketing, Finance and executive leadership make fact-based decisions by integrating and analyzing data.

We are looking for a Data Engineer who has passion for data and comes with data engineering background. Someone who has experience in designing and coding batch as well as real time ETL and one who wants to be part of a team that is actively designing and implementing the big data lake and analytical architecture on Hadoop. You will have the opportunity to grow in responsibility, work on exciting and challenging projects, train on emerging technologies and help set the future of the Data Solution Delivery team.

What you’re good at
Designing schemas, data models and data architecture for Hadoop and HBase environments
Building and maintaining code for real time data ingestion using Java, MapR-Streams (Kafka) and STORM.
Implementing data flow scripts using Unix / Hive QL / Pig scripting
Designing, building and support data processing pipelines to transform data using Hadoop technologies
Designing, building data assets in MapR-DB (HBASE), and HIVE
Developing and executing quality assurance and test scripts
Working with business analysts to understand business requirements and use cases
What you have
Minimum of 2 years of experience in understanding of best practices for building and designing ETL code Strong SQL experience with the ability to develop, tune and debug complex SQL applications is required
Knowledge in schema design, developing data models and proven ability to work with complex data is preferred
Hands-on experience in Java object oriented programming (At least 2 years)
Hands-on experience with Hadoop, MapReduce, Hive, Pig, Flume, STORM, SPARK, Kafka and HBASE is preferred
Understanding Hadoop file format and compressions is preferred
Familiarity with MapR distribution of Hadoop is preferred
Understanding of best practices for building Data Lake and analytical architecture on Hadoop is preferred
Scripting / programming with UNIX, Java, Python, Scala etc. is preferred
Strong SQL experience with the ability to develop, tune and debug complex SQL applications is required
Knowledge in real time data ingestion into Hadoop is preferred
Experience in working in large environments such as RDBMS, EDW, NoSQL, etc. is preferred
Knowledge of Big Data ETL such as Informatica BDM and Talend tools is preferred
Understanding security, encryption and masking using Kerberos, MapR-tickets, Vormetric and Voltage is preferred
Experience with Test Driven Code Development, SCM tools such as GIT, Jenkins is preferred
Experience with Graph database is preferred