Lead Data Engineer – KNIME
Our Opportunity:
Chewy is looking for a Lead Data Pipeline Engineer based in our Dania Beach, FL or Boston, MA office. In this role, you will help define and build the next generation of data pipeline process, infrastructure and implementation for one of the fastest growing E-commerce platforms.
This role will be responsible for maintaining and leveraging the BI owned, customer facing, Data Pipeline tool (Knime) to provide solutions to the various teams within Chewy. This role will develop technical solutions to the challenges of our business users; solving data accessibility, manipulation of data, curation of content from disparate sources, and work with our various Data Analytics and Data Science teams to enable their work.
Who You Are:
- A team player who recognizes the value of teamwork and respect
- Strong self-starter, comfortable working in ambiguity or lightly defined projects
- Someone who is looking to grow and invest in their career
- Ability to analyze a problem and develop testing solutions to accommodate the unique challenges of the project
- Loves to automate and build scaleable solutions to large and nuanced data challenges
- Passion wins at Chewy, we want to see your passion and love of all-things data
- Someone that searches for problems to be solved and people in need of support and takes initiative to resolve their root cause issue
What You’ll Do:
- Develop a framework for intake, construction and productionalization of Data Pipelines within the data community to provide self-service data pipeline creation to end-users
- Architect and define methodologies and processes to improve the ability of users to procure and leverage data from outside the organization
- Design platform integrations to allow the seamless interaction between systems and start down-stream workflows
- Support the creation and implementation of a data quality framework maximize potential for automated test execution across multiple data platforms
- Actively participate in the scoping, design and development of data pipelines to support end-user data needs while ensuring safeguards are in place to protect shared resources
- Ensure that technical solutions follow best practices, are reliable, easily maintainable and are scalable under sustained load
- Create or modify complex framework elements, modules, routines, scripts and data objects from a detailed design where efficiency, performance and reliability are critical
- Design and implement tools and frameworks to simplify test development, including simulators/mocks and data generators
- Build test infrastructure to fit into our CI platform
- Assess and evaluate new tools and data integration methodologies
- Provided peer feedback in design and code reviews
What You’ll Need:
- B.S. in Computer Science, Engineering, or other technical discipline
- Ability to distill complex technical requirements into consumable action-items for developers
- 3-5 years as a Quality Engineer, Software Engineer or ETL Engineer
- 3-5 years of experience using ETL tools such as Informatica, Alteryx, or KNIME
- 3-5 years of experience in SQL with ability to query databases independently to find root cause of issues
- Ability to write and manipulate Python/Unix/PowerShell scripts
- Familiarity with RDBS and Columnar database design
Bonus:
- Knowledge of Atlassian suite of tools (Jira, Bamboo, Confluence)
- Experience working in Agile teams (Scrum, Kanban)
- Experience with KNIME
- Familiarity with at least one code language, ideally Python
If you have a disability under the Americans with Disabilities Act or similar law, or you require a religious accommodation, and you wish to discuss potential accommodations related to applying for employment at our company, please contact [email protected].
To access Chewy’s Privacy Policy, which contains information regarding information collected from job applicants and how we use it, please click here: Chewy Privacy Policy (https://www.chewy.com/app/content/privacy).