Senior Data Engineer
Senior Data Engineer
Rapid7 (Nasdaq: RPD) is advancing security with visibility, analytics, and automation delivered through our Insight cloud. Our solutions simplify the complex, allowing security teams to work more effectively with IT and development to reduce vulnerabilities, monitor for malicious behavior, investigate and shut down attacks, and automate routine tasks. Over 9,300 customers rely on Rapid7 technology, services, and research to improve security outcomes and securely advance their organization. For more information, visit our website, check out our blog, or follow us on LinkedIn.
The Opportunity
Rapid7 seeks a Sr. Data Engineer to build and maintain data infrastructure within the Data Engineering team's data platform. You will be responsible for deploying data pipelines and machine learning models in the cloud, implementing DevOps practices and developing data models within Rapid7’s Data Engineering Ecosystem. You must have demonstrated experience mentoring engineers, be adept with agile tools and methodology, demonstrate advanced analytical skills, technical and business knowledge and have a strong understanding of how to leverage the best industry standard tools and methodology to solve problems. The Sr. Data Engineer will work closely with Data Analysts, Data Scientists, and business teams to build data products to address business problems.
In this role you will:
- Lead and manage Rapid7 core data infrastructure with teammates using dbt, Fivetran, Snowflake, Airflow, GitHub, and AWS
- Architect and develop data practices for data integration, data modeling, unit testing, and data productionization
- Build and maintain the applications that ingest, analyze and store Rapid7's enterprise data
- Mentor and provide guidance to peer data engineers based on your experiences and technical expertise
- Perform data engineering projects within Snowflake such as developing data pipelines, data models and metadata management solutions
- Optimize data lifecycle management and process to improve delivery efficiency, data quality and data integrity
- Automate existing code and processes using scripting, CI/CD, infrastructure-as-code and configuration management tools
- Refine and manage Data Security Policies and RBAC (Role-based Access Control)
- Lead & evaluate data tooling as needed and research opportunities for data acquisition and processing of batch and streaming data
- Partner with the IT, Infrastructure and engineering teams on integration efforts between systems that impact data & Analytics
In return you will bring:
- 4+ years in a hands-on data engineering role performing data pipelining, infrastructure, integration and/or technical development of data architecture
- 4+ years of experience in at least one programming such as Python, Java, Scala etc. as well as advanced hands-on SQL work
- 3+ of experience working with a modern cloud data warehouse (preferably Snowflake) and orchestration tools (preferably Airflow)
- 2+ years of experience with a major cloud provider (preferably AWS); hands-on experience with code deployment in cloud environments using Terraform is a plus!
- Hands on experience working with the “modern data stack” (dbt, Fivetran, Stitch, etc)
- Experience as a leader within a data engineering team and ability to mentor teammates
- Strong written and verbal communication skills
- Highly collaborative in working teammates and stakeholders
- Experience with a CI/CD tool such as Github Actions and AWS Code Pipeline, a plus
- Working knowledge of data architecture, data warehousing, and metadata management
- BS or MS in Computer Science, Analytics, Statistics, Informatics, Information Systems or another quantitative field. or equivalent experience and certifications will be considered