Rapid7 is a leading provider of security data and analytics solutions that enable organizations to implement an active, analytics-driven approach to cyber security. We combine our extensive experience in security data and analytics and deep insight into attacker behaviors and techniques to make sense of the wealth of data available to organizations about their IT environments and users. Our solutions empower organizations to prevent attacks by providing visibility into vulnerabilities and to rapidly detect compromises, respond to breaches, and correct the underlying causes of attacks. Rapid7 is trusted by more than 9000 organizations across 125 countries, including 52% of the Fortune 100. To learn more about Rapid7 or get involved in our threat research, visit www.rapid7.com.
Rapid7 is looking for a Data Engineer to design and implement data models, own the ETL process & influence data warehousing strategy. You'll work with modern data warehousing technologies such as Snowflake, Airflow, Docker and more. In this role you'll be working in tandem with BI & Analytics analysts and managers to assure that we have the right data at the right time to support our business needs. All of this in support of our “single source of truth” and continuing our progression up the analytics maturity curve towards pervasive predictive and prescriptive analytics.
If you have hands-on experience preparing ELT solutions and data warehouses, plus demonstrated examples of translating business objectives and requirements into the data needed to support key analyses, then we're looking for you. At Rapid7 you'll collaborate with a creative, analytical and data-driven team to bring a single source of truth and self-service analytics to over 1600 employees in a culture that values innovation and teamwork.
Create and maintain optimal data pipeline architecture, and implement structures and process to assure efficiency, repeatability and standardization in the use of data within the BI team and Rapid7 as a whole
Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL, Python, and other languages.
Collaborate with the IT and Infrastructure teams on integration efforts between systems that impact data & Analytics
Assemble complex data sets that meet reporting and analytics requirements from business stakeholders
Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
Self-starter orientation, able to handle the prioritization of multiple project and tasks under time-bound deadlines
Develop processes to improve data management and promote data standards, data quality & data security
BS or MS in Computer Science, Analytics, Statistics, Informatics, Information Systems or another quantitative field. or equivalent experience and certifications will be considered
More than 1 year of experience with modern data warehouse (Snowflake, Redshift, Azure, etcl)
More than 3 years experience building data pipelines
Extremely Proficient in SQL
Proficiency in Python, R, or other scripting language
Extensive experience in data integration (ETL) tools such as Informatica, SSIS, Talend, and Fivetran, Amazon Glue
Understanding of a version control system such as Git
Working knowledge and solid understanding of data architecture, data warehousing, and metadata management
Track record of managing and improving complex processes and very strong organizational and problem-solving skills