Job Description:
Overview
As a Data Engineer, you will be the central architect for backend development ETL projects, responsible for ensuring a reliable, optimized data pipeline. You will also collaborate closely with our Directors and Managers to ensure they have access to the reports and data that are critical for their diverse needs. This requires understanding many aspects of the business and the systems that drive them, including: ecommerce, manufacturing, and fulfillment.
This role offers a unique opportunity to engage in meaningful projects, and solve critical challenges. You will drive initiatives that empower departments with data, automated manual reporting, and enable our adoption of AI in innovative ways. While you will gain exposure to transformative AI projects, your primary focus will be on developing a robust data lake that facilitates reporting and analysis.
The Data Engineer is a fully onsite role in Columbus, Ohio. Remote work is not available. Applicants must be authorized to work in the United States for any employer.
Responsibilities
Maintain efficient data pipelines, addressing data challenges such as inconsistency, quality issues, and complex transformation
Ingest internal and external data into the bronze layer
Collaborate with the team to understand pipeline structure and mechanize ad hoc data streams
Transform raw data into clean and optimized usable building blocks in the silver layer
Work with departments to understand business needs and prepare and format data for reporting purposes in the gold layer
Utilize GCP services such as Cloud Scheduler and Cloud Functions
Develop comprehensive testing strategies to ensure data integrity and pipeline functionality using Python notebooks. Collaborate closely with the Data Manager to design and execute data reporting pipelines
Lead technical project development with a high degree of autonomy, owning key tech improvements such as writing Python SDKs and researching efficient methods of data transformation
Qualifications
2-5 years of relevant data engineering experience
Bachelor’s degree in Computer Science or a related field
Proficiency in Python and its data-related libraries (Pandas, NumPy, SciPy)
2 years of experience designing and optimizing complex SQL pipelines in GCP
Experience as a lead developer defining project requirements
Experience working with diverse databases, including e-commerce
Experience ingesting data from diverse APIs
Experience with version control using Git
Excellent communication skills and problem-solving abilities
Demonstrated ability to work independently and collaboratively
High attention to detail, particularly with data security best practices
Experience working with infrastructure as code tools such as Terraform
Experience developing SDKs for data ingestion workflows
By applying to Rogue, regardless of the platform you choose to use, you are agreeing to Rogue's preferred methods of communication (i.e. text message). Submitting an application, through whatever online forum is ultimately used, constitutes a knowing and voluntary agreement to send and receive text messages during the recruitment process.