Job Description
Our Opportunity: We are looking for a
Data Engineer III at our facility in Minneapolis, Minnesota, to own/contribute towards data platform on AWS cloud.
What You'll Do: - Contribute towards the architecture and build/maintain infrastructure using Cloud computing technology.
- Build and maintain scheduling/workflow orchestration applications.
- Build containerized applications with microservices architecture.
- Establish continuous integration and continuous deliver pipeline.
- Performance tuning of data pipelines.
- Monitor data pipelines for accuracy, missing data, enhancements, changes, and billing volumes to ensure all data is captured and processed accurately and when needed.
- Reconcile data issues and alerts between various systems, finding opportunities to innovate and drive improvements.
- Develop and maintain complex data ingestion and transformations for data originating from multiple data sources (structured/unstructured).
- Design and implement dimension data modeling (Star Schema, Snowflake and Galaxy) for on-premise and cloud data warehouse infrastructure.
- Implement the strategy, design, execution, system configuration, and operations of the data platform that help in data-driven decision making.
- Lead the evaluation, implementation, and deployment of emerging tools and technologies.
- Work with cross-functional stakeholders in defining and documenting requirements for building high-quality and impactful data products.
- Create operational reports using visualization/business intelligence tools.
What You'll Need: - Bachelor's degree in Electrical Engineering, Computer Science, Computer Engineering, or related field and 5 years of experience;
- Will accept a Master's degree and 3 years of experience.
- Experience must include 3 years with: implement the strategy, design, execution, system configuration, and operations of the data platform that help in data-driven decision making;
- Building scalable analytics solution, including data processing, storage, and serving large-scale data through batch and stream;
- MPP database (e.g., AWS Redshift, Exasol) big data platforms, open-source table formats like Delta Lake, Hudi, lceberg;
- Building and maintaining infrastructure using AWS;
- Orchestration tools like Airflow;
- Cloud native DevOps platform using Gitlab, AWS CloudFormation, Terraform;
- Dimensional data modeling (Star, Snowflake, Galaxy chema) and ER modeling;
- Python;
- SQL, data warehouse, performance tuning, and data pipelines;
- AWS Glue, Pyspark;
- AWS ecosystem and services including AWS S3, ECS, EC2, EMR, Athena, Lambda, Glue;
- RESTful API's;
- NoSQL database like Cassandra;
- Implement federated logical data platform;
- Design and implement data pipelines using modern data engineering approaches and tools:
- Spark, PySpark, Docker, cloud native DWH (e.g., Snowflake, Exasol, Redshift), AWS Glue, S3, Athena, Lambda, Postgres (AWS RDS Aurora); and
- CI/CD platforms.
- The position is eligible for the Employee Referral Program.
Chewy is committed to equal opportunity. We value and embrace diversity and inclusion of all Team Members. If you have a disability under the Americans with Disabilities Act or similar law, and you need an accommodation during the application process or to perform these job requirements, or if you need a religious accommodation, please contact CAAR@chewy.com. If you have a question regarding your application, please contact HR@chewy.com. To access Chewy's Customer Privacy Policy, please click here. To access Chewy's California CPRA Job Applicant Privacy Policy, please click here.
Job Tags