Data Engineer III

Full Time
Minneapolis, MN, USA
1 month ago

Our Opportunity:

We are looking for a Data Engineer III at our facility in Minneapolis, Minnesota, to own/contribute towards data platform on AWS cloud.

What You’ll Do: 

  • Contribute towards the architecture and build/maintain infrastructure using Cloud computing technology.
  • Build and maintain scheduling/workflow orchestration applications.
  • Build containerized applications with microservices architecture.
  • Establish continuous integration and continuous deliver pipeline.
  • Performance tuning of data pipelines.
  • Monitor data pipelines for accuracy, missing data, enhancements, changes, and billing volumes to ensure all data is captured and processed accurately and when needed.
  • Reconcile data issues and alerts between various systems, finding opportunities to innovate and drive improvements.
  • Develop and maintain complex data ingestion and transformations for data originating from multiple data sources (structured/unstructured).
  • Design and implement dimension data modeling (Star Schema, Snowflake and Galaxy) for on-premise and cloud data warehouse infrastructure.
  • Implement the strategy, design, execution, system configuration, and operations of the data platform that help in data-driven decision making.
  • Lead the evaluation, implementation, and deployment of emerging tools and technologies.
  • Work with cross-functional stakeholders in defining and documenting requirements for building high-quality and impactful data products.
  • Create operational reports using visualization/business intelligence tools.

What You’ll Need:

  • Bachelor’s degree in Electrical Engineering, Computer Science, Computer Engineering, or related field and 5 years of experience;
  • Will accept a Master’s degree and 3 years of experience.
  • Experience must include 3 years with: implement the strategy, design, execution, system configuration, and operations of the data platform that help in data-driven decision making;
  • Building scalable analytics solution, including data processing, storage, and serving large-scale data through batch and stream;
  • MPP database (e.g., AWS Redshift, Exasol) big data platforms, open-source table formats like Delta Lake, Hudi, lceberg;
  • Building and maintaining infrastructure using AWS;
  • Orchestration tools like Airflow;
  • Cloud native DevOps platform using Gitlab, AWS CloudFormation, Terraform;
  • Dimensional data modeling (Star, Snowflake, Galaxy chema) and ER modeling;
  • Python;
  • SQL, data warehouse, performance tuning, and data pipelines;
  • AWS Glue, Pyspark;
  • AWS ecosystem and services including AWS S3, ECS, EC2, EMR, Athena, Lambda, Glue;
  • RESTful API’s;
  • NoSQL database like Cassandra;
  • Implement federated logical data platform;
  • Design and implement data pipelines using modern data engineering approaches and tools:
  • Spark, PySpark, Docker, cloud native DWH (e.g., Snowflake, Exasol, Redshift), AWS Glue, S3, Athena, Lambda, Postgres (AWS RDS Aurora); and
  • CI/CD platforms.
  • The position is eligible for the Employee Referral Program.

Chewy is committed to equal opportunity. We value and embrace diversity and inclusion of all Team Members. If you have a disability under the Americans with Disabilities Act or similar law, and you need an accommodation during the application process or to perform these job requirements, or if you need a religious accommodation, please contact CAAR@chewy.com.

 

If you have a question regarding your application, please contact HR@chewy.com.

 

To access Chewy's Customer Privacy Policy, please click here. To access Chewy's California CPRA Job Applicant Privacy Policy, please click here.