DESCRIPTION The Amazon ShipTech Analytics organization is looking for a smart and ambitious individual to support the development of analytical tools and data processing solutions for the Amazon transportation business. This role will work within a multi-functional BI and data engineering team with a focus on AMZL (Amazon Logistics) and other rapidly growing delivery programs where Amazon is innovating to find new ways to deliver to our customers.
The ideal candidate thrives working with large volumes of data, enjoys the challenge of highly complex technical contexts, and is passionate about data and analytics. The candidate is an expert within data modeling, ETL design and cloud/big-data technologies and passionately partners with the business to identify strategic opportunities where improvements in data infrastructure creates large-scale business impact. The candidate should be a self-starter; comfortable with ambiguity, able to think big, and enjoy working in a fast-paced and global team. It's a big ask, and we're excited to talk to those up to the challenge!
Key job responsibilities
· Build end-to-end data pipelines to ingest and transform data from different types of data sources and systems; from traditional ETL pipelines to event data streams
· Utilize data from disparate data sources to build meaningful datasets for analytics, reporting and ML use cases
· Evaluate and implement various big-data technologies and solutions (e.g. Redshift, Hive/EMR, Spark, SNS, SQS, Kinesis) to optimize processing of extremely large datasets
· Understand and analyze business processes, logical data models and relational database implementations
· Write high performing and optimized SQL queries
· Execute research projects, and generate practical results and recommendations
· Design and implement automated data processing solutions and data quality controls
We are open to hiring candidates to work out of one of the following locations:
Bellevue, WA, USA BASIC QUALIFICATIONS - Bachelor's degree
- 5+ years of data engineering experience
- Experience in at least one modern scripting or programming language, such as Python, Java, Scala, or NodeJS
- Experience with data modeling, warehousing and building ETL pipelines
- Experience with big data technologies such as: Hadoop, Hive, Spark, EMR
- Experience with SQL