Global hospitality company is looking for an experienced PySpark Data Engineer for their Midtown NYC headquarters.
Responsibilities:
- Lead the in-house development of ETL/ELT data pipelines and data science projects.
- Create and own data products, such as a recommendation engine or predictive model.
- Provide key insights and analyses to senior stakeholders and executives.
- Work with the Data Engineering team to design and develop data models.
Qualifications And Requirements
- 5+ years of relevant professional experience in building AWS big data pipelines using Apache Spark.
- Strong hands-on experience with programming in Python.
- Expertise in SQL and analytical data modeling.
- Hands-on experience in pipeline orchestration tools, like Apache Airflow.
- Experience in PostgreSQL, Redshift, S3, AWS Lambda, Kinesis and Athena.
- Previous work with cloud-based BI tools (such as Looker, and Quicksight) is a plus.
- Strong organizational, problem-solving, and communications skills (must be able to do basic technical writing)