Data Engineer (AWS + Spark + Python / PySpark)

Candidates with less than 10 years of total experience would not be considered Hire Type Permanent Full Time OR Contract (6 Months+) Location Los Angeles, CA (Remote till Covid 19) Mandatory Skills AWS, Redshift, Python and PySpark, Spark SQL, NoSQL Experience on AWS and its web service offering S3, Redshift, EC2, EMR, Lambda, CloudWatch, RDS, Step functions, Spark streaming etc. Good knowledge of Configuring and working on Multi node clusters and distributed data processing framework Spark Hands on 3 years of experience with EMR Apache Spark Hadoop technologies Experience with must have Linux, Python and PySpark, Spark SQL Experience in working with large volumes of data Tera-bytes, analyze the data structures Experience in designing scalable data pipelines, complex event processing, analytics components using big data technology Spark, Python, Scala, PySpark Expert in SQL, NoSQL database Experience in process orchestration tools Apache Airflow, Apache NiFi etc. Hands on knowledge of design, development and enhancement of Data Lakes, constantly evolve with emerging tools and technologies. – Soumen Mondal

Related Post