Data Engineer (AWS + Spark + Python / PySpark)

Candidates with less than 10 years of total experience would not be considered Hire Type Permanent Full Time OR Contract (6 Months+) Location Los Angeles, CA (Remote till Covid 19) Mandatory Skills AWS, Redshift, Python and PySpark, Spark SQL, NoSQL Experience on AWS and its web service offering S3, Redshift, EC2, EMR, Lambda, CloudWatch, RDS, Step functions, Spark streaming etc. Good knowledge of Configuring and working on Multi node clusters and distributed data processing framework Spark Hands on 3 years of experience with EMR Apache Spark Hadoop technologies Experience with must have Linux, Python and PySpark, Spark SQL Experience in working with large volumes of data Tera-bytes, analyze the data structures Experience in designing scalable data pipelines, complex event processing, analytics components using big data technology Spark, Python, Scala, PySpark Expert in SQL, NoSQL database Experience in process orchestration tools Apache Airflow, Apache NiFi etc. Hands on knowledge of design, development and enhancement of Data Lakes, constantly evolve with emerging tools and technologies. – Soumen Mondal

Related Post

Software EngineerSoftware Engineer

Job Description: Were looking for engineers who are hardworking, inquisitive, and responsible individuals to help deliver the best digital web experience to our customers. The ideal candidate is deeply entrenched