Experience on AWS and its web service offering S3, Redshift, EC2, EMR, Lambda, CloudWatch, RDS, Step functions, Spark streaming etc.
Good knowledge of Configuring and working on Multi node clusters and distributed data processing framework Spark.
Hands on 3 years of experience with EMR, Apache Spark and Hadoop technologies
Experience with must have Linux, Python and PySpark, Spark SQL, Experience with Java and Scala
Experience in working with large volumes of data Tera-bytes, analyze the data structures and design in Hadoop cluster effectively.
Experience in designing scalable data pipelines, complex event processing, analytics components using big data technology Spark Python Scala PySpark
Expert in redshift and NoSQL databases
Experience in process orchestration tools Apache Airflow, Apache NiFi, etc.
Hands on knowledge of Big Data Analytics, Predictive Analytics, Nifty with design, development and enhancement of Data Lakes, constantly evolve with emerging tools and technologies.
Signup Today!