- Marietta, GA, US
- $30 /hr
- Available now
Development of Big Data projects using Hadoop, HDFS and data lakes.
Design, build, and support cloud and open source systems to process geospatial data assets via an API-basedmore...Development of Big Data projects using Hadoop, HDFS and data lakes.
Design, build, and support cloud and open source systems to process geospatial data assets via an API-based platform.
Processed high volumes of streaming geospatial data from IoT sensors with Kafka, Storm and Spark streaming, showing devices status and location on the digital maps application.
Importing and exporting data between HDFS and RDBMS using Sqoop.
Pig Latin scripts and Hive (HiveQL) to perform data transformations and incremental loads.
Using Flume to handle streaming data and loaded the data into Hadoop clusters.
Extensive knowledge of NoSQL databases such as HBase, MongoDB, and Cassandra.
Multi Clustered environment, deployment and configuration using Cloudera and HortonWorks platforms.
Creation of RDD’s, Datasets and Dataframes for the input data and performed transformations using Spark Python.
Development and support of MapReduce jobs in Scala for data cleaning and preprocessing.
Big Data Analytics, ETL, Data Analysis and Visualization using Cloudera and HortonWorks platforms.
Good experience in working with cloud environment with Amazon Web Services (AWS) EMR, IAM, Lambda, API Gateway, Cognito, CloudFormation, CloudWatch, DynamoDB, Data Pipeline, EC2, and S3.
Hadoop Engineer and AWS Solutions Architect