Matomo Code

Computer Programmer

Date Posted: 03-15-2018, Job Code: CP0318001, Job Timing: Full – time

Job Roles and Responsibilities:

  • Work with the client, business users and other stake holders to gather requirements for Data Pipeline and analytics reporting which helps executive teams in making decisions.
  • Responsible for designing and coding analytics system which helps executive teams in making decisions in bidding, sales and other areas.
  • Responsibilities include analysis of various applications, designing of the enterprise applications.
  • Based on the requirements gathering do the impact analysis.
  • Estimate the efforts to complete the requirements.
  • Prepare the design document for the Data Pipeline and analytics reporting based on the client requirements in Big data technology.
  • Develop the Data Pipeline and Analytics reports using Big Data technologies.
  • Do performance tuning in the analytics programs for better performance and faster execution.
  • Setup the Hadoop cluster for development and testing.
  • Test the huge data in hundreds of nodes in Cloud environment.
  • Involve in daily scrum meeting as part of Scrum methodologies followed in the project.
  • Responsible for building scalable distributed data solutions using Hadoop technologies.
  • Work on Hadoop cluster (CDH 5.x) with several nodes.
  • Write Spark RDD/Data frame/SQL to power data for extraction, transformation and aggregation from multiple file formats including JSON, CSV & other compressed file formats.
  • Experienced in handling large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Effective & efficient Joins, Transformations and other during ingestion process itself.
  • Use Sqoop for importing and exporting data from RDBMS into HDFS and Hive.
  • Work on avro, parquet file formats by using snappy compression.
  • Automate jobs with Oozie and schedule them with Autosys.
  • Develop Spark RDD and Data frames and SQL scripts for ETL transformations.
  • Experience in developing customized functions in python to extend Hive functionality.
  • Experience in building and optimizing existing algorithms in Hadoop using Spark-SQL, Data Frames and Pair RDD’s.

Education Requirement:

  • A Bachelor’s degree equivalent in Computer Science/ Applications, Information Technology, Engineering Management or relevant field.

 FOR FURTHER DETAILS, PLEASE CONTACT:

HR Manager
eGrove Systems Corporation
603 791 4890
E-mail: usitcareers@egrovesys.com

ALWAYS ON THE LOOK OUT

We’re always looking for the right people to join our team. Even if you don’t find suitable roles listed on our website at this time, please send us your details to be considered for other positions in future.

CALL US ON 603 791 4890 (or) email usitcareers@egrovesys.com with your details

APPLY NOW:

If you like us to email you when a position is available, kindly fill out the below form.

Job Alert
reCAPTCHA
Get Quote
close slider
Contact Us
Your IP Address is : 35.175.201.245
Confirm that you are not a Bot