Hadoop Developer Jobs - Expinfo Inc in Plano, Texas, TX

Hadoop Developer

Job Code: 2323
Company Name: Expinfo Inc
Job Location: Plano, Texas
Zip Code: 75093
Job Category/Title: Hadoop Developer
Employment Category: Full-time
Position Type:
Salary: Negotiable
Pay Rate: Negotiable
Travel Required: No
Interview Type:
Required IT Skills: HBase, Pig, Hive, Sqoop, Flume, Oozie, Python, Perl
Job Description: Responsibilities:
- Understanding the requirements of input to output transformations.
- Hadoop developers spend lot of time in cleaning data as per business requirements using streaming API’s or user defined functions.
- Install, configure and maintain enterprise hadoop environment.
- Loading data from different datasets and deciding on which file format is efficient for a task. Hadoop developers source large volumes of data from diverse data platforms into Hadoop platform.
- Defining Hadoop Job Flows.
- Build distributed, reliable and scalable data pipelines to ingest and process data in real-time. Hadoop developer deals with fetching impression streams, transaction behaviours, clickstream data and other unstructured data.
- Managing Hadoop jobs using scheduler.
- Reviewing and managing hadoop log files.
- Design and implement column family schemas of Hive and HBase within HDFS.
- Assign schemas and create Hive tables.
- Managing and deploying HBase clusters.
- Maintain the privacy and security of hadoop clusters.
- Fine tune hadoop applications for high performance and throughput.
- Troubleshoot and debug any hadoop ecosystem run time issues.
- Develop efficient pig and hive scripts with joins on datasets using various techniques.
- Assess the quality of datasets for a hadoop data lake.
- Apply different HDFS formats and structure like Parquet, Avro, etc. to speed up analytics.
- Build new hadoop clusters



Required Skills :
- Analytical and problem solving skills.
- Business acumen and domain knowledge
- The most obvious, knowledge of hadoop ecosystem and its components –HBase, Pig, Hive, Sqoop, Flume, Oozie, etc.
- Know-how on the java essentials for hadoop.
- Know-how on basic Linux administration
- Knowledge of scripting languages like Python or Perl.
- Data modelling experience with OLTP and OLAP
- Basic knowledge of popular ETL tools like Pentaho, Informatica, Talend, etc.
- Analytical and problem solving skills.
- Business acumen and domain knowledge
No. of Posts: 1
Comments/JobNotes:
Job Expiry Date: May 30, 2018
Telecommute: No
Contact Information Login To View