Send me more jobs like this

Weekend Drive on 30th Nov_Bigdata/Data Engineer/Bigdata Admin_Bangalore Whitefield_Karthika

5 - 10 years
Posted: 2019-11-27

Industry
IT/Computers - Software
Function
IT
Role
Software Engineer/ Programmer
Posted On
27th Nov 2019
Job Ref code
JPC028443
Job Description
Job Description :

Job Description: PFB the JD

Data Engineer: Big Data ETL Developers: Proficient understanding of distributed computing principles, Management of Hadoop cluster, with all included services
• Proficiency with Hive-QL and scripting language
• Knowledge of various ETL techniques and frameworks, such as Flume
• Understanding of Hadoop, Hive and SQL
• Very Good hold on Hive Scripting Language
• Candidate should have Data Processing ability (ETL techniques) using hive scripting experience.
• Candidate MUST NOT be limited to Data Migration capability from legacy DB to Hadoop Cluster
• Candidate must be able to Analyze, Develop and Debug the Hive Scripts on his own.
• Proficient with Partitioning, Analytical aggregation and dealing with large tables.
• Understanding of MySQL is a good to have criteria

Data Tester: BigData Testers
• Experience in Hadoop, MapReduce, Hive with hands-on experience in supporting Hadoop applications and testing.
• •Experience in Hive queries, Linux/Unix – Linux is preferable, hands on experience in Spark 20
• Experience in testing ETL,Reports, BI and reporting in a high volume environment.
• Experience validating data mapping, selection criteria, aggregations, sorting, lookups, transformations, data loads
• Hands-on experience in generating test data and test related procedures, packages, triggers

Maintenance Engineer: Bigdata Administrator

• Deploying a hadoop cluster, maintaining a hadoop cluster, adding and removing nodes using cluster monitoring tools like Ganglia Nagios or Cloudera Manager, configuring the NameNode high availability and keeping a track of all the running hadoop jobs.
• Implementing, managing and administering the overall hadoop infrastructure. Ensure that the hadoop cluster is up and running all the time. Monitoring the cluster connectivity and performance.
• Responsible for capacity planning and estimating the requirements for lowering or increasing the capacity of the hadoop cluster. Responsible for deciding the of the hadoop cluster based on the data to be stored in HDFS.
• Manage and review Hadoop log files. Backup and recovery tasks. Resource and security management. Troubleshooting application errors and ensuring that they do not occur again.
Key Skills: PFB the JD

Data Engineer: Big Data ETL Developers: Proficient understanding of distributed computing principles, Management of Hadoop cluster, with all included services
• Proficiency with Hive-QL and scripting language
• Knowledge of various ETL techniques and frameworks, such as Flume
• Understanding of Hadoop, Hive and SQL
• Very Good hold on Hive Scripting Language
• Candidate should have Data Processing ability (ETL techniques) using hive scripting experience.
• Candidate MUST NOT be limited to Data Migration capability from legacy DB to Hadoop Cluster
• Candidate must be able to Analyze, Develop and Debug the Hive Scripts on his own.
• Proficient with Partitioning, Analytical aggregation and dealing with large tables.
• Understanding of MySQL is a good to have criteria

Data Tester: BigData Testers
• Experience in Hadoop, MapReduce, Hive with hands-on experience in supporting Hadoop applications and testing.
• •Experience in Hive queries, Linux/Unix – Linux is preferable, hands on experience in Spark 20
• Experience in testing ETL,Reports, BI and reporting in a high volume environment.
• Experience validating data mapping, selection criteria, aggregations, sorting, lookups, transformations, data loads
• Hands-on experience in generating test data and test related procedures, packages, triggers

Maintenance Engineer: Bigdata Administrator

• Deploying a hadoop cluster, maintaining a hadoop cluster, adding and removing nodes using cluster monitoring tools like Ganglia Nagios or Cloudera Manager, configuring the NameNode high availability and keeping a track of all the running hadoop jobs.
• Implementing, managing and administering the overall hadoop infrastructure. Ensure that the hadoop cluster is up and running all the time. Monitoring the cluster connectivity and performance.
• Responsible for capacity planning and estimating the requirements for lowering or increasing the capacity of the hadoop cluster. Responsible for deciding the of the hadoop cluster based on the data to be stored in HDFS.
• Manage and review Hadoop log files. Backup and recovery tasks. Resource and security management. Troubleshooting application errors and ensuring that they do not occur again.

Similar Jobs
View All Similar Jobs


Walkin for you