Weekend Drive on 30th Nov_Bigdata/Data Engineer/Bigdata Admin_Bangalore Whitefield_Karthika

Weekend Drive on 30th Nov_Bigdata/Data Engineer/Bigdata Admin_Bangalore Whitefield_Karthika

Arminus Software Private Limited
Bengaluru / Bangalore
5 - 10 Years
Not Specified

Job Description


Job Description :
Job Description: PFB the JD
Data Engineer: Big Data ETL Developers: Proficient understanding of distributed computing principles, Management of Hadoop cluster, with all included services
• Proficiency with Hive-QL and scripting language
• Knowledge of various ETL techniques and frameworks, such as Flume
• Understanding of Hadoop, Hive and SQL
• Very Good hold on Hive Scripting Language
• Candidate should have Data Processing ability (ETL techniques) using hive scripting experience.
• Candidate MUST NOT be limited to Data Migration capability from legacy DB to Hadoop Cluster
• Candidate must be able to Analyze, Develop and Debug the Hive Scripts on his own.
• Proficient with Partitioning, Analytical aggregation and dealing with large tables.
• Understanding of MySQL is a good to have criteria
Data Tester: BigData Testers
• Experience in Hadoop, MapReduce, Hive with hands-on experience in supporting Hadoop applications and testing.
• •Experience in Hive queries, Linux/Unix – Linux is preferable, hands on experience in Spark 20
• Experience in testing ETL,Reports, BI and reporting in a high volume environment.
• Experience validating data mapping, selection criteria, aggregations, sorting, lookups, transformations, data loads
• Hands-on experience in generating test data and test related procedures, packages, triggers
Maintenance Engineer: Bigdata Administrator
• Deploying a hadoop cluster, maintaining a hadoop cluster, adding and removing nodes using cluster monitoring tools like Ganglia Nagios or Cloudera Manager, configuring the NameNode high availability and keeping a track of all the running hadoop jobs.
• Implementing, managing and administering the overall hadoop infrastructure. Ensure that the hadoop cluster is up and running all the time. Monitoring the cluster connectivity and performance.
• Responsible for capacity planning and estimating the requirements for lowering or increasing the capacity of the hadoop cluster. Responsible for deciding the of the hadoop cluster based on the data to be stored in HDFS.
• Manage and review Hadoop log files. Backup and recovery tasks. Resource and security management. Troubleshooting application errors and ensuring that they do not occur again.
Key Skills: PFB the JD
Data Engineer: Big Data ETL Developers: Proficient understanding of distributed computing principles, Management of Hadoop cluster, with all included services
• Proficiency with Hive-QL and scripting language
• Knowledge of various ETL techniques and frameworks, such as Flume
• Understanding of Hadoop, Hive and SQL
• Very Good hold on Hive Scripting Language
• Candidate should have Data Processing ability (ETL techniques) using hive scripting experience.
• Candidate MUST NOT be limited to Data Migration capability from legacy DB to Hadoop Cluster
• Candidate must be able to Analyze, Develop and Debug the Hive Scripts on his own.
• Proficient with Partitioning, Analytical aggregation and dealing with large tables.
• Understanding of MySQL is a good to have criteria
Data Tester: BigData Testers
• Experience in Hadoop, MapReduce, Hive with hands-on experience in supporting Hadoop applications and testing.
• •Experience in Hive queries, Linux/Unix – Linux is preferable, hands on experience in Spark 20
• Experience in testing ETL,Reports, BI and reporting in a high volume environment.
• Experience validating data mapping, selection criteria, aggregations, sorting, lookups, transformations, data loads
• Hands-on experience in generating test data and test related procedures, packages, triggers
Maintenance Engineer: Bigdata Administrator
• Deploying a hadoop cluster, maintaining a hadoop cluster, adding and removing nodes using cluster monitoring tools like Ganglia Nagios or Cloudera Manager, configuring the NameNode high availability and keeping a track of all the running hadoop jobs.
• Implementing, managing and administering the overall hadoop infrastructure. Ensure that the hadoop cluster is up and running all the time. Monitoring the cluster connectivity and performance.
• Responsible for capacity planning and estimating the requirements for lowering or increasing the capacity of the hadoop cluster. Responsible for deciding the of the hadoop cluster based on the data to be stored in HDFS.
• Manage and review Hadoop log files. Backup and recovery tasks. Resource and security management. Troubleshooting application errors and ensuring that they do not occur again.

Job Details

Similar Jobs

People Also Considered

Data Not Available

Career Advice to Find Better

Simple body text this will replace with orginal content