The recruiter has not been active on this job recently. You may apply but please expect a delayed response.
Bengaluru (Bangalore)
5 - 10 yrs
₹10L - ₹20L / yr
Data Warehouse (DWH)
Apache Hive
ETL
DWH Cloud
Hadoop
+3 more
candidate will be responsible for all aspects of data acquisition, data transformation, and analytics scheduling and operationalization to drive high-visibility, cross-division outcomes. Expected deliverables will include the development of Big Data ELT jobs using a mix of technologies, stitching together complex and seemingly unrelated data sets for mass consumption, and automating and scaling analytics into the GRAND's Data Lake.
Key Responsibilities :
- Create a GRAND Data Lake and Warehouse which pools all the data from different regions and stores of GRAND in GCC
- Ensure Source Data Quality Measurement, enrichment and reporting of Data Quality
- Manage All ETL and Data Model Update Routines
- Integrate new data sources into DWH
- Manage DWH Cloud (AWS/AZURE/Google) and Infrastructure
Skills Needed :
- Very strong in SQL. Demonstrated experience with RDBMS, Unix Shell scripting preferred (e.g., SQL, Postgres, Mongo DB etc)
- Experience with UNIX and comfortable working with the shell (bash or KRON preferred)
- Good understanding of Data warehousing concepts.
Big data systems : Hadoop, NoSQL, HBase, HDFS, MapReduce
- Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
- Working with data delivery teams to set up new Hadoop users. This job includes setting up Linux users, setting up and testing HDFS, Hive, Pig and MapReduce access for the new users.
- Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, and other tools.
- Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
- Screen Hadoop cluster job performances and capacity planning
- Monitor Hadoop cluster connectivity and security
- File system management and monitoring.
- HDFS support and maintenance.
- Collaborating with application teams to install operating system and
- Hadoop updates, patches, version upgrades when required.
- Defines, develops, documents and maintains Hive based ETL mappings and scripts
Read more
Did not find a job you were looking for?
Search for relevant jobs from 10000+ companies such as Google, Amazon & Uber actively hiring on Cutshort.
Capace Software, founded in 2013 and now 50+ strong, is a Private listed Digital Transformation and Product Engineering outfit. We design, build, and sustain enterprise-grade software to fulfil our clients' most ambitious business aspirations or resolve their most complex business problems using cutting-edge technology. We strive to deliver value accretive services to our clients over a long period of time as an extension of their own team.
TVL Media stands as a testament to the power of vision, innovation, and hard work. Our story is one of transformation and growth, fueled by a passion for helping businesses succeed in an ever-evolving digital world. As we look to the future, we remain committed to pushing the boundaries of what’s possible and empowering our clients to achieve their fullest potential.
Peak Hire Solutions is a leading Recruitment Firm that provides our clients with innovative IT / Non-IT Recruitment Solutions. We pride ourselves on our creativity, quality, and professionalism. Join our team and be a part of shaping the future of Recruitment.