2+ Apache Sqoop Jobs in Chennai | Apache Sqoop Job openings in Chennai
Apply to 2+ Apache Sqoop Jobs in Chennai on CutShort.io. Explore the latest Apache Sqoop Job opportunities across top companies like Google, Amazon & Adobe.
Apache sqoop jobs in other cities
Apache Sqoop JobsApache Sqoop Jobs in Bangalore (Bengaluru)Apache Sqoop Jobs in MumbaiBengaluru (Bangalore), Chennai
4 - 8 yrs
₹9L - ₹14L / yr
Apache Spark
Apache Kafka
Druid Database
Big Data
Apache Sqoop
+5 more
Key skill set : Apache NiFi, Kafka Connect (Confluent), Sqoop, Kylo, Spark, Druid, Presto, RESTful services, Lambda / Kappa architectures
Responsibilities :
- Build a scalable, reliable, operable and performant big data platform for both streaming and batch analytics
- Design and implement data aggregation, cleansing and transformation layers
Skills :
- Around 4+ years of hands-on experience designing and operating large data platforms
- Experience in Big data Ingestion, Transformation and stream/batch processing technologies using Apache NiFi, Apache Kafka, Kafka Connect (Confluent), Sqoop, Spark, Storm, Hive etc;
- Experience in designing and building streaming data platforms in Lambda, Kappa architectures
- Should have working experience in one of NoSQL, OLAP data stores like Druid, Cassandra, Elasticsearch, Pinot etc;
- Experience in one of data warehousing tools like RedShift, BigQuery, Azure SQL Data Warehouse
- Exposure to other Data Ingestion, Data Lake and querying frameworks like Marmaray, Kylo, Drill, Presto
- Experience in designing and consuming microservices
- Exposure to security and governance tools like Apache Ranger, Apache Atlas
- Any contributions to open source projects a plus
- Experience in performance benchmarks will be a plus
Read more
Chennai
1 - 5 yrs
₹1L - ₹6L / yr
Hadoop
Big Data
HDFS
Apache Sqoop
Apache Flume
+2 more
• Looking for Big Data Engineer with 3+ years of experience.
• Hands-on experience with MapReduce-based platforms, like Pig, Spark, Shark.
• Hands-on experience with data pipeline tools like Kafka, Storm, Spark Streaming.
• Store and query data with Sqoop, Hive, MySQL, HBase, Cassandra, MongoDB, Drill, Phoenix, and Presto.
• Hands-on experience in managing Big Data on a cluster with HDFS and MapReduce.
• Handle streaming data in real time with Kafka, Flume, Spark Streaming, Flink, and Storm.
• Experience with Azure cloud, Cognitive Services, Databricks is preferred.
Read more
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs