
Location: Bangalore/Pune/Hyderabad/Nagpur
4-5 years of overall experience in software development.
- Experience on Hadoop (Apache/Cloudera/Hortonworks) and/or other Map Reduce Platforms
- Experience on Hive, Pig, Sqoop, Flume and/or Mahout
- Experience on NO-SQL – HBase, Cassandra, MongoDB
- Hands on experience with Spark development, Knowledge of Storm, Kafka, Scala
- Good knowledge of Java
- Good background of Configuration Management/Ticketing systems like Maven/Ant/JIRA etc.
- Knowledge around any Data Integration and/or EDW tools is plus
- Good to have knowledge of using Python/Perl/Shell
Please note - Hbase hive and spark are must.

Similar jobs
We are seeking a skilled Java Developer with hands-on experience in Java and Spark to build scalable data processing solutions. You'll contribute to high-performance data pipelines and analytics platforms in a dynamic Agile environment.
Key Responsibilities
- Design and develop Java applications integrated with Apache Spark for ETL processes, data transformations, and analytics.
- Build and optimize Spark jobs (Spark SQL, DataFrames, Streaming) for large-scale data processing.
- Collaborate with data engineers and analysts to implement robust data workflows.
- Write clean, maintainable Java code following best practices (Spring Boot, Microservices preferred).
- Perform code reviews, unit testing, and contribute to CI/CD pipelines.
- Troubleshoot and optimize Spark performance for production workloads.
- Document technical solutions and mentor junior developers.
Required Skills & Qualifications
- 4-7 years of hands-on Java development experience.
- Strong expertise in Apache Spark (Spark Core, Spark SQL, PySpark basics).
- Proficiency in Java 8/11+ with multithreading and collections frameworks.
- Experience with data processing (ETL, data pipelines, big data).
- Familiarity with build tools (Maven/Gradle) and version control (Git).
- Strong problem-solving skills and Bangalore location availability.
- Excellent communication skills for cross-team collaboration.
Good to Have
- Experience with Snowflake for cloud data warehousing.
- Knowledge of DBT (Data Build Tool) for analytics engineering.
- Python scripting for data manipulation and automation.
- Exposure to AWS/GCP/Azure cloud platforms.
- Familiarity with Kafka, Airflow, or containerization (Docker/Kubernetes).
Requirements
- 3+ years work experience with production-grade python. Contribution to open source repos is preferred
- Experience writing concurrent and distributed programs, AWS lambda, Kubernetes, Docker, Spark is preferred.
- Experience with one relational & 1 non-relational DB is preferred
- Prior work in the ML domain will be a big boost
What You’ll Do
- Help realize the product vision: Production-ready machine learning models with monitoring within moments, not months.
- Help companies deploy their machine learning models at scale across a wide range of use-cases and sectors.
- Build integrations with other platforms to make it easy for our customers to use our product without changing their workflow.
- Write maintainable, scalable performant python code
- Building gRPC, rest API servers
- Working with Thrift, Protobufs, etc.
• Participation in the requirements analysis, design, development and testing of applications.
• The candidate is expected to write code himself/herself.
• The candidate is expected to write high level code, code review, unit testing and deployment.
• Practical application of design principles with a focus on the user experience, usability, template
designs, cross browser issues and client server concepts.
• Contributes to the development of project estimates, scheduling, and deliverables.
• Works closely with QA team to determine testing requirements to ensure full coverage and best
quality of product.
• There is also the opportunity to mentor and guide junior team members in excelling their jobs.
Job Specifications
• BE/B. Tech. Computer Science or MCA from a reputed University.
• 6+ Years of experience in software development, with emphasis on JAVA/J2EE Server side
programming.
• Hands on experience in Core Java, Multithreading, RMI, Socket programing, JDBC, NIO,
webservices and Design patterns.
• Should have Knowledge of distributed system, distributed caching, messaging frameworks, ESB
etc.
• Knowledge of Linux operating system and PostgreSQL/MySQL/MongoDB/Cassandra database is
essential.
• Additionally, knowledge of HBase, Hadoop and Hive are desirable.
• Familiarity with message queue systems and AMQP and Kafka is desirable.
• Should have experience as a participant in Agile methodologies.
• Should have excellent written and verbal communication skills and presentation skills.
• This is not a Fullstack requirement, we are purely looking out for Backend resources








