11+ HDFS Jobs in Pune | HDFS Job openings in Pune
Apply to 11+ HDFS Jobs in Pune on CutShort.io. Explore the latest HDFS Job opportunities across top companies like Google, Amazon & Adobe.
Job Details
- Job Title: Lead Software Engineer - Java, Python, API Development
- Industry: Global digital transformation solutions provider
- Domain - Information technology (IT)
- Experience Required: 8-10 years
- Employment Type: Full Time
- Job Location: Pune & Trivandrum/ Thiruvananthapuram
- CTC Range: Best in Industry
Job Description
Job Summary
We are seeking a Lead Software Engineer with strong hands-on expertise in Java and Python to design, build, and optimize scalable backend applications and APIs. The ideal candidate will bring deep experience in cloud technologies, large-scale data processing, and leading the design of high-performance, reliable backend systems.
Key Responsibilities
- Design, develop, and maintain backend services and APIs using Java and Python
- Build and optimize Java-based APIs for large-scale data processing
- Ensure high performance, scalability, and reliability of backend systems
- Architect and manage backend services on cloud platforms (AWS, GCP, or Azure)
- Collaborate with cross-functional teams to deliver production-ready solutions
- Lead technical design discussions and guide best practices
Requirements
- 8+ years of experience in backend software development
- Strong proficiency in Java and Python
- Proven experience building scalable APIs and data-driven applications
- Hands-on experience with cloud services and distributed systems
- Solid understanding of databases, microservices, and API performance optimization
Nice to Have
- Experience with Spring Boot, Flask, or FastAPI
- Familiarity with Docker, Kubernetes, and CI/CD pipelines
- Exposure to Kafka, Spark, or other big data tools
Skills
Java, Python, API Development, Data Processing, AWS Backend
Skills: Java, API development, Data Processing, AWS backend, Python,
Must-Haves
Java (8+ years), Python (8+ years), API Development (8+ years), Cloud Services (AWS/GCP/Azure), Database & Microservices
8+ years of experience in backend software development
Strong proficiency in Java and Python
Proven experience building scalable APIs and data-driven applications
Hands-on experience with cloud services and distributed systems
Solid understanding of databases, microservices, and API performance optimization
Mandatory Skills: Java API AND AWS
******
Notice period - 0 to 15 days only
Job stability is mandatory
Location: Pune, Trivandrum
Job Role: We are seeking a skilled Java Developer to contribute to the development and enhancement renowned banking application, which supports automatic reconciliation and unified data reporting for their clients. This role involves working on high-impact enhancements, data pipeline integration, and platform modernization. The ideal candidate will be a quick learner, self-motivated, and able to ramp up quickly in a fast-paced environment.
Key Responsibilities:
Design, develop, and maintain Java-based applications using Java 17 and Spring Boot.
Implement and manage message routing using Apache Camel.
Develop and monitor data pipelines using Kafka.
Support and enhance existing cloud-native applications.
Work with OpenShift Container Platform (OCP 4) for container orchestration and deployments.
Utilize Jenkins for CI/CD pipeline automation and management.
Collaborate with cross-functional teams to integrate multiple data sources into a unified reporting platform.
Participate in code reviews, unit testing, and performance tuning.
Troubleshoot and resolve production issues in collaboration with operations teams.
Document development processes and system configurations.
Required Skills:
Strong proficiency in Java 17 and Spring Boot frameworks.
Hands-on experience with Apache Camel for message routing and transformation.
Solid experience in Kafka development and monitoring tools.
Good understanding of cloud pipeline architectures and deployment strategies.
Experience working with OpenShift (OCP 4).
Familiarity with Jenkins for CI/CD and automated deployments.
Understanding of cloud deployment platforms (AWS, Azure, or GCP preferred).
Strong analytical and debugging skills.
Ability to learn quickly and adapt to evolving project requirements.
Nice to Have:
Experience in financial services or transaction reporting platforms.
Familiarity with microservices architecture and containerization best practices.
Knowledge of monitoring tools (e.g., Prometheus, Grafana).
2. Design software and make technology choices across the stack (from data storage to application to front-end)
3. Understand a range of tier-1 systems/services that power our product to make scalable changes to critical path code
4. Own the design and delivery of an integral piece of a tier-1 system or application
5. Work closely with product managers, UX designers, and end users and integrate software components into a fully functional system
6. Work on the management and execution of project plans and delivery commitments
7. Take ownership of product/feature end-to-end for all phases from the development to the production
8. Ensure the developed features are scalable and highly available with no quality concerns
9. Work closely with senior engineers for refining and implementation
10. Manage and execute project plans and delivery commitments
11. Create and execute appropriate quality plans, project plans, test strategies, and processes for development activities in concert with business and project management efforts
Requirements
• Extensive and expert programming experience in at least one general programming language (e. g.
Java, C, C++) & tech stack to write maintainable, scalable, unit-tested code.
• Experience with multi-threading and concurrency programming.
• Extensive experience in object oriented design skills, knowledge of design patterns, and a huge passion
and ability to design intuitive modules and class-level interfaces.
• Excellent coding skills - should be able to convert design into code fluently.
• Knowledge of Test Driven Development.
• Good understanding of databases (e. g. MySQL) and NoSQL (e. g. HBase, Elasticsearch, Aerospike etc).
• Strong desire to solve complex and interesting real world problems.
• Experience with full life cycle development in any programming language on a Linux platform.
• Go-getter attitude that reflects in energy and intent behind assigned tasks.
• Worked in a startup-like environment with high levels of ownership and commitment.
• BTech, MTech or Ph. D. in Computer Science or related technical discipline (or equivalent).
• Experience in building highly scalable business applications, which involve implementing large complex
business flows and dealing with huge amounts of data.
• 3+ years of experience in the art of writing code and solving problems on a large scale.
• Open communicator who shares thoughts and opinions frequently, listens intently, and takes
constructive feedback.
Location: Bangalore/Pune/Hyderabad/Nagpur
4-5 years of overall experience in software development.
- Experience on Hadoop (Apache/Cloudera/Hortonworks) and/or other Map Reduce Platforms
- Experience on Hive, Pig, Sqoop, Flume and/or Mahout
- Experience on NO-SQL – HBase, Cassandra, MongoDB
- Hands on experience with Spark development, Knowledge of Storm, Kafka, Scala
- Good knowledge of Java
- Good background of Configuration Management/Ticketing systems like Maven/Ant/JIRA etc.
- Knowledge around any Data Integration and/or EDW tools is plus
- Good to have knowledge of using Python/Perl/Shell
Please note - Hbase hive and spark are must.
Desired Candidate Profile
- A team focus with strong collaboration and communication skills
- Exceptional ability to quickly grasp high-level business goals, derive requirements, and translate them into effective technical solutions
- Exceptional object-oriented thinking, design and programming skills (Java 8 or 11)
- Expertise with the following technologies : Data Structures, Design Patterns ,Code Versioning Tools(Github/bitbucket/..), XML, JSON, Spring Batch Restful, Spring Cloud, Grafana(Knowledge/Experience), Kafka, Spring Boot, Microservices, DB/NoSQL, Docker, Kubernetes, AWS/GCP, Architecture design (Patterns) Agile, JIRA.
- Penchant toward self-motivation and continuous improvement; these words should describe you: dedicated, energetic, curious, conscientious, and flexible.
- 3-8+ years of experience programming in a backend language (Java / Python), with a good understanding of troubleshooting errors.
- 5+ years of experience in Confluent Kafka / 3+ years of experience in Confluent Kafka
- Cloud Kafka, Control Central, Rest Proxy, HA Proxy, Confluent Kafka Connect, Confluent Kafka Security features
- 2 - 6 years of software development experience
- Good grasp on programming fundamentals including OOP, Design Patterns and Data Structures
- Excellent analytical, logical and problem-solving skills
- Software Development Engineer
- Good understanding of complexities involved in designing/developing large scale systems
- Strong system design skills
- Experience in technologies like Elasticsearch, Redis, Kafka etc
- Good knowledge of relational and NoSQL databases
- Familiarity with common machine learning algorithms. In-depth knowledge is a plus
- Experience of working with big data technologies like Hadoop, Spark, Hive is a big plus
- Ability to understand business requirements and take ownership of the work
- Exhibit passion and enthusiasm for building and maintaining large scale platforms
- You will be responsible for design, development and testing of Products
- Contributing in all phases of the development lifecycle
- Writing well designed, testable, efficient code
- Ensure designs are in compliance with specifications
- Prepare and produce releases of software components
- Support continuous improvement by investigating alternatives and technologies and presenting these for architectural review
- Some of the technologies you will be working on: Core Java, Solr, Hadoop, Spark, Elastic search, Clustering, Text Mining, NLP, Mahout and Lucene etc.



