Cutshort logo
Apache Kafka Jobs in Bangalore (Bengaluru)

50+ Apache Kafka Jobs in Bangalore (Bengaluru) | Apache Kafka Job openings in Bangalore (Bengaluru)

Apply to 50+ Apache Kafka Jobs in Bangalore (Bengaluru) on CutShort.io. Explore the latest Apache Kafka Job opportunities across top companies like Google, Amazon & Adobe.

icon
Kanerika Software

at Kanerika Software

3 candid answers
2 recruiters
Ariba Khan
Posted by Ariba Khan
Bengaluru (Bangalore)
3 - 5 yrs
Upto ₹22L / yr (Varies
)
skill iconJava
skill iconAmazon Web Services (AWS)
Apache Kafka
skill iconSpring Boot
Microservices

About Kanerika:

Kanerika Inc. is a premier global software products and services firm that specializes in providing innovative solutions and services for data-driven enterprises. Our focus is to empower businesses to achieve their digital transformation goals and maximize their business impact through the effective use of data and AI.

We leverage cutting-edge technologies in data analytics, data governance, AI-ML, GenAI/ LLM and industry best practices to deliver custom solutions that help organizations optimize their operations, enhance customer experiences, and drive growth.


Awards and Recognitions:

Kanerika has won several awards over the years, including:

1. Best Place to Work 2023 by Great Place to Work®

2. Top 10 Most Recommended RPA Start-Ups in 2022 by RPA Today

3. NASSCOM Emerge 50 Award in 2014

4. Frost & Sullivan India 2021 Technology Innovation Award for its Kompass composable solution architecture

5. Kanerika has also been recognized for its commitment to customer privacy and data security, having achieved ISO 27701, SOC2, and GDPR compliances.


Working for us:

Kanerika is rated 4.6/5 on Glassdoor, for many good reasons. We truly value our employees' growth, well-being, and diversity, and people’s experiences bear this out. At Kanerika, we offer a host of enticing benefits that create an environment where you can thrive both personally and professionally. From our inclusive hiring practices and mandatory training on creating a safe work environment to our flexible working hours and generous parental leave, we prioritize the well-being and success of our employees.


Our commitment to professional development is evident through our mentorship programs, job training initiatives, and support for professional certifications. Additionally, our company-sponsored outings and various time-off benefits ensure a healthy work-life balance. Join us at Kanerika and become part of a vibrant and diverse community where your talents are recognized, your growth is nurtured, and your contributions make a real impact. See the benefits section below for the perks you’ll get while working for Kanerika.


About the role:

As a Senior Java Developer, you will utilize your extensive Java programming skills and expertise to design and develop robust and scalable applications. You will collaborate with cross- functional teams, provide technical leadership, and contribute to the entire software development life cycle. With your deep understanding of Java technologies and frameworks, you will ensure the delivery of high-quality solutions that meet the project requirements and adhere to coding standards.


Role Responsibilities:

  • Discuss new features and collaborate with the development and UX team, commercial product manager and product owner to get functionalities specified and implemented.
  • Agree the technical implementation with involved component owners and estimate its work effort.
  • Write great code and do code reviews for other engineers
  • Implement, test and demonstrate new product features in an agile process.
  • Develop complete sets of functionalities including the backend and frontend.
  • Create new microservices, or migrate existing services, to run on a cloud infrastructure
  • Work on further usability, performance improvements or quality assurance, including bug fixes and test automation.
  • Watch out for potential security issues and fix them, or better avoid them altogether 


Role requirements:

  • BTech computer science or equivalent
  • Java development skills with at least 3 to 5 years of experience. Knowledge of the most popular java libraries and frameworks: JPA, Spring, Kafka, etc
  • Have a degree in computer science, or a similar background, or you just have enough professional experience to blow right through all your challenges
  • Are a great communicator, analytic, goal-oriented, quality-focused, yet still agile person who likes to work with software engineers; you will interact a lot with architects, developers from other teams, component owners and system engineers
  • Have a clear overview of all layers in computer software development, including REST APIs and how to make and integrate them in our products
  • Have Java server-side development and SQL (PostgreSQL) and nice to have NOSQL (Mongo Db or Dynamo DB) database knowledge.
  • Are open to pick-up innovative technologies as needed by the team. Have or want to build experience with cloud and DevOps infrastructure (like Kubernetes, Docker, Terraform, Concourse, etc.)
  • Speak English fluently


Employee Benefits:

1. Culture:

  1. Open Door Policy: Encourages open communication and accessibility to management.
  2. Open Office Floor Plan: Fosters a collaborative and interactive work environment.
  3. Flexible Working Hours: Allows employees to have flexibility in their work schedules.
  4. Employee Referral Bonus: Rewards employees for referring qualified candidates.
  5. Appraisal Process Twice a Year: Provides regular performance evaluations and feedback.


2. Inclusivity and Diversity:

  1. Hiring practices that promote diversity: Ensures a diverse and inclusive workforce.
  2. Mandatory POSH training: Promotes a safe and respectful work environment.


3. Health Insurance and Wellness Benefits:

  1. GMC and Term Insurance: Offers medical coverage and financial protection.
  2. Health Insurance: Provides coverage for medical expenses.
  3. Disability Insurance: Offers financial support in case of disability.


4. Child Care & Parental Leave Benefits:

  1. Company-sponsored family events: Creates opportunities for employees and their       families to bond.
  2. Generous Parental Leave: Allows parents to take time off after the birth or adoption of a child.
  3. Family Medical Leave: Offers leave for employees to take care of family members' medical needs.


5. Perks and Time-Off Benefits:

  1. Company-sponsored outings: Organizes recreational activities for employees.
  2. Gratuity: Provides a monetary benefit as a token of appreciation.
  3. Provident Fund: Helps employees save for retirement.
  4. Generous PTO: Offers more than the industry standard for paid time off.
  5. Paid sick days: Allows employees to take paid time off when they are unwell.
  6. Paid holidays: Gives employees paid time off for designated holidays.
  7. Bereavement Leave: Provides time off for employees to grieve the loss of a loved one.


6. Professional Development Benefits:

  1. L&D with FLEX- Enterprise Learning Repository: Provides access to a learning repository for professional development.
  2. Mentorship Program: Offers guidance and support from experienced professionals.
  3. Job Training: Provides training to enhance job-related skills.
  4. Professional Certification Reimbursements: Assists employees in obtaining professional      certifications.
  5. Promote from Within: Encourages internal growth and advancement opportunities.
Read more
Technology Industry

Technology Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
6 - 9 yrs
₹36L - ₹48L / yr
skill iconPython
TypeScript
skill iconNodeJS (Node.js)
ReAct (Reason + Act)
skill iconReact Native
+13 more

Review Criteria:

  • Strong Software Engineer fullstack profile using NodeJS / Python and React
  • 6+ YOE in Software Development using Python OR NodeJS (For backend) & React (For frontend)
  • Must have strong experience in working on Typescript
  • Must have experience in message-based systems like Kafka, RabbitMq, Redis
  • Databases - PostgreSQL & NoSQL databases like MongoDB
  • Product Companies Only
  • Tier 1 Engineering Institutes preferred (IIT, NIT, BITS, IIIT, DTU or equivalent)

 

Preferred:

  • Experience in Fin-Tech, Payment, POS and Retail products is highly preferred
  • Experience in mentoring, coaching the team.


Role & Responsibilities:

We are currently seeking a Senior Engineer to join our Financial Services team, contributing to the design and development of scalable system.

 

The Ideal Candidate Will Be Able To-

  • Take ownership of delivering performant, scalable and high-quality cloud-based software, both frontend and backend side.
  • Mentor team members to develop in line with product requirements.
  • Collaborate with Senior Architect for design and technology choices for product development roadmap.
  • Do code reviews.


Ideal Candidate:

  • Thorough knowledge of developing cloud-based software including backend APIs and react based frontend.
  • Thorough knowledge of scalable design patterns and message-based systems such as Kafka, RabbitMq, Redis, MongoDB, ORM, SQL etc.
  • Experience with AWS services such as S3, IAM, Lambda etc.
  • Expert level coding skills in Python FastAPI/Django, NodeJs, TypeScript, ReactJs.
  • Eye for user responsive designs on the frontend.


Read more
AryuPay Technologies
Bhavana Chaudhari
Posted by Bhavana Chaudhari
Bengaluru (Bangalore)
4 - 8 yrs
₹4L - ₹9L / yr
skill iconDjango
skill iconPython
RESTful APIs
skill iconFlask
skill iconPostgreSQL
+7 more

We are seeking a highly skilled and experienced Python Developer with a strong background in fintech to join our dynamic team. The ideal candidate will have at least 7+ years of professional experience in Python development, with a proven track record of delivering high-quality software solutions in the fintech industry.

Responsibilities:

Design, build, and maintain RESTful APIs using Django and Django Rest Framework.

Integrate AI/ML models into existing applications to enhance functionality and provide data-driven insights.

Collaborate with cross-functional teams, including product managers, designers, and other developers, to define and implement new features and functionalities.

Manage deployment processes, ensuring smooth and efficient delivery of applications.

Implement and maintain payment gateway solutions to facilitate secure transactions.

Conduct code reviews, provide constructive feedback, and mentor junior members of the development team.

Stay up-to-date with emerging technologies and industry trends, and evaluate their potential impact on our products and services.

Maintain clear and comprehensive documentation for all development processes and integrations.

Requirements:

Proficiency in Python and Django/Django Rest Framework.

Experience with REST API development and integration.

Knowledge of AI/ML concepts and practical experience integrating AI/ML models.

Hands-on experience with deployment tools and processes.

Familiarity with payment gateway integration and management.

Strong understanding of database systems (SQL, PostgreSQL, MySQL).

Experience with version control systems (Git).

Strong problem-solving skills and attention to detail.

Excellent communication and teamwork skills.

Job Types: Full-time, Permanent

Work Location: In person

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Archana M
Posted by Archana M
Bengaluru (Bangalore)
8 - 12 yrs
Best in industry
Distributed Systems
skill iconLeadership
LLD
HLD
DDD
+5 more

Location: Hybrid (Bangalore)

Travel: Quarterly travel to Seattle(US)

Education: B.Tech from premium institutes only

Note: Only immediate joiners required/ 0 to 15 Days — no other applications accepted.

Role Summary

We are seeking top-tier Lead Engineers who can design, build, and deliver large-scale distributed systems with high performance, reliability, and operational excellence. The ideal candidate will be a hands-on engineer with expert system design ability, deep understanding of distributed architectures, and strong communication and leadership skills.

The Lead Engineer must be able to convert complex and ambiguous requirements into a fully engineered architecture and implementation plan covering components, data flows, infrastructure, observability, and operations.

Key Responsibilities

1. End-to-End System Architecture

  • Architect scalable, reliable, and secure systems from initial concept through production rollout.
  • Define system boundaries, components, service responsibilities, and integration points.
  • Produce high-level (HLD) and low-level design (LLD) documents.
  • Ensure designs meet performance, reliability, security, and cost objectives.
  • Make informed design trade-offs with solid technical reasoning.

2. Component & Communication Design

  • Break complex systems into independently deployable services.
  • Define APIs, communication contracts, data models, and event schemas.
  • Apply modern architecture patterns such as microservices, event-driven design, DDD, CQRS, and hexagonal architecture.
  • Ensure component clarity, maintainability, and extensibility.

3. Communication Protocol & Middleware

  • Design both sync and async communication layers: REST, RPC, gRPC, message queues, event streams (Kafka/Kinesis/Pulsar).
  • Define retry/timeout strategies, circuit breakers, rate limiting, and versioning strategies.
  • Handle backpressure, partitioning, delivery semantics (at-least/at-most/exactly once).

4. Data Architecture & Storage Strategy

  • Architect data models and storage strategies for SQL and NoSQL databases, distributed caches, blob stores, and search indexes.
  • Define sharding/partitioning, replication, consistency, indexing, backup/restore, and schema evolution strategies.
  • Design real-time and batch data processing pipelines.

5. Operational Readiness

  • Define observability (metrics, logs, traces) requirements.
  • Collaborate with DevOps to ensure deployment, monitoring, alerts, and incident management readiness.
  • Provide production support as a senior technical owner.

6. Leadership & Influence

  • Lead technical discussions, design reviews, and cross-team collaboration.
  • Mentor engineers and help elevate team practices.
  • Influence technology direction and architectural standards.

Required Qualifications

  • 10+ years of professional software engineering experience with strong backend and distributed systems background.
  • Proven track record of leading large-scale architecture and delivery of production systems.
  • Expert in system design with the ability to simplify ambiguity and craft robust solutions.
  • Strong programming experience in one or more languages (Java, Go, Python, C++).
  • Deep understanding of distributed systems, message streaming, queues, RPC/REST, and event-driven architecture.
  • Experience with cloud platforms (AWS/Azure/GCP) and container technologies (Kubernetes/Docker).
  • Strong communication, documentation, and leadership skills.

Preferred Skills

  • Experience with large-scale messaging/streaming (Kafka/Pulsar), caching, and NoSQL.
  • Experience designing for high availability, fault tolerance, and performance at scale.
  • Mentoring and leading global engineering teams.
  • Familiarity with observability tooling (Grafana, Prometheus, Jaeger). 


Read more
Ride-hailing Industry

Ride-hailing Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
5 - 7 yrs
₹42L - ₹45L / yr
DevOps
skill iconPython
Shell Scripting
Infrastructure
Terraform
+16 more

JOB DETAILS:

- Job Title: Senior Devops Engineer 2

- Industry: Ride-hailing

- Experience: 5-7 years

- Working Days: 5 days/week

- Work Mode: ONSITE

- Job Location: Bangalore

- CTC Range: Best in Industry


Required Skills: Cloud & Infrastructure Operations, Kubernetes & Container Orchestration, Monitoring, Reliability & Observability, Proficiency with Terraform, Ansible etc., Strong problem-solving skills with scripting (Python/Go/Shell)

 

Criteria:

1.   Candidate must be from a product-based or scalable app-based start-ups company with experience handling large-scale production traffic.

2.   Minimum 5 yrs of experience working as a DevOps/Infrastructure Consultant

3.   Own end-to-end infrastructure right from non-prod to prod environment including self-managed

4.   Candidate must have experience in database migration from scratch 

5.   Must have a firm hold on the container orchestration tool Kubernetes

6.   Must have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

7.   Understanding programming languages like GO/Python, and Java

8.   Working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

9.   Working experience on Cloud platform - AWS

10. Candidate should have Minimum 1.5 years stability per organization, and a clear reason for relocation.

 

Description 

Job Summary:

As a DevOps Engineer at company, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.

 

Job Responsibilities:

● Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs

● Codify our infrastructure

● Do what it takes to keep the uptime above 99.99%

● Understand the bigger picture and sail through the ambiguities

● Scale technology considering cost and observability and manage end-to-end processes

● Understand DevOps philosophy and evangelize the principles across the organization

● Strong communication and collaboration skills to break down the silos

 

Job Requirements:

● B.Tech. / B.E. degree in Computer Science or equivalent software engineering degree/experience

● Minimum 5 yrs of experience working as a DevOps/Infrastructure Consultant

● Must have a firm hold on the container orchestration tool Kubernetes

● Must have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

● Strong problem-solving skills, and ability to write scripts using any scripting language

● Understanding programming languages like GO/Python, and Java

● Comfortable working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

 

What’s there for you?

Company’s team handles everything – infra, tooling, and self-manages a bunch of databases, such as

● 150+ microservices with event-driven architecture across different tech stacks Golang/ java/ node

● More than 100,000 Request per second on our edge gateways

● ~20,000 events per second on self-managed Kafka

● 100s of TB of data on self-managed databases

● 100s of real-time continuous deployment to production

● Self-managed infra supporting

● 100% OSS

Read more
Ride-hailing Industry

Ride-hailing Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
6 - 9 yrs
₹47L - ₹50L / yr
DevOps
skill iconPython
Shell Scripting
skill iconKubernetes
Terraform
+15 more

JOB DETAILS:

- Job Title: Lead DevOps Engineer

- Industry: Ride-hailing

- Experience: 6-9 years

- Working Days: 5 days/week

- Work Mode: ONSITE

- Job Location: Bangalore

- CTC Range: Best in Industry


Required Skills: Cloud & Infrastructure Operations, Kubernetes & Container Orchestration, Monitoring, Reliability & Observability, Proficiency with Terraform, Ansible etc., Strong problem-solving skills with scripting (Python/Go/Shell)

 

Criteria:

1.   Candidate must be from a product-based or scalable app-based start-ups company with experience handling large-scale production traffic.

2.   Minimum 6 yrs of experience working as a DevOps/Infrastructure Consultant

3.   Candidate must have 2 years of experience as an lead (handling team of 3 to 4 members at least)

4.   Own end-to-end infrastructure right from non-prod to prod environment including self-managed

5.   Candidate must have Self experience in database migration from scratch 

6.   Must have a firm hold on the container orchestration tool Kubernetes

7.   Should have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

8.   Understanding programming languages like GO/Python, and Java

9.   Working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

10.   Working experience on Cloud platform -AWS

11. Candidate should have Minimum 1.5 years stability per organization, and a clear reason for relocation.

 

Description

Job Summary:

As a DevOps Engineer at company, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.

 

Job Responsibilities:

● Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs

● Codify our infrastructure

● Do what it takes to keep the uptime above 99.99%

● Understand the bigger picture and sail through the ambiguities

● Scale technology considering cost and observability and manage end-to-end processes

● Understand DevOps philosophy and evangelize the principles across the organization

● Strong communication and collaboration skills to break down the silos

 

Job Requirements:

● B.Tech. / B.E. degree in Computer Science or equivalent software engineering degree/experience

● Minimum 6 yrs of experience working as a DevOps/Infrastructure Consultant

● Must have a firm hold on the container orchestration tool Kubernetes

● Must have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

● Strong problem-solving skills, and ability to write scripts using any scripting language

● Understanding programming languages like GO/Python, and Java

● Comfortable working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

 

What’s there for you?

Company’s team handles everything – infra, tooling, and self-manages a bunch of databases, such as

● 150+ microservices with event-driven architecture across different tech stacks Golang/ java/ node

● More than 100,000 Request per second on our edge gateways

● ~20,000 events per second on self-managed Kafka

● 100s of TB of data on self-managed databases

● 100s of real-time continuous deployment to production

● Self-managed infra supporting

● 100% OSS

Read more
Ride-hailing Industry

Ride-hailing Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
4 - 6 yrs
₹34L - ₹37L / yr
DevOps
skill iconPython
Shell Scripting
skill iconKubernetes
Monitoring
+18 more

JOB DETAILS:

- Job Title: Senior Devops Engineer 1

- Industry: Ride-hailing

- Experience: 4-6 years

- Working Days: 5 days/week

- Work Mode: ONSITE

- Job Location: Bangalore

- CTC Range: Best in Industry


Required Skills: Cloud & Infrastructure Operations, Kubernetes & Container Orchestration, Monitoring, Reliability & Observability, Proficiency with Terraform, Ansible etc., Strong problem-solving skills with scripting (Python/Go/Shell)

 

Criteria:

1. Candidate must be from a product-based or scalable app-based startups company with experience handling large-scale production traffic.

2. Candidate must have strong Linux expertise with hands-on production troubleshooting and working knowledge of databases and middleware (Mongo, Redis, Cassandra, Elasticsearch, Kafka).

3. Candidate must have solid experience with Kubernetes.

4. Candidate should have strong knowledge of configuration management tools like Ansible, Terraform, and Chef / Puppet. Add on- Prometheus & Grafana etc.

5. Candidate must be an individual contributor with strong ownership.

6. Candidate must have hands-on experience with DATABASE MIGRATIONS and observability tools such as Prometheus and Grafana.

7. Candidate must have working knowledge of Go/Python and Java.

8. Candidate should have working experience on Cloud platform - AWS

9. Candidate should have Minimum 1.5 years stability per organization, and a clear reason for relocation.

 

Description 

Job Summary:

As a DevOps Engineer at company, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.

 

Job Responsibilities:

- Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs.

- Understanding the needs of stakeholders and conveying this to developers.

- Working on ways to automate and improve development and release processes.

- Identifying technical problems and developing software updates and ‘fixes’.

- Working with software developers to ensure that development follows established processes and works as intended.

- Do what it takes to keep the uptime above 99.99%.

- Understand DevOps philosophy and evangelize the principles across the organization.

- Strong communication and collaboration skills to break down the silos

 

Job Requirements:

- B.Tech. / B.E. degree in Computer Science or equivalent software engineering degree/experience.

- Minimum 4 yrs of experience working as a DevOps/Infrastructure Consultant.

- Strong background in operating systems like Linux.

- Understands the container orchestration tool Kubernetes.

- Proficient Knowledge of configuration management tools like Ansible, Terraform, and Chef / Puppet. Add on- Prometheus & Grafana etc.

- Problem-solving attitude, and ability to write scripts using any scripting language.

- Understanding programming languages like GO/Python, and Java.

- Basic understanding of databases and middlewares like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

- Should be able to take ownership of tasks, and must be responsible. - Good communication skills

 

Read more
Bengaluru (Bangalore)
5 - 10 yrs
₹25L - ₹50L / yr
skill iconNodeJS (Node.js)
skill iconReact.js
skill iconPython
skill iconJava
Data engineering
+10 more

Job Title : Senior Software Engineer (Full Stack — AI/ML & Data Applications)

Experience : 5 to 10 Years

Location : Bengaluru, India

Employment Type : Full-Time | Onsite


Role Overview :

We are seeking a Senior Full Stack Software Engineer with strong technical leadership and hands-on expertise in AI/ML, data-centric applications, and scalable full-stack architectures.

In this role, you will design and implement complex applications integrating ML/AI models, lead full-cycle development, and mentor engineering teams.


Mandatory Skills :

Full Stack Development (React/Angular/Vue + Node.js/Python/Java), Data Engineering (Spark/Kafka/ETL), ML/AI Model Integration (TensorFlow/PyTorch/scikit-learn), Cloud & DevOps (AWS/GCP/Azure, Docker, Kubernetes, CI/CD), SQL/NoSQL Databases (PostgreSQL/MongoDB).


Key Responsibilities :

  • Architect, design, and develop scalable full-stack applications for data and AI-driven products.
  • Build and optimize data ingestion, processing, and pipeline frameworks for large datasets.
  • Deploy, integrate, and scale ML/AI models in production environments.
  • Drive system design, architecture discussions, and API/interface standards.
  • Ensure engineering best practices across code quality, testing, performance, and security.
  • Mentor and guide junior developers through reviews and technical decision-making.
  • Collaborate cross-functionally with product, design, and data teams to align solutions with business needs.
  • Monitor, diagnose, and optimize performance issues across the application stack.
  • Maintain comprehensive technical documentation for scalability and knowledge-sharing.

Required Skills & Experience :

  • Education : B.E./B.Tech/M.E./M.Tech in Computer Science, Data Science, or equivalent fields.
  • Experience : 5+ years in software development with at least 2+ years in a senior or lead role.
  • Full Stack Proficiency :
  • Front-end : React / Angular / Vue.js
  • Back-end : Node.js / Python / Java
  • Data Engineering : Experience with data frameworks such as Apache Spark, Kafka, and ETL pipeline development.
  • AI/ML Expertise : Practical exposure to TensorFlow, PyTorch, or scikit-learn and deploying ML models at scale.
  • Databases : Strong knowledge of SQL & NoSQL systems (PostgreSQL, MongoDB) and warehousing tools (Snowflake, BigQuery).
  • Cloud & DevOps : Working knowledge of AWS, GCP, or Azure; containerization & orchestration (Docker, Kubernetes); CI/CD; MLflow/SageMaker is a plus.
  • Visualization : Familiarity with modern data visualization tools (D3.js, Tableau, Power BI).

Soft Skills :

  • Excellent communication and cross-functional collaboration skills.
  • Strong analytical mindset with structured problem-solving ability.
  • Self-driven with ownership mentality and adaptability in fast-paced environments.

Preferred Qualifications (Bonus) :

  • Experience deploying distributed, large-scale ML or data-driven platforms.
  • Understanding of data governance, privacy, and security compliance.
  • Exposure to domain-driven data/AI use cases in fintech, healthcare, retail, or e-commerce.
  • Experience working in Agile environments (Scrum/Kanban).
  • Active open-source contributions or a strong GitHub technical portfolio.
Read more
Inflectionio

at Inflectionio

1 candid answer
Renu Philip
Posted by Renu Philip
Bengaluru (Bangalore)
3 - 5 yrs
₹20L - ₹30L / yr
skill iconAmazon Web Services (AWS)
skill iconKubernetes
skill iconJenkins
Chef
CI/CD
+6 more

We are looking for a DevOps Engineer with hands-on experience in managing production infrastructure using AWS, Kubernetes, and Terraform. The ideal candidate will have exposure to CI/CD tools and queueing systems, along with a strong ability to automate and optimize workflows.


Responsibilities: 

* Manage and optimize production infrastructure on AWS, ensuring scalability and reliability.

* Deploy and orchestrate containerized applications using Kubernetes.

* Implement and maintain infrastructure as code (IaC) using Terraform.

* Set up and manage CI/CD pipelines using tools like Jenkins or Chef to streamline deployment processes.

* Troubleshoot and resolve infrastructure issues to ensure high availability and performance.

* Collaborate with cross-functional teams to define technical requirements and deliver solutions.

* Nice-to-have: Manage queueing systems like Amazon SQS, Kafka, or RabbitMQ.



Requirements: 

* 2+ years of experience with AWS, including practical exposure to its services in production environments.

* Demonstrated expertise in Kubernetes for container orchestration.

* Proficiency in using Terraform for managing infrastructure as code.

* Exposure to at least one CI/CD tool, such as Jenkins or Chef.

* Nice-to-have: Experience managing queueing systems like SQS, Kafka, or RabbitMQ.

Read more
Technology Industry

Technology Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
6 - 9 yrs
₹30L - ₹48L / yr
skill iconPython
skill iconReact.js
skill iconNodeJS (Node.js)
TypeScript
ReAct (Reason + Act)
+13 more

Review Criteria:

  • Strong Software Engineer fullstack profile using NodeJS / Python and React
  • 6+ YOE in Software Development using Python OR NodeJS (For backend) & React (For frontend)
  • Must have strong experience in working on Typescript
  • Must have experience in message-based systems like Kafka, RabbitMq, Redis
  • Databases - PostgreSQL & NoSQL databases like MongoDB
  • Product Companies Only
  • Tier 1 Engineering Institutes (IIT, NIT, BITS, IIIT, DTU or equivalent)

 

Preferred:

  • Experience in Fin-Tech, Payment, POS and Retail products is highly preferred
  • Experience in mentoring, coaching the team.


Role & Responsibilities:

We are currently seeking a Senior Engineer to join our Financial Services team, contributing to the design and development of scalable system.

 

The Ideal Candidate Will Be Able To-

  • Take ownership of delivering performant, scalable and high-quality cloud-based software, both frontend and backend side.
  • Mentor team members to develop in line with product requirements.
  • Collaborate with Senior Architect for design and technology choices for product development roadmap.
  • Do code reviews.


Ideal Candidate:

  • Thorough knowledge of developing cloud-based software including backend APIs and react based frontend.
  • Thorough knowledge of scalable design patterns and message-based systems such as Kafka, RabbitMq, Redis, MongoDB, ORM, SQL etc.
  • Experience with AWS services such as S3, IAM, Lambda etc.
  • Expert level coding skills in Python FastAPI/Django, NodeJs, TypeScript, ReactJs.
  • Eye for user responsive designs on the frontend.


Perks, Benefits and Work Culture:

  • We prioritize people above all else. While we're recognized for our innovative technology solutions, it's our people who drive our success. That’s why we offer a comprehensive and competitive benefits package designed to support your well-being and growth:
  • Medical Insurance with coverage up to INR 8,00,000 for the employee and their family
Read more
Tarento Group

at Tarento Group

3 candid answers
1 recruiter
Reshika Mendiratta
Posted by Reshika Mendiratta
Bengaluru (Bangalore)
8yrs+
Upto ₹30L / yr (Varies
)
skill iconJava
skill iconSpring Boot
Microservices
Windows Azure
RESTful APIs
+7 more

About Tarento:

 

Tarento is a fast-growing technology consulting company headquartered in Stockholm, with a strong presence in India and clients across the globe. We specialize in digital transformation, product engineering, and enterprise solutions, working across diverse industries including retail, manufacturing, and healthcare. Our teams combine Nordic values with Indian expertise to deliver innovative, scalable, and high-impact solutions.

 

We're proud to be recognized as a Great Place to Work, a testament to our inclusive culture, strong leadership, and commitment to employee well-being and growth. At Tarento, you’ll be part of a collaborative environment where ideas are valued, learning is continuous, and careers are built on passion and purpose.


Job Summary:

We are seeking a highly skilled and self-driven Senior Java Backend Developer with strong experience in designing and deploying scalable microservices using Spring Boot and Azure Cloud. The ideal candidate will have hands-on expertise in modern Java development, containerization, messaging systems like Kafka, and knowledge of CI/CD and DevOps practices.


Key Responsibilities:

  • Design, develop, and deploy microservices using Spring Boot on Azure cloud platforms.
  • Implement and maintain RESTful APIs, ensuring high performance and scalability.
  • Work with Java 11+ features including Streams, Functional Programming, and Collections framework.
  • Develop and manage Docker containers, enabling efficient development and deployment pipelines.
  • Integrate messaging services like Apache Kafka into microservice architectures.
  • Design and maintain data models using PostgreSQL or other SQL databases.
  • Implement unit testing using JUnit and mocking frameworks to ensure code quality.
  • Develop and execute API automation tests using Cucumber or similar tools.
  • Collaborate with QA, DevOps, and other teams for seamless CI/CD integration and deployment pipelines.
  • Work with Kubernetes for orchestrating containerized services.
  • Utilize Couchbase or similar NoSQL technologies when necessary.
  • Participate in code reviews, design discussions, and contribute to best practices and standards.


Required Skills & Qualifications:

  • Strong experience in Java (11 or above) and Spring Boot framework.
  • Solid understanding of microservices architecture and deployment on Azure.
  • Hands-on experience with Docker, and exposure to Kubernetes.
  • Proficiency in Kafka, with real-world project experience.
  • Working knowledge of PostgreSQL (or any SQL DB) and data modeling principles.
  • Experience in writing unit tests using JUnit and mocking tools.
  • Experience with Cucumber or similar frameworks for API automation testing.
  • Exposure to CI/CD tools, DevOps processes, and Git-based workflows.


Nice to Have:

  • Azure certifications (e.g., Azure Developer Associate)
  • Familiarity with Couchbase or other NoSQL databases.
  • Familiarity with other cloud providers (AWS, GCP)
  • Knowledge of observability tools (Prometheus, Grafana, ELK)


Soft Skills:

  • Strong problem-solving and analytical skills.
  • Excellent verbal and written communication.
  • Ability to work in an agile environment and contribute to continuous improvement.


Why Join Us:

  • Work on cutting-edge microservice architectures
  • Strong learning and development culture
  • Opportunity to innovate and influence technical decisions
  • Collaborative and inclusive work environment
Read more
AI-First Company

AI-First Company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore), Mumbai, Hyderabad, Gurugram
5 - 17 yrs
₹30L - ₹45L / yr
Data engineering
Data architecture
SQL
Data modeling
GCS
+47 more

ROLES AND RESPONSIBILITIES:

You will be responsible for architecting, implementing, and optimizing Dremio-based data Lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.


  • Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
  • Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
  • Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
  • Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
  • Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
  • Support self-service analytics by enabling governed data products and semantic layers.
  • Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
  • Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.


IDEAL CANDIDATE:

  • Bachelor’s or Master’s in Computer Science, Information Systems, or related field.
  • 5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
  • Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
  • Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
  • Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
  • Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
  • Excellent problem-solving, documentation, and stakeholder communication skills.


PREFERRED:

  • Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) and data catalogs (Collibra, Alation, Purview).
  • Exposure to Snowflake, Databricks, or BigQuery environments.
  • Experience in high-tech, manufacturing, or enterprise data modernization programs.
Read more
iMerit
Bengaluru (Bangalore)
6 - 9 yrs
₹10L - ₹15L / yr
DevOps
Terraform
Apache Kafka
skill iconPython
skill iconGo Programming (Golang)
+4 more

Exp: 7- 10 Years

CTC: up to 35 LPA


Skills:

  • 6–10 years DevOps / SRE / Cloud Infrastructure experience
  • Expert-level Kubernetes (networking, security, scaling, controllers)
  • Terraform Infrastructure-as-Code mastery
  • Hands-on Kafka production experience
  • AWS cloud architecture and networking expertise
  • Strong scripting in Python, Go, or Bash
  • GitOps and CI/CD tooling experience


Key Responsibilities:

  • Design highly available, secure cloud infrastructure supporting distributed microservices at scale
  • Lead multi-cluster Kubernetes strategy optimized for GPU and multi-tenant workloads
  • Implement Infrastructure-as-Code using Terraform across full infrastructure lifecycle
  • Optimize Kafka-based data pipelines for throughput, fault tolerance, and low latency
  • Deliver zero-downtime CI/CD pipelines using GitOps-driven deployment models
  • Establish SRE practices with SLOs, p95 and p99 monitoring, and FinOps discipline
  • Ensure production-ready disaster recovery and business continuity testing



If interested Kindly share your updated resume at 82008 31681

Read more
Tarento Group

at Tarento Group

3 candid answers
1 recruiter
Reshika Mendiratta
Posted by Reshika Mendiratta
Bengaluru (Bangalore)
4yrs+
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
Windows Azure
RESTful APIs
+5 more

Job Summary:

We are seeking a highly skilled and self-driven Java Backend Developer with strong experience in designing and deploying scalable microservices using Spring Boot and Azure Cloud. The ideal candidate will have hands-on expertise in modern Java development, containerization, messaging systems like Kafka, and knowledge of CI/CD and DevOps practices.Key Responsibilities:

  • Design, develop, and deploy microservices using Spring Boot on Azure cloud platforms.
  • Implement and maintain RESTful APIs, ensuring high performance and scalability.
  • Work with Java 11+ features including Streams, Functional Programming, and Collections framework.
  • Develop and manage Docker containers, enabling efficient development and deployment pipelines.
  • Integrate messaging services like Apache Kafka into microservice architectures.
  • Design and maintain data models using PostgreSQL or other SQL databases.
  • Implement unit testing using JUnit and mocking frameworks to ensure code quality.
  • Develop and execute API automation tests using Cucumber or similar tools.
  • Collaborate with QA, DevOps, and other teams for seamless CI/CD integration and deployment pipelines.
  • Work with Kubernetes for orchestrating containerized services.
  • Utilize Couchbase or similar NoSQL technologies when necessary.
  • Participate in code reviews, design discussions, and contribute to best practices and standards.

Required Skills & Qualifications:

  • Strong experience in Java (11 or above) and Spring Boot framework.
  • Solid understanding of microservices architecture and deployment on Azure.
  • Hands-on experience with Docker, and exposure to Kubernetes.
  • Proficiency in Kafka, with real-world project experience.
  • Working knowledge of PostgreSQL (or any SQL DB) and data modeling principles.
  • Experience in writing unit tests using JUnit and mocking tools.
  • Experience with Cucumber or similar frameworks for API automation testing.
  • Exposure to CI/CD toolsDevOps processes, and Git-based workflows.

Nice to Have:

  • Azure certifications (e.g., Azure Developer Associate)
  • Familiarity with Couchbase or other NoSQL databases.
  • Familiarity with other cloud providers (AWS, GCP)
  • Knowledge of observability tools (Prometheus, Grafana, ELK)

Soft Skills:

  • Strong problem-solving and analytical skills.
  • Excellent verbal and written communication.
  • Ability to work in an agile environment and contribute to continuous improvement.

Why Join Us:

  • Work on cutting-edge microservice architectures
  • Strong learning and development culture
  • Opportunity to innovate and influence technical decisions
  • Collaborative and inclusive work environment
Read more
Capace Software Private Limited
Bengaluru (Bangalore), Bhopal
5 - 10 yrs
₹4L - ₹10L / yr
skill iconDjango
CI/CD
Software deployment
RESTful APIs
skill iconFlask
+8 more

Senior Python Django Developer 

Experience: Back-end development: 6 years (Required)


Location:  Bangalore/ Bhopal

Job Description:

We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the start-up environment.

This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.

Responsibilities:

  • Design and develop scalable, secure, and high-performance applications using Python (Django framework).
  • Architect system components, define database schemas, and optimize backend services for speed and efficiency.
  • Lead and implement design patterns and software architecture best practices.
  • Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
  • Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
  • Drive performance improvements, monitor system health, and troubleshoot production issues.
  • Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
  • Contribute to technical decision-making and mentor junior developers.

Requirements:

  • 6 to 10 years of professional backend development experience with Python and Django.
  • Strong background in payments/financial systems or FinTech applications.
  • Proven experience in designing software architecture in a microservices or modular monolith environment.
  • Experience working in fast-paced startup environments with agile practices.
  • Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
  • Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
  • Hands-on experience with test-driven development (TDD) and frameworks like pytest, unittest, or factory_boy.
  • Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).

Preferred Skills:

  • Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
  • Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
  • Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
  • Contributions to open-source or personal finance-related projects.

Job Types: Full-time, Permanent


Schedule:

  • Day shift

Supplemental Pay:

  • Performance bonus
  • Yearly bonus

Ability to commute/relocate:

  • JP Nagar, 5th Phase, Bangalore, Karnataka or Indrapuri, Bhopal, Madhya Pradesh: Reliably commute or willing to relocate with an employer-provided relocation package (Preferred)


Read more
DigitalSprint AI Solutions

at DigitalSprint AI Solutions

1 candid answer
Nalini Sanka
Posted by Nalini Sanka
Bengaluru (Bangalore)
6 - 12 yrs
₹15L - ₹25L / yr
skill iconJava
skill iconSpring Boot
skill iconAmazon Web Services (AWS)
Apache Kafka
skill iconJenkins

Requirements

  • 6–12 years of backend development experience.
  • Strong expertise in Java 11+, Spring Boot, REST APIs, AWS.
  • Solid experience with distributed, high-volume systems.
  • Strong knowledge of RDBMS (e.g., MySQL, Oracle) and NoSQL databases (e.g., DynamoDB, MongoDB, Cassandra).
  • Hands-on with CI/CD (Jenkins) and caching technologies Redis or Similar.
  • Strong debugging and system troubleshooting skills.
  • Experience in payments system is a must.


Read more
Tradelab Technologies
Bengaluru (Bangalore)
2 - 4 yrs
₹10L - ₹15L / yr
MySQL
skill iconRedis
Apache Kafka

We're Hiring for Golang Developer (2–4 years experience)

Company: Tradelab Technologies

Location: Bangalore (Preferred candidates from Bangalore or nearby locations only)


We are looking for a Golang Developer with strong experience in backend development, microservices, and system-level programming. The role involves contributing to high-performance trading systems, low-latency architecture, and scalable backend solutions.


Key Responsibilities

• Develop and maintain backend services using Golang

• Build scalable, secure and high-performance microservices

• Work with REST APIs, WebSockets, message queues and distributed systems

• Collaborate with DevOps, frontend and product teams for smooth delivery

• Optimize performance, troubleshoot issues and ensure system stability


Skills & Experience Required

• 2–4 years of experience in Golang development

• Strong fundamentals in data structures, concurrency and networking

• Experience with MySQL/Redis/Kafka or similar technologies

• Understanding of microservices, APIs and cloud environments

• Experience in fintech/trading systems is a plus


If you are passionate about backend engineering and want to work on fast, scalable trading systems.


Read more
Tradelab Technologies
Aakanksha Yadav
Posted by Aakanksha Yadav
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
skill iconPython
RestAPI
FastAPI
RabbitMQ
Apache Kafka
+3 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What We Expect:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and

a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex

performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What You Will Do:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Experience with GoLang, C/C++, Erlang, or Elixir.

• Exposure to trading, fintech, or low-latency systems.

• Familiarity with microservices and CI/CD pipelines.



Read more
AI company

AI company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore), Mumbai, Hyderabad, Gurugram
5 - 17 yrs
₹30L - ₹45L / yr
Data architecture
Data engineering
SQL
Data modeling
GCS
+21 more

Review Criteria

  • Strong Dremio / Lakehouse Data Architect profile
  • 5+ years of experience in Data Architecture / Data Engineering, with minimum 3+ years hands-on in Dremio
  • Strong expertise in SQL optimization, data modeling, query performance tuning, and designing analytical schemas for large-scale systems
  • Deep experience with cloud object storage (S3 / ADLS / GCS) and file formats such as Parquet, Delta, Iceberg along with distributed query planning concepts
  • Hands-on experience integrating data via APIs, JDBC, Delta/Parquet, object storage, and coordinating with data engineering pipelines (Airflow, DBT, Kafka, Spark, etc.)
  • Proven experience designing and implementing lakehouse architecture including ingestion, curation, semantic modeling, reflections/caching optimization, and enabling governed analytics
  • Strong understanding of data governance, lineage, RBAC-based access control, and enterprise security best practices
  • Excellent communication skills with ability to work closely with BI, data science, and engineering teams; strong documentation discipline
  • Candidates must come from enterprise data modernization, cloud-native, or analytics-driven companies


Preferred

  • Preferred (Nice-to-have) – Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) or data catalogs (Collibra, Alation, Purview); familiarity with Snowflake, Databricks, or BigQuery environments


Job Specific Criteria

  • CV Attachment is mandatory
  • How many years of experience you have with Dremio?
  • Which is your preferred job location (Mumbai / Bengaluru / Hyderabad / Gurgaon)?
  • Are you okay with 3 Days WFO?
  • Virtual Interview requires video to be on, are you okay with it?


Role & Responsibilities

You will be responsible for architecting, implementing, and optimizing Dremio-based data lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.

  • Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
  • Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
  • Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
  • Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
  • Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
  • Support self-service analytics by enabling governed data products and semantic layers.
  • Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
  • Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.


Ideal Candidate

  • Bachelor’s or master’s in computer science, Information Systems, or related field.
  • 5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
  • Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
  • Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
  • Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
  • Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
  • Excellent problem-solving, documentation, and stakeholder communication skills.
Read more
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
skill iconGo Programming (Golang)
CI/CD
Apache Kafka
RabbitMQ
skill iconDocker
+1 more

Location: Bengalore, India, Exp: 3-5 Yrs

Backend Developer (Golang) - Trading & Fintech


About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What we expect:

 You should already be exceptional at Golang. If you need hand-holding, this isn’t the place for you.

 You thrive on challenges, not on perks or financial rewards.

 You measure success by your own growth, not external validation.

 Taking calculated risks excites you—you’re here to build, break, and learn.

 You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading environment.

 You understand the stakes—milliseconds can make or break trades, and precision is everything.



What you will do:

 Develop and optimize high-performance backend systems in Golang for trading platforms and financial services.

 Architect low-latency, high-throughput microservices that push the boundaries ofspeed and efficiency.

 Build event-driven, fault-tolerant systems that can handle massive real-time data streams.

 Own your work—no babysitting, no micromanagement.

 Work alongside equally driven engineers who expect nothing less than brilliance.


Must have skills:

 Learn faster than you ever thought possible.

Proven expertise in Golang (if you need to prove yourself, this isn’t the role for you).

 Deep understanding of concurrency, memory management, and system design.

 Experience with Trading, market data processing, or low-latency systems.

 Strong knowledge of distributed systems, message queues (Kafka, RabbitMQ), and real-time processing.

 Hands-on with Docker, Kubernetes, and CI/CD pipelines.

 A portfolio of work that speaks louder than a resume.


Nice-to-Have Skills:

 Past experience in fintech, trading systems, or algorithmic trading.

 Contributions to open-source Golang projects.

 A history of building something impactful from scratch.

 Understanding of FIX protocol, WebSockets, and streaming APIs.


Why Join Us?

 Work with a team that expects and delivers excellence.

 A culture where risk-taking is rewarded, and complacency is not.

 Limitless opportunities for growth—if you can handle the pace.

 A place where learning is currency, and outperformance is the only metric that matters.


The opportunity to build systems that move markets, execute trades in microseconds, and redefine fintech. This isn’t just a job—it’s a proving ground. Ready to take the leap? Apply now.



Read more
prep study
Pooja Sharma
Posted by Pooja Sharma
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
FastAPI
RESTful APIs
RabbitMQ
Apache Kafka
skill iconElastic Search
+2 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.

What We Expect:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.

Read more
Bengaluru (Bangalore)
4 - 7 yrs
₹15L - ₹35L / yr
skill iconJava
skill iconSpring Boot
Apache Kafka

SE/SSE - Backend (Java)



JAVA: Software Engineer / Senior Software Engineer


Job Overview: RevSure.AI is building a world-class team of Engineers with a mandate to architect, design, build, scale, and maintain our cutting-edge Revops platform and derive insights from the customer data. If you are a great JavaScript developer (or, for that matter, in any major programming language) with experience in building SaaS web applications and looking for an opportunity to build world-class products using cutting-edge technologies, please read on -


Responsibilities:

● Contribute to the growth and development of our product by writing and maintaining code, focusing on quality and maintainability. We believe that our teams own the code they build.

● Share your knowledge, both inside and outside your own team.

● Design, implement, test, and monitor valuable solutions that achieve the teamʼs goals while keeping a smooth delivery flow and balancing tradeoffs between scope, time, and effort.

● Continuously improve your and your teamʼs way of working, expanding the boundaries of your teamʼs autonomy and self-organisation.


Requirements:


Must-Have:

● Good interpersonal and communication skills. An analytical mind and an eye for detail

● Experience of 4-7 years in building large applications, preferably from the heavy data processing & analytics side

● Experience in architecting, designing, and building end-to-end System Design of multi tenant SAAS applications

● Experience architecting interactive configuration experiences around DAGs, configuration-heavy reporting feature, interactive data pipeline design interfaces, and charting.

● Experience in Java, Spring Boot, and GCP

● In-depth understanding of design patterns, OOPs, and Functional programming.

● Passionate programmer focused on Backend Programming.


Good to have:

● Working knowledge of GCP managed services like Kubernetes, PubSub


Please feel free to share your resume to chandana at revsure.ai


**Based on your experience, the title could vary. We care more about mindset , ownership, and craft than just titles.

Read more
Tradelab Software Private Limited
Pooja Sharma
Posted by Pooja Sharma
Bengaluru (Bangalore)
5 - 8 yrs
₹10L - ₹15L / yr
concurrency
Apache Kafka
RabbitMQ
skill iconDocker
skill iconKubernetes

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What We Expect:

• You should already be exceptional at Backend. If you need hand-holding, this isn’t the place for you.

• You thrive on challenges, not on perks or financial rewards.

• You measure success by your own growth, not external validation.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a pay check; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What You Will Do:

• Develop and optimize high-performance backend systems in for trading platforms and financial services.

• Architect low-latency, high-throughput microservices that push the boundaries of speed and efficiency.

• Build event-driven, fault-tolerant systems that can handle massive real-time data streams.

• Own your work—no babysitting, no micromanagement.

• Work alongside equally driven engineers who expect nothing less than brilliance.

• Learn faster than you ever thought possible.


Must-Have Skills:

Proven expertise in Backend (if you need to prove yourself, this isn’t the role for you).

• Deep understanding of concurrency, memory management, and system design.

• Experience with Trading, market data processing, or low-latency systems.

• Strong knowledge of distributed systems, message queues (Kafka, RabbitMQ), and real-time processing.

• Hands-on with Docker, Kubernetes, and CI/CD pipelines.

• A portfolio of work that speaks louder than a resume.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Contributions to open-source projects.

• A history of building something impactful from scratch.

• Understanding of FIX protocol, WebSockets, and streaming APIs.


Why Join Us?

• Work with a team that expects and delivers excellence.

• A culture where risk-taking is rewarded, and complacency is not.

• Limitless opportunities for growth—if you can handle the pace.

• A place where learning is currency, and outperformance is the only metric that matters.

• The opportunity to build systems that move markets, execute trades in microseconds, and redefine fintech.

Read more
Tradelab Software Private Limited
Pooja Sharma
Posted by Pooja Sharma
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
RESTful APIs
RabbitMQ
Apache Kafka
skill iconRedis
skill iconElastic Search
+2 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance. driven backend systems, we want you.


What We Expect:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and

a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex

performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What You Will Do:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field

Read more
ADTRAN

at ADTRAN

1 recruiter
Reema Meshram
Posted by Reema Meshram
Bengaluru (Bangalore)
3 - 7 yrs
₹12L - ₹35L / yr
skill iconScala
Functional programming
Akka
Apache Kafka
skill iconPostgreSQL

Key Responsibilities

  • Provide technical leadership in the design, development, and delivery of scalable, high-performance software systems.
  • Partner with product managers, architects, and cross-functional teams to define technical strategy and ensure alignment with business objectives.
  • Lead by example in writing high-quality, testable, and maintainable code.
  • Drive best practices in software engineering, including code reviews, system design, and performance optimization.
  • Mentor and guide engineers across teams, fostering a culture of technical excellence and continuous learning.
  • Evaluate and introduce new technologies, tools, and frameworks to improve productivity, scale and system robustness.

Required Skills & Qualifications

  • Strong foundation in computer science fundamentals: data structures, algorithms, and functional programming techniques.
  • Expertise in Scala, with strong preference for functional programming.
  • Solid experience in software design, implementation, and debugging, including inter-process communication and multi-threading.
  • Hands-on experience with distributed systems and event-driven architectures.
  • Familiarity with databases (Postgres preferred).
  • Proficiency with Apache Kafka for messaging and persistence.
  • Working knowledge of Python for unit and integration testing.
  • Basic to intermediate experience with Ansible for automation.
  • Strong problem-solving, analytical, and communication skills.

Nice-to-Have / Bonus Skills

  • Experience with modeling in YANG.
  • Experience with Scala libraries such as Cats Effect (2/3), Monix, and Akka.
  • Experience working in Agile/Scrum environments.

What We Offer

  • Opportunity to work on cutting-edge technologies in a collaborative environment.
  • A role with strong ownership, technical influence, and visibility across teams.
  • Competitive compensation and benefits.


Read more
Lookup

at Lookup

2 candid answers
1 recruiter
Ajay Kumar
Posted by Ajay Kumar
Bengaluru (Bangalore)
3 - 8 yrs
₹30L - ₹35L / yr
skill iconPython
skill iconPostgreSQL
FastAPI
skill iconRedis
skill iconAmazon Web Services (AWS)
+2 more

Our Mission

To make video as accessible to machines as text and voice are today.


At lookup, we believe the world's most valuable asset is trapped. Video is everywhere, but it's unsearchable—a black box of insight that no one can open or atleast open affordably. We’re changing that. We're building the search engine for the visual world, so anyone can find or do anything with video just by asking.


Text is queryable. Voice is transcribed. Video, the largest and richest data source of all, is still a black box. A computer can't understand it, and so its value remains trapped.


Our mission at lookup is to fix this.


About the Role

We are looking for founding Backend Engineers to build a highly performant, reliable, and scalable API platform that makes enterprise video knowledge readily available for video search, summarization, and natural‑language Q&A. You will partner closely with our ML team working on vision‑language models to productionize research and deliver fast, trustworthy APIs for customers.


Examples of technical challenges you will work on include: distributed video storage, a unified application framework and data model for indexing large video libraries, low‑latency clip retrieval, vector search at scale, and end‑to‑end build, test, deploy, and observability in cloud environments.


What You’ll Do:

  • Design and build robust backend services and APIs (REST, gRPC) for vector search, video summarization, and video Q&A.
  • Own API performance and reliability, including low‑latency retrieval, pagination, rate limiting, and backwards‑compatible versioning.
  • Design schemas and tune queries in Postgres, and integrate with unstructured storage.
  • Implement observability across metrics, logs, and traces. Set error budgets and SLOs.
  • Write clear design docs and ship high‑quality, well‑tested code.
  • Collaborate with ML engineers to integrate and productionize VLMs and retrieval pipelines.
  • Take ownership of architecture from inception to production launch.


Who You Are:

  • 3+ years of professional experience in backend development.
  • Proven experience building and scaling polished WebSocket, gRPC, and REST APIs.
  • Exposure to distributed systems and container orchestration (Docker and Kubernetes).
  • Hands‑on experience with AWS.
  • Strong knowledge of SQL (Postgres) and NoSQL (e.g., Cassandra), including schema design, query optimization, and scaling.
  • Familiarity with our stack is a plus, but not mandatory: Python (FastAPI), Celery, Kafka, Postgres, Redis, Weaviate, React.
  • Ability to diagnose complex issues, identify root causes, and implement effective fixes.
  • Comfortable working in a fast‑paced startup environment.


Nice to have:

  • Hands-on work with LLM agents, vector embeddings, or RAG applications.
  • Building video streaming pipelines and storage systems at scale (FFmpeg, RTSP, WebRTC).
  • Proficiency with modern frontend frameworks (React, TypeScript, Tailwind CSS) and responsive UI design.


Location & Culture

  • Full-time, in-office role in Bangalore (we’re building fast and hands-on).
  • Must be comfortable with a high-paced environment and collaboration across PST time zones for our US customers and investors.
  • Expect startup speed — daily founder syncs, rapid design-to-prototype cycles, and a culture of deep ownership.


Why You’ll Love This Role

  • Work on the frontier of video understanding and real-world AI — products that can redefine trust and automation.
  • Build core APIs that make video queryable and power real customer use.
  • Own systems end to end: performance, reliability, and developer experience.
  • Work closely with founders and collaborate in person in Bangalore.
  • Competitive salary with meaningful early equity.
Read more
Bengaluru (Bangalore)
2 - 5 yrs
₹10L - ₹15L / yr
Apache Kafka
RabbitMQ
skill iconDocker
skill iconKubernetes
CI/CD

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What You Will Do:

• Develop and optimize high-performance backend systems in Golang for trading platforms and financial

services.

• Architect low-latency, high-throughput microservices that push the boundaries of speed and efficiency.

• Build event-driven, fault-tolerant systems that can handle massive real-time data streams.

• Own your work—no babysitting, no micromanagement.

• Work alongside equally driven engineers who expect nothing less than brilliance.

• Learn faster than you ever thought possible.


Must-Have Skills:

Proven expertise in Golang (if you need to prove yourself, this isn’t the role for you).

• Deep understanding of concurrency, memory management, and system design.

• Experience with Trading, market data processing, or low-latency systems.

• Strong knowledge of distributed systems, message queues (Kafka, RabbitMQ), and real-time processing.

• Hands-on with Docker, Kubernetes, and CI/CD pipelines.

• A portfolio of work that speaks louder than a resume.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Contributions to open-source Golang projects.

• A history of building something impactful from scratch.

• Understanding of FIX protocol, WebSockets, and streaming APIs.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Robin Silverster
Posted by Robin Silverster
Bengaluru (Bangalore)
5 - 11 yrs
₹10L - ₹35L / yr
skill iconPython
Spark
Apache Kafka
Snow flake schema
databricks
+1 more

Required Skills:

· 8+ years of being a practitioner in data engineering or a related field.

· Proficiency in programming skills in Python

· Experience with data processing frameworks like Apache Spark or Hadoop.

· Experience working on Databricks.

· Familiarity with cloud platforms (AWS, Azure) and their data services.

· Experience with data warehousing concepts and technologies.

· Experience with message queues and streaming platforms (e.g., Kafka).

· Excellent communication and collaboration skills.

· Ability to work independently and as part of a geographically distributed team.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Manisha Gouda
Posted by Manisha Gouda
Bengaluru (Bangalore)
3 - 6 yrs
₹5L - ₹18L / yr
DevOps
Red Hat Linux
Apache Kafka
IBM WebSphere MQ
skill iconKubernetes
+1 more

- Experience comparable to DevOps SIRE providing SME-tevel application or platform support with responsibility for designing and automating operational procedures and best practices 


-Experience writing python and shell scripts to perform health checks and automations 


- Experience with Linux System Administration (preferably Red Hat)


- Hands-on experience with multi-tenant hosting environments for middleware applications (for example: centrally managed platform or infrastructure as a service) 


- Experience with implementing observabitity, monitoring and alerting tools 


- Excellent written and oral English communication skills. The candidate must write user-facing documentation, prepare and deliver presentations to an internal audience and effectively interact with upper management, colleagues, and customers 


- Independent problem-solving skills, self-motivated, and a mindset for taking ownership


- A minimum of 5 years of infrastructure production support or DevOps experience 




Additional Technical Skills 


Experience with broker-based messaging infrastructure such as Apache Kafka, IBM MQ (or similar technology like ActiveMQ, Azure Service Bus) including configuration and performance tuning 

Experience with public/private cloud and containerization technologies (e.g. Kubernetes) 

Experience with Agile development methodology, Cl/CD and automated build pipelines 

Experience with DevOps methodology (e.g. Phoenix Project) 

Experience with tools such as Jira, Confluence and ServiceNow 

Experience working with JSON, XML, Google Protocol Buffers, Avro, FIX 

Experience with troubleshooting tools such as TCPdump and Wireshark

Experience with NoSQL databases such as MongoDB and Redis interest and understanding of emerging IT trends 

Experience with system architecture design


Read more
Publicis Sapient

at Publicis Sapient

10 recruiters
Dipika
Posted by Dipika
Bengaluru (Bangalore), Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Hyderabad, Pune
5 - 7 yrs
₹5L - ₹20L / yr
skill iconJava
Microservices
06692
Apache Kafka
Apache ActiveMQ
+3 more

1 Senior Associate Technology L1 – Java Microservices


Company Description

Publicis Sapient is a digital transformation partner helping established organizations get to their future, digitally-enabled state, both in the way they work and the way they serve their customers. We help unlock value through a start-up mindset and modern methods, fusing strategy, consulting and customer experience with agile engineering and problem-solving creativity. United by our core values and our purpose of helping people thrive in the brave pursuit of next, our 20,000+ people in 53 offices around the world combine experience across technology, data sciences, consulting and customer obsession to accelerate our clients’ businesses through designing the products and services their customers truly value.


Job Description

We are looking for a Senior Associate Technology Level 1 - Java Microservices Developer to join our team of bright thinkers and doers. You’ll use your problem-solving creativity to design, architect, and develop high-end technology solutions that solve our clients’ most complex and challenging problems across different industries.

We are on a mission to transform the world, and you will be instrumental in shaping how we do it with your ideas, thoughts, and solutions.


Your Impact:

• Drive the design, planning, and implementation of multifaceted applications, giving you breadth and depth of knowledge across the entire project lifecycle.

• Combine your technical expertise and problem-solving passion to work closely with clients, turning • complex ideas into end-to-end solutions that transform our clients’ business

• Constantly innovate and evaluate emerging technologies and methods to provide scalable and elegant solutions that help clients achieve their business goals.


Qualifications

➢ 5 to 7 Years of software development experience

➢ Strong development skills in Java JDK 1.8 or above

➢ Java fundamentals like Exceptional handling, Serialization/Deserialization and Immutability concepts

➢ Good fundamental knowledge in Enums, Collections, Annotations, Generics, Auto boxing and Data Structure

➢ Database RDBMS/No SQL (SQL, Joins, Indexing)

➢ Multithreading (Re-entrant Lock, Fork & Join, Sync, Executor Framework)

➢ Spring Core & Spring Boot, security, transactions ➢ Hands-on experience with JMS (ActiveMQ, RabbitMQ, Kafka etc)

➢ Memory Mgmt (JVM configuration, Profiling, GC), profiling, Perf tunning, Testing, Jmeter/similar tool)

➢ Devops (CI/CD: Maven/Gradle, Jenkins, Quality plugins, Docker and containersization)

➢ Logical/Analytical skills. Thorough understanding of OOPS concepts, Design principles and implementation of

➢ different type of Design patterns. ➢ Hands-on experience with any of the logging frameworks (SLF4J/LogBack/Log4j) ➢ Experience of writing Junit test cases using Mockito / Powermock frameworks.

➢ Should have practical experience with Maven/Gradle and knowledge of version control systems like Git/SVN etc.

➢ Good communication skills and ability to work with global teams to define and deliver on projects.

➢ Sound understanding/experience in software development process, test-driven development.

➢ Cloud – AWS / AZURE / GCP / PCF or any private cloud would also be fine

➢ Experience in Microservices

Read more
Geektrust
Agency job
via Scaling Theory by Sriraagavi Baskaran
Bengaluru (Bangalore)
2.5 - 5 yrs
₹10L - ₹30L / yr
skill iconJava
Microsoft SharePoint
Apache Kafka
skill iconElastic Search

Role Overview


We are seeking a motivated and technically versatile CMS Engineer (IC2) to support our transition from SharePoint to Contentful, while also contributing to the broader CMS ecosystem. This is an excellent opportunity for an early-career engineer to work on enterprise-grade platforms and microservice-based architectures.


Key Responsibilities

·         3-5 years of experience with SharePoint Online and/or enterprise CMS platforms.

·         Familiarity with Contentful or other headless CMS solutions is a strong plus.

·         Hands-on experience with JavaSpring Boot, and relational databases (e.g., PostgreSQL).

·         Exposure to KafkaElasticsearch, or similar distributed technologies is desirable.

·         Solid problem-solving and communication skills with an eagerness to learn.

 

What would you do here

SharePoint to Contentful Migration | Backend + CMS Integration

 

·         Assist in maintaining and enhancing SharePoint Online content and features during the transition period.

·         Support the migration of pages, documents, and metadata from SharePoint to Contentful.

·         Contribute to the design and development of backend services that integrate with Contentful using Java, Spring Boot, and REST APIs.

·         Write reusable services for content delivery, search indexing (via Elasticsearch), and event processing (via Kafka).

·         Help develop APIs for CMS-based applications that interact with PostgreSQL databases.

·         Troubleshoot CMS-related issues and support testing efforts during platform migration.


Read more
Genspark

at Genspark

2 candid answers
Agency job
via hirezyai by HR Hirezyai
Bengaluru (Bangalore), Chennai, Coimbatore
5 - 9 yrs
₹9L - ₹25L / yr
Apache Kafka
Apache
MLOps
skill iconAmazon Web Services (AWS)

The candidate should have extensive experience in designing and developing scalable data pipelines and real-time data processing solutions. As a key member of the team, the Senior Data Engineer will play a critical role in building end-to-end data workflows, supporting machine learning model deployment, and driving MLOps practices in a fast-paced, agile environment. Strong expertise in Apache Kafka, Apache Flink, AWS SageMaker, and Terraform is essential. Additional experience with infrastructure automation and CI/CD for ML models is a significant advantage.

Key Responsibilities

  1. Design, develop, and maintain high-performance ETL and real-time data pipelines using Apache Kafka and Apache Flink.
  2. Build scalable and automated MLOps pipelines for training, validation, and deployment of models using AWS SageMaker and associated services.
  3. Implement and manage Infrastructure as Code (IaC) using Terraform to provision and manage AWS environments.
  4. Collaborate with data scientists, ML engineers, and DevOps teams to streamline model deployment workflows and ensure reliable production delivery.
  5. Optimize data storage and retrieval strategies for large-scale structured and unstructured datasets.
  6. Develop data transformation logic and integrate data from various internal and external sources into data lakes and warehouses.
  7. Monitor, troubleshoot, and enhance performance of data systems in a cloud-native, fast-evolving production setup.
  8. Ensure adherence to data governance, privacy, and security standards across all data handling activities.
  9. Document data engineering solutions and workflows to facilitate cross-functional understanding and ongoing maintenance.


Read more
IT service Based
Gurugram, Bengaluru (Bangalore)
5 - 8 yrs
₹7L - ₹15L / yr
skill iconJava
Multithreading
skill iconSpring Boot
Microservices
Apache Kafka
+1 more

Looking for Java Developer for Gurugram and Bangalore Location with 5+ years of experience win Java + Microservices , Multithreading , Spring boot , Kafka and any MQ Series

Read more
AI Powered Logistics Company

AI Powered Logistics Company

Agency job
via Recruiting Bond by Pavan Kumar
Bengaluru (Bangalore)
5 - 8 yrs
₹25L - ₹32L / yr
skill iconNodeJS (Node.js)
NOSQL Databases
SQL
skill iconMongoDB
RabbitMQ
+19 more

Job Title: Backend Engineer - NodeJS, NestJS, and Python

Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)


About the role:

We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.


What You'll Do 🛠️ 

  • Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
  • Develop and integrate RESTful APIs and microservices to support scalable systems.
  • Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
  • Collaborate with cross-functional teams including frontend, DevOps, and QA.
  • Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
  • Monitor and troubleshoot production systems, ensuring high availability and performance.
  • Implement security and data protection best practices.


What You'll Bring 💼 

  • 4 to 6 years of professional experience as a backend developer.
  • Strong proficiency in Node.js and NestJS framework.
  • Good hands-on experience with Python (Django/Flask experience is a plus).
  • Solid understanding of relational and non-relational databases.
  • Proficient in writing complex NoSQL queries and SQL queries
  • Experience with microservices architecture and distributed systems.
  • Familiarity with version control systems like Git.
  • Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
  • Excellent problem-solving skills and a collaborative mindset.

 

Bonus Points ➕ 

  • Experience with CI/CD pipelines.
  • Exposure to cloud platforms like AWS, GCP or Azure.
  • Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)


Why this role matters

You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Read more
AI Powered Logistics Company

AI Powered Logistics Company

Agency job
via Recruiting Bond by Pavan Kumar
Bengaluru (Bangalore)
2 - 4 yrs
₹12L - ₹15L / yr
skill iconNodeJS (Node.js)
skill iconFlask
RESTful APIs
skill iconMongoDB
SQL
+13 more

Job Title: Backend Engineer - NodeJS, NestJS, and Python

Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)


About the role:

We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.


What You'll Do 🛠️ 

  • Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
  • Develop and integrate RESTful APIs and microservices to support scalable systems.
  • Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
  • Collaborate with cross-functional teams including frontend, DevOps, and QA.
  • Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
  • Monitor and troubleshoot production systems, ensuring high availability and performance.
  • Implement security and data protection best practices.


What You'll Bring 💼 

  • 4 to 6 years of professional experience as a backend developer.
  • Strong proficiency in Node.js and NestJS framework.
  • Good hands-on experience with Python (Django/Flask experience is a plus).
  • Solid understanding of relational and non-relational databases.
  • Proficient in writing complex NoSQL queries and SQL queries
  • Experience with microservices architecture and distributed systems.
  • Familiarity with version control systems like Git.
  • Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
  • Excellent problem-solving skills and a collaborative mindset.

 

Bonus Points ➕ 

  • Experience with CI/CD pipelines.
  • Exposure to cloud platforms like AWS, GCP or Azure.
  • Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)


Why this role matters

You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Rashmi SR
Posted by Rashmi SR
Bengaluru (Bangalore)
4 - 9 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Apache Kafka
Data Structures
Algorithms
+2 more

Java Developer – Job Description Wissen Technology is now hiring for a Java Developer - Bangalore with hands-on experience in Core Java, algorithms, data structures, multithreading and SQL. We are solving complex technical problems in the industry and need talented software engineers to join our mission and be a part of a global software development team. A brilliant opportunity to become a part of a highly motivated and expert team which has made a mark as a high-end technical consulting. Required Skills: • Exp. - 4 to 7 years. • Experience in Core Java and Spring Boot. • Extensive experience in developing enterprise-scale applications and systems. Should possess good architectural knowledge and be aware of enterprise application design patterns. • Should have the ability to analyze, design, develop and test complex, low-latency client facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-performance server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills. • Strong interpersonal, communication and analytical skills. • Should have the ability to express their design ideas and thoughts. About Wissen Technology: Wissen Technology is a niche global consulting and solutions company that brings unparalleled domain expertise in Banking and Finance, Telecom and Startups. Wissen Technology is a part of Wissen Group and was established in the year 2015. Wissen has offices in the US, India, UK, Australia, Mexico, and Canada, with best-in-class infrastructure and development facilities. Wissen has successfully delivered projects worth $1 Billion for more than 25 of the Fortune 500 companies. The Wissen Group overall includes more than 4000 highly skilled professionals. Wissen Technology provides exceptional value in mission critical projects for its clients, through thought leadership, ownership, and assured on-time deliveries that are always ‘first time right’. Our team consists of 1200+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world. Wissen Technology offers an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation. We have been certified as a Great Place to Work® for two consecutive years (2020-2022) and voted as the Top 20 AI/ML vendor by CIO Insider. 

Read more
Deqode

at Deqode

1 recruiter
Alisha Das
Posted by Alisha Das
Bengaluru (Bangalore), Hyderabad, Delhi, Gurugram, Noida, Ghaziabad, Faridabad
5 - 7 yrs
₹10L - ₹25L / yr
Microsoft Windows Azure
Data engineering
skill iconPython
Apache Kafka

Role Overview:

We are seeking a Senior Software Engineer (SSE) with strong expertise in Kafka, Python, and Azure Databricks to lead and contribute to our healthcare data engineering initiatives. This role is pivotal in building scalable, real-time data pipelines and processing large-scale healthcare datasets in a secure and compliant cloud environment.

The ideal candidate will have a solid background in real-time streaming, big data processing, and cloud platforms, along with strong leadership and stakeholder engagement capabilities.

Key Responsibilities:

  • Design and develop scalable real-time data streaming solutions using Apache Kafka and Python.
  • Architect and implement ETL/ELT pipelines using Azure Databricks for both structured and unstructured healthcare data.
  • Optimize and maintain Kafka applications, Python scripts, and Databricks workflows to ensure performance and reliability.
  • Ensure data integrity, security, and compliance with healthcare standards such as HIPAA and HITRUST.
  • Collaborate with data scientists, analysts, and business stakeholders to gather requirements and translate them into robust data solutions.
  • Mentor junior engineers, perform code reviews, and promote engineering best practices.
  • Stay current with evolving technologies in cloud, big data, and healthcare data standards.
  • Contribute to the development of CI/CD pipelines and containerized environments (Docker, Kubernetes).

Required Skills & Qualifications:

  • 4+ years of hands-on experience in data engineering roles.
  • Strong proficiency in Kafka (including Kafka Streams, Kafka Connect, Schema Registry).
  • Proficient in Python for data processing and automation.
  • Experience with Azure Databricks (or readiness to ramp up quickly).
  • Solid understanding of cloud platforms, with a preference for Azure (AWS/GCP is a plus).
  • Strong knowledge of SQL and NoSQL databases; data modeling for large-scale systems.
  • Familiarity with containerization tools like Docker and orchestration using Kubernetes.
  • Exposure to CI/CD pipelines for data applications.
  • Prior experience with healthcare datasets (EHR, HL7, FHIR, claims data) is highly desirable.
  • Excellent problem-solving abilities and a proactive mindset.
  • Strong communication and interpersonal skills to work in cross-functional teams.


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Robin Silverster
Posted by Robin Silverster
Bengaluru (Bangalore)
5 - 8 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Apache Kafka
Microservices
Multithreading
+2 more

Java Developer – Job Description Wissen Technology is now hiring for a Java Developer - Bangalore with hands-on experience in Core Java, algorithms, data structures, multithreading and SQL. We are solving complex technical problems in the industry and need talented software engineers to join our mission and be a part of a global software development team. A brilliant opportunity to become a part of a highly motivated and expert team which has made a mark as a high-end technical consulting. Required Skills: • Exp. - 4 to 7 years. • Experience in Core Java and Spring Boot. • Extensive experience in developing enterprise-scale applications and systems. Should possess good architectural knowledge and be aware of enterprise application design patterns. • Should have the ability to analyze, design, develop and test complex, low-latency client facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-performance server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills. • Strong interpersonal, communication and analytical skills. • Should have the ability to express their design ideas and thoughts. About Wissen Technology: Wissen Technology is a niche global consulting and solutions company that brings unparalleled domain expertise in Banking and Finance, Telecom and Startups. Wissen Technology is a part of Wissen Group and was established in the year 2015. Wissen has offices in the US, India, UK, Australia, Mexico, and Canada, with best-in-class infrastructure and development facilities. Wissen has successfully delivered projects worth $1 Billion for more than 25 of the Fortune 500 companies. The Wissen Group overall includes more than 4000 highly skilled professionals. Wissen Technology provides exceptional value in mission critical projects for its clients, through thought leadership, ownership, and assured on-time deliveries that are always ‘first time right’. Our team consists of 1200+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world. Wissen Technology offers an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation. We have been certified as a Great Place to Work® for two consecutive years (2020-2022) and voted as the Top 20 AI/ML vendor by CIO Insider. 

Read more
QAgile Services

at QAgile Services

1 recruiter
Radhika Chotai
Posted by Radhika Chotai
Bengaluru (Bangalore)
3 - 8 yrs
₹17L - ₹25L / yr
skill iconMongoDB
skill iconPython
skill iconFlask
skill iconDjango
Windows Azure
+4 more

Employment type- Contract basis


Key Responsibilities

  • Design, develop, and maintain scalable data pipelines using PySpark and distributed computing frameworks.
  • Implement ETL processes and integrate data from structured and unstructured sources into cloud data warehouses.
  • Work across Azure or AWS cloud ecosystems to deploy and manage big data workflows.
  • Optimize performance of SQL queries and develop stored procedures for data transformation and analytics.
  • Collaborate with Data Scientists, Analysts, and Business teams to ensure reliable data availability and quality.
  • Maintain documentation and implement best practices for data architecture, governance, and security.

⚙️ Required Skills

  • Programming: Proficient in PySpark, Python, and SQL, MongoDB
  • Cloud Platforms: Hands-on experience with Azure Data Factory, Databricks, or AWS Glue/Redshift.
  • Data Engineering Tools: Familiarity with Apache Spark, Kafka, Airflow, or similar tools.
  • Data Warehousing: Strong knowledge of designing and working with data warehouses like Snowflake, BigQuery, Synapse, or Redshift.
  • Data Modeling: Experience in dimensional modeling, star/snowflake schema, and data lake architecture.
  • CI/CD & Version Control: Exposure to Git, Terraform, or other DevOps tools is a plus.

🧰 Preferred Qualifications

  • Bachelor's or Master's in Computer Science, Engineering, or related field.
  • Certifications in Azure/AWS are highly desirable.
  • Knowledge of business intelligence tools (Power BI, Tableau) is a bonus.
Read more
Deqode

at Deqode

1 recruiter
Apoorva Jain
Posted by Apoorva Jain
Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Pune, Bengaluru (Bangalore), Mumbai, Nagpur, Ahmedabad, Kochi (Cochin), Chennai
6 - 11 yrs
₹4L - ₹15L / yr
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka
Spring

Job Description:

We are looking for a Senior Java Developer with strong expertise in Apache Kafka and backend systems. The ideal candidate will have hands-on experience in Java (8/11+), Spring Boot, and building scalable, real-time data pipelines using Kafka.

Key Responsibilities:

  • Develop and maintain backend services using Java and Spring Boot
  • Design and implement Kafka-based messaging and streaming solutions
  • Optimize Kafka performance (topics, partitions, consumers)
  • Collaborate with cross-functional teams to deliver scalable microservices
  • Ensure code quality and maintain best practices in a distributed environment

Required Skills:

  • 6+ years in Java development
  • 3+ years of hands-on Kafka experience (producers, consumers, streams)
  • Strong knowledge of Spring Boot, REST APIs, and microservices
  • Familiarity with Kafka Connect, Schema Registry, and stream processing
  • Experience with containerization (Docker), CI/CD, and cloud platforms (AWS/GCP/Azure)


Read more
KJBN labs

at KJBN labs

2 candid answers
sakthi ganesh
Posted by sakthi ganesh
Bengaluru (Bangalore)
4 - 7 yrs
₹10L - ₹30L / yr
Hadoop
Apache Kafka
Spark
skill iconPython
skill iconJava
+8 more

Senior Data Engineer Job Description

Overview

The Senior Data Engineer will design, develop, and maintain scalable data pipelines and

infrastructure to support data-driven decision-making and advanced analytics. This role requires deep

expertise in data engineering, strong problem-solving skills, and the ability to collaborate with

cross-functional teams to deliver robust data solutions.

Key Responsibilities


Data Pipeline Development: Design, build, and optimize scalable, secure, and reliable data

pipelines to ingest, process, and transform large volumes of structured and unstructured data.

Data Architecture: Architect and maintain data storage solutions, including data lakes, data

warehouses, and databases, ensuring performance, scalability, and cost-efficiency.

Data Integration: Integrate data from diverse sources, including APIs, third-party systems,

and streaming platforms, ensuring data quality and consistency.

Performance Optimization: Monitor and optimize data systems for performance, scalability,

and cost, implementing best practices for partitioning, indexing, and caching.

Collaboration: Work closely with data scientists, analysts, and software engineers to

understand data needs and deliver solutions that enable advanced analytics, machine

learning, and reporting.

Data Governance: Implement data governance policies, ensuring compliance with data

security, privacy regulations (e.g., GDPR, CCPA), and internal standards.

Automation: Develop automated processes for data ingestion, transformation, and validation

to improve efficiency and reduce manual intervention.

Mentorship: Guide and mentor junior data engineers, fostering a culture of technical

excellence and continuous learning.

Troubleshooting: Diagnose and resolve complex data-related issues, ensuring high

availability and reliability of data systems.

Required Qualifications

Education: Bachelor’s or Master’s degree in Computer Science, Engineering, Data Science,

or a related field.

Experience: 5+ years of experience in data engineering or a related role, with a proven track

record of building scalable data pipelines and infrastructure.

Technical Skills:

Proficiency in programming languages such as Python, Java, or Scala.

Expertise in SQL and experience with NoSQL databases (e.g., MongoDB, Cassandra).

Strong experience with cloud platforms (e.g., AWS, Azure, GCP) and their data services

(e.g., Redshift, BigQuery, Snowflake).

Hands-on experience with ETL/ELT tools (e.g., Apache Airflow, Talend, Informatica) and

data integration frameworks.

Familiarity with big data technologies (e.g., Hadoop, Spark, Kafka) and distributed

systems.

Knowledge of containerization and orchestration tools (e.g., Docker, Kubernetes) is a

plus.

Soft Skills:

Excellent problem-solving and analytical skills.

Strong communication and collaboration abilities.

Ability to work in a fast-paced, dynamic environment and manage multiple priorities.

Certifications (optional but preferred): Cloud certifications (e.g., AWS Certified Data Analytics,

Google Professional Data Engineer) or relevant data engineering certifications.

Preferred Qualifica

Experience with real-time data processing and streaming architectures.

Familiarity with machine learning pipelines and MLOps practices.

Knowledge of data visualization tools (e.g., Tableau, Power BI) and their integration with data

pipelines.

Experience in industries with high data complexity, such as finance, healthcare, or

e-commerce.

Work Environment

Location: Hybrid/Remote/On-site (depending on company policy).

Team: Collaborative, cross-functional team environment with data scientists, analysts, and

business stakeholders.

Hours: Full-time, with occasional on-call responsibilities for critical data systems.

Read more
NeoGenCode Technologies Pvt Ltd
Akshay Patil
Posted by Akshay Patil
Hyderabad, Bengaluru (Bangalore), Mumbai, Pune, Gurugram, Chennai
4 - 10 yrs
₹5L - ₹20L / yr
Red Hat PAM Developer
Red Hat Process Automation Manager (PAM)
jBPM
JBPM (Java Business Process Management)
BPMN 2.0 (Business Process Model and Notation)
+6 more

Job Title : Red Hat PAM Developer

Experience Required :

  • Relevant Experience : 4+ Years
  • Total Experience : Up to 8 Years

No. of Positions : 4

Work Locations : Hyderabad / Bangalore / Mumbai / Pune / Gurgaon / Chennai

Work Mode : Hybrid

Work Timings : 1:00 PM to 10:00 PM IST

Interview Mode : Virtual

Interview Rounds : 2


Mandatory Skills :

  • Excellent communication skills – must be comfortable in client-facing roles
  • Red Hat Process Automation Manager (PAM)
  • JBPM (Java Business Process Management)
  • BPMN 2.0 (Business Process Model and Notation) – low-code platform
  • DMN (Decision Model and Notation) – business processes and rules
  • Spring Boot
  • JavaScript

Good-to-Have Skills :

  • Red Hat Fuse
  • Apache Kafka
  • Apache Camel
Read more
Wissen Technology

at Wissen Technology

4 recruiters
Robin Silverster
Posted by Robin Silverster
Bengaluru (Bangalore)
11 - 18 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
Multithreading
Data Structures
+3 more

Java Developer – Job Description


Wissen Technology is now hiring for a Java Developer - Bangalore with hands-on experience in Core Java, algorithms, data structures, multithreading and SQL. We are solving complex technical problems in the industry and need talented software engineers to join our mission and be a part of a global software development team. A brilliant opportunity to become a part of a highly motivated and expert team which has made a mark as a high-end technical consulting. Required Skills: • Exp. - 5 to 12 years. • Experience in Core Java and Spring Boot. • Extensive experience in developing enterprise-scale applications and systems. Should possess good architectural knowledge and be aware of enterprise application design patterns. • Should have the ability to analyze, design, develop and test complex, low-latency client facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-performance server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills. • Strong interpersonal, communication and analytical skills. • Should have the ability to express their design ideas and thoughts. About Wissen Technology: Wissen Technology is a niche global consulting and solutions company that brings unparalleled domain expertise in Banking and Finance, Telecom and Startups. Wissen Technology is a part of Wissen Group and was established in the year 2015. Wissen has offices in the US, India, UK, Australia, Mexico, and Canada, with best-in-class infrastructure and development facilities. Wissen has successfully delivered projects worth $1 Billion for more than 25 of the Fortune 500 companies. The Wissen Group overall includes more than 4000 highly skilled professionals. Wissen Technology provides exceptional value in mission critical projects for its clients, through thought leadership, ownership, and assured on-time deliveries that are always ‘first time right’. Our team consists of 1200+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world. Wissen Technology offers an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation. We have been certified as a Great Place to Work® for two consecutive years (2020-2022) and voted as the Top 20 AI/ML vendor by CIO Insider. 

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Seema Srivastava
Posted by Seema Srivastava
Bengaluru (Bangalore), Mumbai
5 - 10 yrs
Best in industry
skill iconJava
06692
Microservices
skill iconAmazon Web Services (AWS)
Apache Kafka
+1 more

Job Description: We are looking for a talented and motivated Software Engineer with

expertise in both Windows and Linux operating systems and solid experience in Java

technologies. The ideal candidate should be proficient in data structures and algorithms, as

well as frameworks like Spring MVC, Spring Boot, and Hibernate. Hands-on experience

working with MySQL databases is also essential for this role.


Responsibilities:

● Design, develop, test, and maintain software applications using Java technologies.

● Implement robust solutions using Spring MVC, Spring Boot, and Hibernate frameworks.

● Develop and optimize database operations with MySQL.

● Analyze and solve complex problems by applying knowledge of data structures and

algorithms.

● Work with both Windows and Linux environments to develop and deploy solutions.

● Collaborate with cross-functional teams to deliver high-quality products on time.

● Ensure application security, performance, and scalability.

● Maintain thorough documentation of technical solutions and processes.

● Debug, troubleshoot, and upgrade legacy systems when required.

Requirements:

● Operating Systems: Expertise in Windows and Linux environments.

● Programming Languages & Technologies: Strong knowledge of Java (Core Java, Java 8+).

● Frameworks: Proficiency in Spring MVC, Spring Boot, and Hibernate.

● Algorithms and Data Structures: Good understanding and practical application of DSA

concepts.

● Databases: Experience with MySQL – writing queries, stored procedures, and performance

tuning.

● Version Control Systems: Experience with tools like Git.

● Deployment: Knowledge of CI/CD pipelines and tools such as Jenkins, Docker (optional)

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Deepa Shankar
Posted by Deepa Shankar
Bengaluru (Bangalore)
4 - 9 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka
Multitasking
+2 more

Required Skillset

• Experience in Core Java 1.8 and above, Data Structures, OOPS, Multithreading, Algorithms, Collections, System Design, Unix/Linux. • Possess good architectural knowledge and be aware of enterprise application design patterns. • Should be able to analyze, design, develop and test complex, low-latency client-facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-volume server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills in Java. • Strong interpersonal, communication and analytical skills. • Should be able to express their design ideas and thoughts.


Job Brief-

• Understand product requirements and come up with solution approaches. • Build and enhance large scale domain centric applications. • Deploy high quality deliverables into production adhering to the security, compliance and SDLC guidelines.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
VenkataRamanan S
Posted by VenkataRamanan S
Bengaluru (Bangalore)
3 - 8 yrs
Best in industry
06692
Microservices
Apache Kafka

Key Responsibilities:

  • Design, develop, and maintain robust and scalable backend applications using Core Java and Spring Boot.
  • Build and manage microservices-based architectures and ensure smooth inter-service communication.
  • Integrate and manage real-time data streaming using Apache Kafka.
  • Write clean, maintainable, and efficient code following best practices.
  • Collaborate with cross-functional teams including QA, DevOps, and product management.
  • Participate in code reviews and provide constructive feedback.
  • Troubleshoot, debug, and optimize applications for performance and scalability.

Required Skills:

  • Strong knowledge of Core Java (Java 8 or above).
  • Hands-on experience with Spring Boot and the Spring ecosystem (Spring MVC, Spring Data, Spring Security).
  • Experience in designing and developing RESTful APIs.
  • Solid understanding of Microservices architecture and related patterns.
  • Practical experience with Apache Kafka for real-time data processing.
  • Familiarity with SQL/NoSQL databases such as MySQL, PostgreSQL, or MongoDB.
  • Good understanding of CI/CD tools and practices.
  • Knowledge of containerization tools like Docker is a plus.
  • Strong problem-solving skills and attention to detail.


Read more
Cloudesign Technology Solutions
Anshul Saxena
Posted by Anshul Saxena
Noida, Hyderabad, Bengaluru (Bangalore)
5 - 8 yrs
₹25L - ₹30L / yr
Apache Kafka
Kafka Cluster
Schema Registry
Streaming
Kafka
+1 more

Job Title: Tech Lead and SSE – Kafka, Python, and Azure Databricks (Healthcare Data Project)

Experience: 4 to 12 years


Role Overview:

We are looking for a highly skilled Tech Lead with expertise in Kafka, Python, and Azure Databricks (preferred) to drive our healthcare data engineering projects. The ideal candidate will have deep experience in real-time data streaming, cloud-based data platforms, and large-scale data processing. This role requires strong technical leadership, problem-solving abilities, and the ability to collaborate with cross-functional teams.


Key Responsibilities:

  • Lead the design, development, and implementation of real-time data pipelines using Kafka, Python, and Azure Databricks.
  • Architect scalable data streaming and processing solutions to support healthcare data workflows.
  • Develop, optimize, and maintain ETL/ELT pipelines for structured and unstructured healthcare data.
  • Ensure data integrity, security, and compliance with healthcare regulations (HIPAA, HITRUST, etc.).
  • Collaborate with data engineers, analysts, and business stakeholders to understand requirements and translate them into technical solutions.
  • Troubleshoot and optimize Kafka streaming applications, Python scripts, and Databricks workflows.
  • Mentor junior engineers, conduct code reviews, and ensure best practices in data engineering.
  • Stay updated with the latest cloud technologies, big data frameworks, and industry trends.


Required Skills & Qualifications:

  • 4+ years of experience in data engineering, with strong proficiency in Kafka and Python.
  • Expertise in Kafka Streams, Kafka Connect, and Schema Registry for real-time data processing.
  • Experience with Azure Databricks (or willingness to learn and adopt it quickly).
  • Hands-on experience with cloud platforms (Azure preferred, AWS or GCP is a plus).
  • Proficiency in SQL, NoSQL databases, and data modeling for big data processing.
  • Knowledge of containerization (Docker, Kubernetes) and CI/CD pipelines for data applications.
  • Experience working with healthcare data (EHR, claims, HL7, FHIR, etc.) is a plus.
  • Strong analytical skills, problem-solving mindset, and ability to lead complex data projects.
  • Excellent communication and stakeholder management skills.


Read more
Talent Pro
Bengaluru (Bangalore)
4 - 8 yrs
₹26L - ₹35L / yr
skill iconJava
skill iconSpring Boot
Google Cloud Platform (GCP)
Distributed Systems
Microservices
+3 more

Role & Responsibilities

Responsible for ensuring that the architecture and design of the platform remains top-notch with respect to scalability, availability, reliability and maintainability

Act as a key technical contributor as well as a hands-on contributing member of the team.

Own end-to-end availability and performance of features, driving rapid product innovation while ensuring a reliable service.

Working closely with the various stakeholders like Program Managers, Product Managers, Reliability and Continuity Engineering(RCE) team, QE team to estimate and execute features/tasks independently.

Maintain and drive tech backlog execution for non-functional requirements of the platform required to keep the platform resilient

Assist in release planning and prioritization based on technical feasibility and engineering constraints

A zeal to continually find new ways to improve architecture, design and ensure timely delivery and high quality.

Read more
Trika Tech
bhagya a
Posted by bhagya a
Bengaluru (Bangalore), Coimbatore
7 - 8 yrs
₹12L - ₹25L / yr
skill iconNodeJS (Node.js)
AWS Lambda
Apache Kafka
skill iconKubernetes

WHO WE ARE


We are a team of digital practitioners with roots stretching back to the earliest days of online commerce, who dedicate themselves to serving our client companies.

We’ve seen the advancements first-hand over the last 25 years and believe our experiences allow us to innovate. Utilizing cutting-edge technology and providing bespoke, innovative services, we believe we can help you stay ahead of the curve.

We take a holistic view of digital strategy. Our approach to transformation is based on conscious Intent to delight customers through continuous Insight and creative Innovation with an enduring culture of problem-solving.

We bring every element together to create innovative, high-performing commerce experiences for enterprise B2C, B2B, D2C and Marketplace brands across industries. From mapping out business and functional requirements, to developing the infrastructure to optimize traditionally fragmented processes, we help you create integrated, future-proofed commerce solutions.

 

WHAT YOU’LL BE DOING

As part of our team, you'll play a key role in building and evolving our Integration Platform as a Service (iPaaS) solution. This platform empowers our clients to seamlessly connect systems, automate workflows, and scale integrations with modern cloud-native tools.

Here’s what your day-to-day will look like:

  • Designing and Building Integrations
  • Collaborate with clients to understand integration needs and build scalable, event-driven solutions using Apache Kafka, AWS Lambda, API Gateway, and EventBridge.

  • Cloud-Native Development
  • Develop and deploy microservices and serverless functions using TypeScript (Node.js), hosted on Kubernetes (EKS) and fully integrated with core AWS services like S3, SQS, and SNS.

  • Managing Data Pipelines
  • Build robust data flows and streaming pipelines using Kafka and NoSQL databases like MongoDB, ensuring high availability and fault tolerance.

  • Client Collaboration
  • Work directly with customers to gather requirements, design integration patterns, and provide guidance on best practices for cloud-native architectures.

  • Driving Platform Evolution
  • Contribute to the ongoing improvement of our iPaaS platform—enhancing observability, scaling capabilities, and CI/CD processes using modern DevOps practices.


WHAT WE NEED IN YOU


  • Solid Experience in Apache Kafka for data streaming and event-driven systems
  • Production experience with Kubernetes (EKS) and containerized deployments
  • Deep knowledge of AWS, including S3, EC2, SQS, SNS, EventBridge, Lambda
  • Proficient in TypeScript (Node.js environment)
  • Experience with MongoDB or other NoSQL databases
  • Familiarity with microservices architecture, async messaging, and DevOps practices
  • AWS Certification (e.g., Solutions Architect or Developer Associate) is a plus


Qualification

  • Graduate - BE / Btech or equivalent.
  • 5 to 8 years of experience
  • Self motivated and quick learner with excellent problem solving skills. 
  • A good team player with nice communication skills.
  • Energy and real passion to work in a startup environment.

Visit our website - https://www.trikatechnologies.com



Read more
NeoGenCode Technologies Pvt Ltd
Akshay Patil
Posted by Akshay Patil
Bengaluru (Bangalore), Pune, Chennai
10 - 20 yrs
₹30L - ₹60L / yr
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka
skill iconAmazon Web Services (AWS)
+8 more

📍 Position : Java Architect

📅 Experience : 10 to 15 Years

🧑‍💼 Open Positions : 3+

📍 Work Location : Bangalore, Pune, Chennai

💼 Work Mode : Hybrid

📅 Notice Period : Immediate joiners preferred; up to 1 month maximum

🔧 Core Responsibilities :

  • Lead architecture design and development for scalable enterprise-level applications.
  • Own and manage all aspects of technical development and delivery.
  • Define and enforce best coding practices, architectural guidelines, and development standards.
  • Plan and estimate the end-to-end technical scope of projects.
  • Conduct code reviews, ensure CI/CD, and implement TDD/BDD methodologies.
  • Mentor and lead individual contributors and small development teams.
  • Collaborate with cross-functional teams, including DevOps, Product, and QA.
  • Engage in high-level and low-level design (HLD/LLD), solutioning, and cloud-native transformations.

🛠️ Required Technical Skills :

  • Strong hands-on expertise in Java, Spring Boot, Microservices architecture
  • Experience with Kafka or similar messaging/event streaming platforms
  • Proficiency in cloud platformsAWS and Azure (must-have)
  • Exposure to frontend technologies (nice-to-have)
  • Solid understanding of HLD, system architecture, and design patterns
  • Good grasp of DevOps concepts, Docker, Kubernetes, and Infrastructure as Code (IaC)
  • Agile/Lean development, Pair Programming, and Continuous Integration practices
  • Polyglot mindset is a plus (Scala, Golang, Python, etc.)

🚀 Ideal Candidate Profile :

  • Currently working in a product-based environment
  • Already functioning as an Architect or Principal Engineer
  • Proven track record as an Individual Contributor (IC)
  • Strong engineering fundamentals with a passion for scalable software systems
  • No compromise on code quality, craftsmanship, and best practices

🧪 Interview Process :

  1. Round 1: Technical pairing round
  2. Rounds 2 & 3: Technical rounds with panel (code pairing + architecture)
  3. Final Round: HR and offer discussion
Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort