Cutshort logo
Apache kafka jobs

50+ Apache Kafka Jobs in India

Apply to 50+ Apache Kafka Jobs on CutShort.io. Find your next job, effortlessly. Browse Apache Kafka Jobs and apply today!

icon
Tamashalive

at Tamashalive

1 candid answer
Aparna. Majumder
Posted by Aparna. Majumder
Bengaluru (Bangalore)
5 - 7 yrs
₹35L - ₹50L / yr
skill iconNodeJS (Node.js)
skill iconExpress
skill iconGo Programming (Golang)
skill iconJava
Apache Kafka
+1 more

About Tamasha

Tamasha is building next-generation live, social, and real-time consumer

experiences across gaming, audio, video, and interactive platforms. Our backend

systems power millions of real-time interactions, low-latency data flows, and high

throughput event-driven services.

Role Overview

As a Software Development Engineer III (Backend), you will operate as a senior

individual contributor owning mission-critical backend systems. You will design

scalable architectures, solve complex distributed systems problems, and influence

engineering standards across teams.

Roles & Responsibilities

 Own backend systems end-to-end, from design and implementation to

production reliability and long-term evolution.

Design scalable, fault-tolerant, and low-latency backend architectures for real

time and event-driven use cases.

Lead system design decisions including data models, service boundaries, async

workflows, and scalability strategies.

Build and operate high-throughput services handling millions of events, requests,

or messages per day.

 Own reliability metrics such as p95/p99 latency, uptime, error budgets, and

participate in incident resolution and post-mortems.

Drive architectural migrations including monolith to microservices and

synchronous to asynchronous flows.

 Collaborate with product, frontend, and data teams to translate complex

requirements into robust backend solutions.

Improve developer productivity by contributing to internal platforms, tooling,

CI/CD pipelines, and coding standards.

Mentor engineers through design reviews, code reviews, and technical guidance

without direct people management.

Proactively identify performance bottlenecks, cost inefficiencies, and scalability

risks and resolve them at the system level.

Tech Stack

Backend: Node.js (NestJS / Express), Golang, Java

Databases: PostgreSQL, MySQL, MongoDB, Redis

Messaging & Async: Kafka, SQS, RabbitMQ, Pub/Sub

Real-time: WebSockets, Server-Sent Events

Cloud & Infra: AWS / GCP, Docker, Kubernetes

Observability: Prometheus, Grafana, OpenTelemetry

What We’re Looking For

• 5+ years of backend engineering experience

• Proven ownership of production systems at scale

• Strong understanding of distributed systems and async architectures

• Experience handling production incidents and system reliability

• Ability to independently design and evolve backend systems


Read more
Tamashalive

at Tamashalive

1 candid answer
Aparna. Majumder
Posted by Aparna. Majumder
Bengaluru (Bangalore)
5 - 7 yrs
₹35L - ₹50L / yr
skill iconNodeJS (Node.js)
skill iconExpress
skill iconGo Programming (Golang)
skill iconJava
Apache Kafka
+1 more

About Tamasha

Tamasha is building next-generation live, social, and real-time consumer

experiences across gaming, audio, video, and interactive platforms. Our backend

systems power millions of real-time interactions, low-latency data flows, and high

throughput event-driven services.

Role Overview

As a Software Development Engineer III (Backend), you will operate as a senior

individual contributor owning mission-critical backend systems. You will design

scalable architectures, solve complex distributed systems problems, and influence

engineering standards across teams.

Roles & Responsibilities

 Own backend systems end-to-end, from design and implementation to

production reliability and long-term evolution.

Design scalable, fault-tolerant, and low-latency backend architectures for real

time and event-driven use cases.

Lead system design decisions including data models, service boundaries, async

workflows, and scalability strategies.

Build and operate high-throughput services handling millions of events, requests,

or messages per day.

 Own reliability metrics such as p95/p99 latency, uptime, error budgets, and

participate in incident resolution and post-mortems.

Drive architectural migrations including monolith to microservices and

synchronous to asynchronous flows.

 Collaborate with product, frontend, and data teams to translate complex

requirements into robust backend solutions.

Improve developer productivity by contributing to internal platforms, tooling,

CI/CD pipelines, and coding standards.

Mentor engineers through design reviews, code reviews, and technical guidance

without direct people management.

Proactively identify performance bottlenecks, cost inefficiencies, and scalability

risks and resolve them at the system level.

Tech Stack

Backend: Node.js (NestJS / Express), Golang, Java

Databases: PostgreSQL, MySQL, MongoDB, Redis

Messaging & Async: Kafka, SQS, RabbitMQ, Pub/Sub

Real-time: WebSockets, Server-Sent Events

Cloud & Infra: AWS / GCP, Docker, Kubernetes

Observability: Prometheus, Grafana, OpenTelemetry

What We’re Looking For

• 5+ years of backend engineering experience

• Proven ownership of production systems at scale

• Strong understanding of distributed systems and async architectures

• Experience handling production incidents and system reliability

• Ability to independently design and evolve backend systems


Read more
PhotonMatters
Human Resource
Posted by Human Resource
Remote only
4 - 13 yrs
₹8L - ₹20L / yr
skill iconPython
ETL
Spark
skill iconAmazon Web Services (AWS)
ELT
+2 more

 

 

 

Job Title: Data Engineer

Experience: 4–14 Years

Work Mode: Remote

Employment Type: Full-Time

 

Position Overview:

We are looking for highly experienced Senior Data Engineers to design, architect, and lead scalable, cloud-based data platforms on AWS. The role involves building enterprise-grade data pipelines, modernizing legacy systems, and developing high-performance scoring engines and analytics solutions and collaborate closely with architecture, analytics, risk, and business teams to deliver secure, reliable, and scalable data solutions.

 

Key Responsibilities:

·      Design and build scalable data pipelines for financial and customer data

·      Build and optimize scoring engines (credit, risk, fraud, customer scoring)

·      Design, develop, and optimize complex ETL/ELT pipelines (batch & real-time)

·      Ensure data quality, governance, reliability, and compliance standards

·      Optimize large-scale data processing using SQL, Spark/PySpark, and cloud technologies

·      Lead cloud data architecture, cost optimization, and performance tuning initiatives

·      Collaborate with Data Science, Analytics, and Product teams to deliver business-ready datasets

·      Mentor junior engineers and establish best practices for data engineering

 

Key Requirements:

·      Strong programming skills in Python and advanced SQL

·      Experience building scalable scoring or rule-based decision engines

·      Hands-on experience with Big Data technologies (Spark/PySpark/Kafka)

·      Strong expertise in designing ETL/ELT pipelines and data modeling

·      Experience with cloud platforms (AWS/Azure) and modern data architectures

·      Solid understanding of data warehousing, data lakes, and performance tuning

·      Knowledge of CI/CD, version control (Git), and production support best practices

Read more
Tarento Group

at Tarento Group

3 candid answers
1 recruiter
Bisman Gill
Posted by Bisman Gill
Bengaluru (Bangalore)
4yrs+
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
Windows Azure
RESTful APIs
+5 more

Job Summary:

We are seeking a highly skilled and self-driven Java Backend Developer with strong experience in designing and deploying scalable microservices using Spring Boot and Azure Cloud. The ideal candidate will have hands-on expertise in modern Java development, containerization, messaging systems like Kafka, and knowledge of CI/CD and DevOps practices.Key Responsibilities:

  • Design, develop, and deploy microservices using Spring Boot on Azure cloud platforms.
  • Implement and maintain RESTful APIs, ensuring high performance and scalability.
  • Work with Java 11+ features including Streams, Functional Programming, and Collections framework.
  • Develop and manage Docker containers, enabling efficient development and deployment pipelines.
  • Integrate messaging services like Apache Kafka into microservice architectures.
  • Design and maintain data models using PostgreSQL or other SQL databases.
  • Implement unit testing using JUnit and mocking frameworks to ensure code quality.
  • Develop and execute API automation tests using Cucumber or similar tools.
  • Collaborate with QA, DevOps, and other teams for seamless CI/CD integration and deployment pipelines.
  • Work with Kubernetes for orchestrating containerized services.
  • Utilize Couchbase or similar NoSQL technologies when necessary.
  • Participate in code reviews, design discussions, and contribute to best practices and standards.

Required Skills & Qualifications:

  • Strong experience in Java (11 or above) and Spring Boot framework.
  • Solid understanding of microservices architecture and deployment on Azure.
  • Hands-on experience with Docker, and exposure to Kubernetes.
  • Proficiency in Kafka, with real-world project experience.
  • Working knowledge of PostgreSQL (or any SQL DB) and data modeling principles.
  • Experience in writing unit tests using JUnit and mocking tools.
  • Experience with Cucumber or similar frameworks for API automation testing.
  • Exposure to CI/CD toolsDevOps processes, and Git-based workflows.

Nice to Have:

  • Azure certifications (e.g., Azure Developer Associate)
  • Familiarity with Couchbase or other NoSQL databases.
  • Familiarity with other cloud providers (AWS, GCP)
  • Knowledge of observability tools (Prometheus, Grafana, ELK)

Soft Skills:

  • Strong problem-solving and analytical skills.
  • Excellent verbal and written communication.
  • Ability to work in an agile environment and contribute to continuous improvement.

Why Join Us:

  • Work on cutting-edge microservice architectures
  • Strong learning and development culture
  • Opportunity to innovate and influence technical decisions
  • Collaborative and inclusive work environment
Read more
Recruiting Bond

at Recruiting Bond

2 candid answers
Pavan Kumar
Posted by Pavan Kumar
Mumbai, Navi Mumbai
10 - 15 yrs
₹55L - ₹80L / yr
Distributed Systems
Systems design
Systems architecture
High-level design
LLD
+77 more

Location: Mumbai, Maharashtra, India

Sector: Technology, Information & Media

Company Size: 500 - 1,000 Employees

Employment: Full-Time, Permanent

Experience: 10 - 14 Years (Engineering Leadership)

Level: Engineering Manager / Group EM


ABOUT THIS MANDATE :


Recruiting Bond has been exclusively retained by one of India's most prominent and well-established digital platform organisations operating at the intersection of Technology, Information, and Media to identify and place an exceptional Engineering Manager who can lead engineering teams through an enterprise-wide AI adoption and digital transformation agenda.


This is a high-impact, hands-on leadership role at the nexus of people, product, and technology. The organisation is executing one of the most ambitious AI transformation programmes in its sector and this Engineering Manager will be a core driver of that change. You will lead multiple squads, own engineering delivery end-to-end, embed AI tooling and practices into the team's DNA, and shape the engineering culture of tomorrow.


We are seeking leaders who code when it matters, who build systems and teams with equal conviction, and who view AI not as a trend but as a fundamental shift in how great software is built.


THE OPPORTUNITY AT A GLANCE :


AI-First Engineering Culture :

  • Own AI adoption across your squads - from LLM tooling integration to automation-first delivery workflows. Make AI a default, not an afterthought.


Hands-On Engineering Leadership :

  • Stay close to the code. Lead architecture reviews, unblock engineers, and set the technical bar - not just the management agenda.


People & Org Builder :

  • Grow engineers into leaders. Build squads of 615 across functions. Drive hiring, career frameworks, and a culture of psychological safety.


KEY RESPONSIBILITIES :


1. Hands-On Technical Engagement :

  • Remain deeply embedded in the technical work participate in design reviews, architecture decisions, and critical code reviews
  • Set and uphold the engineering quality bar : performance benchmarks, security standards, test coverage, and release quality
  • Provide technical direction on backend platform strategy, API design, service decomposition, and data architecture
  • Identify and resolve systemic technical debt and architectural risks across team-owned services
  • Unblock engineers by diving into complex problems debugging, pair programming, and system analysis when it matters
  • Own key technical decisions in collaboration with Tech Leads and Principal Engineers; balance pragmatism with long-term sustainability


2. AI Adoption, Integration & Transformation (2026 Mandate) :

  • Define and execute the team's AI adoption roadmap - from developer tooling to product-facing AI features
  • Champion the integration of GenAI tools (GitHub Copilot, Cursor, Claude, ChatGPT) across the full engineering workflow coding, testing, documentation, incident response
  • Embed LLM-powered capabilities into the product : recommendation engines, intelligent search, conversational interfaces, content generation, and predictive systems
  • Lead evaluation and adoption of AI-assisted SDLC practices : automated code review, AI-generated test suites, intelligent observability, and anomaly detection
  • Partner with Data Science and ML Platform teams to productionise ML models with robust MLOps pipelines
  • Build team literacy in prompt engineering, RAG (Retrieval-Augmented Generation), and AI agent frameworks
  • Create an experimentation culture : run structured AI pilots, measure productivity impact, and scale what works
  • Stay ahead of the AI tooling landscape and advise senior leadership on strategic AI investments and engineering implications


3. People Leadership & Team Development :

  • Lead, manage, and grow squads of 6 - 15 engineers across seniority levels (L2 through L6 / Junior through Staff)
  • Conduct structured 1 : 1s, career growth conversations, and development planning with every direct report
  • Design and execute personalised AI upskilling programmes ensure every engineer develops practical AI fluency by end of 2026
  • Build and maintain a high-performance team culture : clarity of ownership, accountability, fast feedback loops, and psychological safety
  • Drive performance management fairly and rigorously recognise top performers, manage underperformance constructively
  • Lead technical hiring end-to-end : define job requirements, conduct bar-raising interviews, and make data-driven hire decisions
  • Contribute to engineering career frameworks and level definitions in partnership with the VP / Director of Engineering


4. Engineering Delivery & Execution Excellence :

  • Own end-to-end delivery for multiple product squads from planning and scoping through production release and post-launch stability
  • Implement and refine agile delivery frameworks (Scrum, Kanban, Shape Up) calibrated to squad needs and product cadence
  • Drive predictable delivery : maintain healthy sprint velocity, manage WIP limits, and ensure dependency resolution across teams.
  • Establish and own engineering KPIs : DORA metrics (deployment frequency, lead time, MTTR, change failure rate), uptime SLOs, and velocity trends
  • Lead incident management : build blameless post-mortem culture, own RCA processes, and drive systemic reliability improvements
  • Balance technical debt repayment with feature velocity negotiate prioritisation transparently with Product leadership


5. Strategic Leadership & Cross-Functional Influence :

  • Serve as the primary engineering partner for Product, Design, Data, and Business stakeholders translate ambiguity into executable engineering plans
  • Participate in quarterly roadmap planning, capacity forecasting, and OKR definition for engineering teams
  • Represent engineering in leadership forums articulate technical constraints, risks, and opportunities in business terms
  • Contribute to org-wide engineering strategy : platform investments, build-vs-buy decisions, and shared infrastructure priorities
  • Build relationships across geographies (Mumbai HQ + distributed teams) to maintain alignment and delivery cohesion
  • Act as a culture carrier and ambassador for engineering excellence, innovation, and responsible AI use


AI TRANSFORMATION LEADERSHIP 2026 EXPECTATIONS :


In 2026, Engineering Managers at this organisation are expected to be active architects of AI transformation not passive observers. The following outlines the specific AI leadership expectations for this role :


AI Developer Productivity

  • Drive measurable uplift in developer velocity through AI tooling adoption. Target : 30%+ reduction in code review cycle time and 40%+ increase in test coverage automation by Q3 2026.


LLM & GenAI Product Features

  • Own delivery of GenAI-powered product capabilities : intelligent content, semantic search, personalisation, and conversational UX in production, at scale.


AI-Augmented Observability

  • Implement AI-driven monitoring and anomaly detection pipelines. Reduce MTTR by leveraging predictive alerting, intelligent runbooks, and auto-remediation scripts.


Team AI Fluency :

  • Build mandatory AI literacy across all engineering levels.
  • Every engineer understands prompt engineering basics, AI ethics guardrails, and responsible AI deployment practices.


Responsible AI Governance :

  • Partner with Security, Legal, and Data Privacy to ensure all AI deployments meet compliance standards, bias mitigation requirements, and explainability benchmarks.


TECHNOLOGY STACK & DOMAIN FAMILIARITY REQUIRED :


  • Languages: Java/ Go/ Python/ Node.js /PHP /Rust (must be hands-on in at least 2)
  • Cloud: AWS / GCP / Azure (multi-cloud exposure strongly preferred)
  • AI & GenAI: OpenAI / Anthropic / Gemini APIs /LangChain /LlamaIndex / RAG / Vector DBs / GitHub
  • Copilot: Cursor /Hugging Face
  • Containers: Docker /Kubernetes /Helm /Service Mesh (Istio / Linkerd)
  • Databases: PostgreSQL /MongoDB / Redis / Cassandra / Elasticsearch / Pinecone (Vector DB)
  • Messaging: Apache Kafka /RabbitMQ /AWS SQS/SNS /Google Pub/Sub
  • MLOps & DataOps: MLflow /Kubeflow / SageMaker / Vertex AI /Airflow /dbt
  • Observability: Datadog /Prometheus /Grafana /OpenTelemetry / Jaeger /ELK Stack
  • CI/CD & IaC: GitHub Actions ArgoCD / Jenkins / Terraform /Ansible /Backstage (IDP)


QUALIFICATIONS & CANDIDATE PROFILE :

Education :

  • B.E. / B.Tech or M.E. / M.Tech from a Tier-I or Tier-II Institution - CS, IS, ECE, AI/ML streams strongly preferred
  • Demonstrated engineering depth and leadership impact may complement institution pedigree


Experience :

  • 10 to 14 years of progressive engineering experience, with at least 3 years in a formal Engineering Manager or equivalent people-leadership role
  • Proven track record of managing and scaling engineering teams (615+ engineers) in a fast-growing SaaS or digital product environment
  • Hands-on backend engineering background must be able to read, write, and critique production code
  • Direct experience driving AI/ML feature delivery or AI tooling adoption within engineering organisations
  • Exposure across start-up, mid-size, and large-scale product organisations, preferred adaptability is a core requirement
  • Strong CS fundamentals: distributed systems, algorithms, system design, and software architecture
  • Demonstrated career stability minimum of 2 years of average tenure per organisation.


The Ideal Engineering Manager in 2026 :

  • Leads with context, not control, empowers engineers while maintaining accountability and quality
  • Is fluent in both people language and technical language, switches registers naturally with engineers and executives alike
  • Sees AI as a force multiplier for the team, not a threat. Actively experiments with and advocates for AI tooling
  • Measures success by team outcomes, not personal output. Takes pride in what the team ships, not what they build alone
  • Creates feedback loops obsessively between product and engineering, between seniors and juniors, between metrics and decisions
  • Has strong opinions, loosely held, brings conviction to discussions but updates on evidence
  • Invests in engineering excellence as seriously as delivery velocity knows that quality and speed are not opposites


WHY THIS ROLE STANDS APART :


AI Transformation at Scale :

  • Lead one of the most significant AI adoption programmes in India's digital media sector.
  • Our decisions will shape how hundreds of engineers work in 2026 and beyond.


Hands-On & Strategic Balance :

  • A rare EM role that actively encourages technical depth.
  • Stay close to the code while owning the people agenda - the best of both worlds.


Established Platform, Real Scale :

  • 5001,000 engineers, proven product-market fit, and the org maturity to execute.
  • This is not a greenfield startup gamble it is a serious company with serious ambition.


Clear Leadership Growth Path :

  • A visible, direct path toward Director / VP of Engineering.
  • Senior leadership is invested in growing its next generation of technology executives.


Read more
Tradelab Software Private Limited
Mumbai
2 - 5 yrs
₹10L - ₹15L / yr
skill iconPython
FastAPI
RabbitMQ
Apache Kafka
skill iconElastic Search
+1 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.



What You Will Do:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and a deep interest in scalable, low-latency systems.

• You should have 2–5 years of experience in Python-based development and be eager to solve complex performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What We Expect:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 2–5 years of experience in Python and backend development.

• Strong understanding of REST APIssockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/KafkaSQL & NoSQL databasesRedis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Experience with GoLangC/C++Erlang, or Elixir.

• Exposure to trading, fintech, or low-latency systems.

• Familiarity with microservices and CI/CD pipelines.

Read more
Recruiting Bond

at Recruiting Bond

2 candid answers
Pavan Kumar
Posted by Pavan Kumar
Bengaluru (Bangalore)
6 - 12 yrs
₹40L - ₹55L / yr
skill iconPython
skill iconGo Programming (Golang)
skill iconNodeJS (Node.js)
skill iconJava
Distributed Systems
+27 more

NOW HIRING · WORLD-CLASS TALENT Backend Tech Lead (Senior Level Engineering Leadership)

Placed by Recruiting Bond on behalf of a Confidential Digital Platform Leader

📍Location: Bengaluru, India (Hybrid / On-Site)

🏢Sector: Technology, Information & Media

👥Company Size: 500 – 1,000 Employees

💼Employment: Full-Time, Permanent

🎯Experience: 6 – 9 Years (Backend Engineering)

🚀 Level: Tech Lead


ABOUT THIS MANDATE

Recruiting Bond has been exclusively retained by one of India's most well-established digital platform organisations — a company operating at the intersection of Technology, Information, and Media — to identify and place a world-class Backend Tech Lead who can drive a transformational engineering agenda at scale.

This is not an ordinary role. The organisation is executing a high-stakes, large-scale modernisation of its backend infrastructure — migrating from legacy monolithic systems to resilient, cloud-native, AI-augmented distributed architectures that serve millions of concurrent users. The person in this seat will be a core pillar of that transformation.

We are looking exclusively for the top 1% — engineers who think in systems, own outcomes, and lead by example.


THE OPPORTUNITY AT A GLANCE

🏗️ Architecture Ownership

Drive system design decisions across the entire backend platform. Shape the future of distributed, fault-tolerant architecture.

🤖 AI-Augmented Engineering

Embed GenAI and LLM tooling directly into the SDLC. Champion automation-first development practices across squads.

🎓 Engineering Leadership

Mentor and grow the next generation of backend engineers. Lead hiring, reviews, and cross-functional technical alignment.


KEY RESPONSIBILITIES


1. Architecture & Platform Modernisation

  • Lead the full migration of legacy monolithic systems to a scalable, cloud-native microservices architecture
  • Design and own distributed, fault-tolerant backend systems with sub-millisecond SLO targets
  • Architect API-first and event-driven platforms using async messaging patterns (Kafka, Pub/Sub, SQS)
  • Resolve systemic performance bottlenecks, concurrency conflicts, and scalability ceilings
  • Establish backend design standards, coding guidelines, and architectural review processes


2. Distributed Systems Engineering (Production-Grade)

  • Design and implement Webhook reliability frameworks with intelligent retry and exponential backoff strategies
  • Build idempotent, versioned APIs with enterprise-grade rate limiting and throttling controls
  • Implement circuit breakers, bulkheads, and resilience patterns using Resilience4j / Hystrix or equivalents
  • Engineer Dead-Letter Queue (DLQ) strategies and event reprocessing pipelines with guaranteed delivery semantics
  • Apply Saga orchestration and choreography patterns for distributed transaction integrity
  • Execute zero-downtime deployments and canary release strategies with rollback capability
  • Design and enforce multi-region disaster recovery and business continuity protocols


3. AI-Driven Engineering Practices

  • Champion LLM and GenAI adoption as first-class tooling across the software development lifecycle
  • Apply prompt engineering techniques for automated code generation, review, and documentation workflows
  • Utilise AI-assisted debugging, root cause analysis, and predictive performance optimisation
  • Build automation-first pipelines that reduce toil and accelerate delivery velocity
  • Evaluate and integrate emerging AI developer tools into the engineering ecosystem


4. Engineering Leadership & Culture

  • Own backend platforms end-to-end with full accountability across development, stability, and performance
  • Actively mentor, coach, and elevate engineers at all levels (L3–L6) through structured 1:1s and code reviews
  • Drive and lead technical hiring — from designing assessments to final hire decisions
  • Partner with Product, Data, DevOps, and Security stakeholders to align engineering with business objectives
  • Represent the engineering org in cross-functional roadmap planning and architecture decision reviews
  • Foster a culture of technical excellence, psychological safety, and high-velocity delivery


TECHNOLOGY STACK (HANDS-ON PROFICIENCY REQUIRED)

Languages: Java (primary) · Go · Python · Node.js · PHP · Rust

Cloud: AWS · GCP · Azure (Multi-cloud exposure preferred)

Containers: Docker · Kubernetes · Helm · Service Mesh (Istio / Linkerd)

Databases: PostgreSQL · MySQL · MongoDB · Cassandra · Redis · Elasticsearch

Messaging: Apache Kafka · RabbitMQ · AWS SQS/SNS · Google Pub/Sub

Observability: Datadog · Prometheus · Grafana · OpenTelemetry · Jaeger · ELK Stack

CI/CD & IaC: GitHub Actions · Jenkins · ArgoCD · Terraform · Ansible

AI & GenAI: OpenAI / Claude APIs · LangChain · RAG Pipelines · GitHub Copilot · Cursor


QUALIFICATIONS & CANDIDATE PROFILE


Education

  • B.E. / B.Tech or M.E. / M.Tech from a Tier-I or Tier-II Institution — CS, IS, ECE, AI/ML streams strongly preferred
  • Exceptional real-world engineering track record may be considered in lieu of institution pedigree


Experience

  • 6 to 9 years of progressive backend engineering experience with demonstrable ownership and impact
  • Proven track record of shipping and scaling production SaaS / Product systems at significant user load
  • Exposure to and success within start-up, mid-size, and large-scale product organisations — the full spectrum
  • Strong computer science fundamentals: algorithms, data structures, distributed systems theory, OS internals
  • Demonstrated career stability — minimum 2 years average tenure per organisation
  • The Ideal Candidate Exemplifies
  • System-level thinking with an ability to hold context across code, architecture, product, and business
  • An ownership mindset — no task is 'not my job'; outcomes and quality are personal commitments
  • Strong written and verbal communication skills for asynchronous, cross-functional collaboration
  • Intellectual curiosity: actively follows engineering trends, contributes to the community (OSS, blogs, talks)
  • Bias for automation, observability, and engineering efficiency at every level
  • A mentor's instinct — genuine desire to grow others and raise the capability of the team around them


WHY THIS ROLE STANDS APART

✅ Transformational Scope

Lead platform modernisation at scale. Your architectural choices will define systems serving millions of users for years.

✅ AI-Forward Engineering Culture

Be at the forefront of AI-augmented development. This org invests in tools and practices that make great engineers exceptional.

✅ Established, Stable Platform

Join a company with 500–1,000 employees, proven product-market fit, and the resources to execute on a serious technical vision.

✅ Career-Defining Leadership

Operate with strategic influence, direct access to senior leadership, and a clear path toward Principal / Staff / VP Engineering.


HOW TO APPLY

This search is being managed exclusively by Recruiting Bond

Submit your application with an updated resume

Only shortlisted candidates will be contacted. All applications are treated with the strictest confidentiality.

⚡ We move fast — qualified candidates can expect a response within 48–72 business hours.


Recruiting Bond | Bengaluru, Karnataka, India | 2026

Read more
Ebrotech Software Solutions
Remote, Delhi, Gurugram, Noida, Ghaziabad, Faridabad
8 - 10 yrs
₹15L - ₹20L / yr
Systems design
Microservices
Apache Kafka
Oracle
Cassandra

Job Description

  • 8+ years of experience to Lead, coach and mentor team of 5-8 full stack and backend engineers
  • Write high quality code
  • Champion coding standards and re-usable components
  • Influence the technical direction of the engineering team
  • Partner with Product Managers in designing and defining new features
  • Serve as a key member of a Scrum team
  • Participate and potentially lead Communities-of-Practice programs.

 

Requirements

  • 8+ years of large-scale distributed ecommerce systems experience
  • Expert understanding of Java, database, and messaging technologies
  • Enthusiasm for constant improvement as a Software Engineer
  • Ability to communicate clearly effectively and motivate team members
  • Reactive Java a plus
  • Passionate about Ecommerce and retail a plus


Technical Skill Set

  • Java (Core & Advanced)
  • Spring Boot, RESTful Services, Reactive Java (preferred)                                                                    
  • Vue.js, Web Components (Lit Framework)
  • Oracle, Cassandra
  • Apache Kafka
  • System Design, Microservices Architecture, Scalable & Event-Driven Systems
Read more
A real time Customer Data Platform and cross channel marketing automation delivers superior experiences that result in an increased revenue for some of the largest enterprises in the world.

A real time Customer Data Platform and cross channel marketing automation delivers superior experiences that result in an increased revenue for some of the largest enterprises in the world.

Agency job
via HyrHub by Neha Koshy
Bengaluru (Bangalore)
1 - 2 yrs
₹7L - ₹14L / yr
skill iconJava
skill iconSpring Boot
Apache Kafka
SQL
skill iconPostgreSQL
+3 more

Key Responsibilities:

  • Design and develop backend components and sub-systems for large-scale platforms under guidance from senior engineers.
  • Contribute to building and evolving the next-generation customer data platform.
  • Write clean, efficient, and well-tested code with a focus on scalability and performance.
  • Explore and experiment with modern technologies—especially open-source frameworks—
  • and build small prototypes or proof-of-concepts.
  • Use AI-assisted development tools to accelerate coding, testing, debugging, and learning while adhering to engineering best practices.
  • Participate in code reviews, design discussions, and continuous improvement of the platform.

Qualifications:

  • 0–2 years of experience (or strong academic/project background) in backend development with Java.
  • Good fundamentals in algorithms, data structures, and basic performance optimizations.
  • Bachelor’s or Master’s degree in Computer Science or IT (B.E / B.Tech / M.Tech / M.S) from premier institutes.

Technical Skill Set:

  • Strong aptitude and analytical skills with emphasis on problem solving and clean coding.
  • Working knowledge of SQL and NoSQL databases.
  • Familiarity with unit testing frameworks and writing testable code is a plus.
  • Basic understanding of distributed systems, messaging, or streaming platforms is a bonus.

AI-Assisted Engineering (LLM-Era Skills):

  • Familiarity with modern AI coding tools such as Cursor, Claude Code, Codex, Windsurf, Opencode, or similar.
  • Ability to use AI tools for code generation, refactoring, test creation, and learning new systems responsibly.
  • Willingness to learn how to combine human judgment with AI assistance for high-quality engineering outcomes.

Soft Skills & Nice to Have

  • Appreciation for technology and its ability to create real business value, especially in data and marketing platforms.
  • Clear written and verbal communication skills.
  • Strong ownership mindset and ability to execute in fast-paced environments.
  • Prior internship or startup experience is a plus.
Read more
SPGConsulting
Bengaluru (Bangalore)
3 - 6 yrs
₹5L - ₹15L / yr
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka
skill iconAmazon Web Services (AWS)
+2 more

Job Title: Java Backend Developer

Experience: ~3-6 years (Mid-to-Senior)

Employment Type: Full-time, Permanent

Location : Bangalore

Role Overview

As a Java Backend Developer, you’ll be responsible for designing, developing, and maintaining scalable backend systems and microservices. You’ll work with cross-functional teams to build high-performance distributed services, APIs, and data-driven applications that power business solutions.

Key Responsibilities

  • Design and implement microservices and backend components using Java (8+) and Spring Boot.
  • Build and consume RESTful APIs and integrate with internal/external services.
  • Work with event-driven systems and messaging using Apache Kafka (producers/consumers).
  • Develop and optimize databases, including SQL (e.g., MySQL/PostgreSQL) and NoSQL (e.g., MongoDB/Cassandra).
  • Participate in CI/CD pipelines, automated builds, and deployments using tools like Git, Maven, Jenkins.
  • Ensure code quality through unit and integration testing, documentation, and code reviews.
  • Collaborate with frontend developers, QA, DevOps, and product teams following Agile methodologies.

Required Skills & Qualifications

  • Bachelor’s degree in Computer Science, Information Technology, or related field.
  • Proven hands-on experience with Core Java and Spring Boot development.
  • Strong understanding of microservices architecture, REST APIs, and distributed systems.
  • Experience with message queues/event streaming (Apache Kafka).
  • Skilled in relational and NoSQL databases and writing optimized queries.
  • Comfortable with CI/CD tools (e.g., Git, Maven, Jenkins) and version control.
  • Good problem-solving, debugging, and collaboration skills.

Preferred / Nice-to-Have

  • Cloud platform experience (AWS / Azure / GCP).
  • Familiarity with containerization (Docker) and orchestration (Kubernetes).
  • Knowledge of performance tuning, caching strategies, observability (metrics/logging).
  • Agile/Scrum development experience.
Read more
Consumer Internet, Technology & Travel and Tourism Platform

Consumer Internet, Technology & Travel and Tourism Platform

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
4 - 7 yrs
₹38L - ₹50L / yr
DevOps
Cloud Computing
Infrastructure
skill iconKubernetes
skill iconDocker
+23 more

Job Details

Job Title: Senior DevOps Engineer

Industry: Consumer Internet, Technology & Travel and Tourism Platform

Function - IT

Experience Required: 4-7 years

Employment Type: Full Time

Job Location: Bengaluru

CTC Range: Best in Industry

 

Criteria:

  • Strong DevOps / Infrastructure Engineer Profiles.
  • Must have 4+ years of hands-on experience working as a DevOps Engineer / Infrastructure Engineer / SRE / DevOps Consultant.
  • Must have hands-on experience with Kubernetes and Docker, including deployment, scaling, or containerized application management.
  • Must have experience with Infrastructure as Code (IaC) or configuration management tools such as Terraform, Ansible, Chef, or Puppet.
  • Must have strong automation and scripting experience using Python, Go, Bash, Shell, or similar scripting languages.
  • Must have working experience with distributed databases or data systems such as MongoDB, Redis, Cassandra, Elasticsearch, or Kafka.
  • Candidate must demonstrate strong expertise in at least one of the following areas - Databases / Distributed Data Systems, Observability & Monitoring, CI/CD Pipelines. Networking Concepts, Kubernetes / Container Platforms
  • Candidates must be from B2C Product-based companies only.
  • (Education) – BE / B.Tech or equivalent

 

Preferred

  • Experience working with microservices or event-driven architectures.
  • Exposure to cloud infrastructure, monitoring, reliability, and scalability practices.
  • (Skills) – Understanding of programming languages such as Go, Python, or Java.
  • Preferred (Environment) – Experience working in high-scale production or fast-growing product startups.

 

Job Description 

As a DevOps Engineer, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.

 

Job Responsibilities:

  • Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs
  • Codify our infrastructure
  • Do what it takes to keep the uptime above 99.99%
  • Understand the bigger picture and sail through the ambiguities
  • Scale technology considering cost and observability and manage end-to-end processes
  • Understand DevOps philosophy and evangelize the principles across the organization
  • Strong communication and collaboration skills to break down the silos


Read more
HighLevel Inc.
Remote only
7yrs+
Best in industry
skill iconNodeJS (Node.js)
skill iconReact.js
skill iconMongoDB
skill iconExpress
MERN Stack
+2 more

About Us

HighLevel is an AI powered, all-in-one white-label sales & marketing platform that empowers agencies, entrepreneurs, and businesses to elevate their digital presence and drive growth. We are proud to support a global and growing community of over 2 million businesses, comprised of agencies, consultants, and businesses of all sizes and industries. HighLevel empowers users with all the tools needed to capture, nurture, and close new leads into repeat customers. As of mid 2025, HighLevel processes over 15 billion API hits and handles more than 2.5 billion message events every day. Our platform manages over 470 terabytes of data distributed across five databases, operates with a network of over 250 microservices, and supports over 1 million domain names.


About the Role

We are looking for a highly experienced Lead Engineer to take technical ownership of our Social Planner product suite, which enables users to connect social accounts, schedule and publish posts, manage comments, and analyze performance through real-time dashboards. This role will also help drive improvements across our Media Optimization system, ensuring we deliver a world-class experience for millions of users.


You will lead a team of talented engineers, shape technical direction, modernize architecture, and ensure the platform remains reliable, scalable, and high-performing as we continue to expand.


This is a high-impact leadership role for someone who thrives in complex distributed systems, loves solving tough technical challenges, and enjoys building products used at massive scale.


Responsibilities

  • Lead architecture, design, and development of the Social Planner & Media Optimization platforms.
  • Own end-to-end reliability, performance, and scalability of distributed systems.
  • Design and maintain fault-tolerant pipelines for posting, media processing, analytics, and comment workflows.
  • Drive modernization efforts: event-driven patterns, improved data models, and microservice decomposition.
  • Establish a long-term technical roadmap and make impactful architectural decisions.
  • Mentor, guide, and support backend, frontend, and full-stack engineers.
  • Conduct meaningful code reviews and enforce engineering best practices.
  • Collaborate with Product, Design, Data, and Support teams to deliver high-impact features.
  • Ensure high standards in code quality and documentation.
  • Improve posting reliability, delivery accuracy, retries, and error handling.
  • Build and optimize dashboards for analytics, scheduling workflows, and comment moderation.
  • Enhance media processing systems: compression, transcoding, storage, and CDN usage.
  • Troubleshoot complex production issues and lead root cause analyses.
  • Optimize performance across databases, search indexes, queues, and backend services.
  • Ensure platform security, privacy, and multi-tenant isolation best practices.

.

Requirements

  • 8+ years of full-stack engineering experience with significant backend/system design exposure.
  • Deep expertise in distributed systems and event-driven architecture.
  • Strong experience with asynchronous job processing and background workers.
  • Solid understanding of both relational and NoSQL databases (MongoDB, ClickHouse preferred).Proven ability to build scalable, multi-tenant SaaS applications.
  • Strong experience with modern frontend frameworks (Vue.js, React, or Angular).Ability to design and optimize interactive dashboards and data-heavy UIs.
  • Excellent understanding of software design patterns, API design, and microservices.
  • Hands-on experience with CI/CD pipelines, TDD, and automated testing frameworks.
  • Strong performance tuning, profiling, and optimization skills (especially Node.js).Experience with caching, rate limiting, and queuing systems.
  • Proven experience integrating third-party APIs (preferably social media APIs).Strong knowledge of Elasticsearch for search, indexing, and analytics.
  • Experience using Redis for caching, session storage, or background queues.
  • Experience leading engineering teams or acting as a technical lead.
  • Strong communication skills with the ability to work cross-functionally.
  • Ability to make clear architectural trade-offs and guide team decision-making.


Bonus Skills

  • Contributions to open-source tools, internal platforms or engineering blogsExperience with observability tools (Prometheus, Grafana, OpenTelemetry).
  • Knowledge of media processing, compression, encoding, and CDN workflows.
  • Exposure to AI-driven content tools or automation systems.
  • Experience with domain-driven design and system decomposition.
  • Familiarity with security standards, OAuth, rate limiting, and multi-tenant data protection.


Our People

With over 1,500 team members across 15+ countries, we operate in a global, remote-first environment. We are building more than software; we are building a global community rooted in creativity, collaboration, and impact. We take pride in cultivating a culture where innovation thrives, ideas are celebrated, and people come first, no matter where they call home.


Our Impact

As of mid 2025, our platform powers over 1.5 billion messages, helps generate over 200 million leads, and facilitates over 20 million conversations for the more than 2 million businesses we serve each month. Behind those numbers are real people growing their companies, connecting with customers, and making their mark - and we get to help make that happen.

Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune, Trivandrum , Thiruvananthapuram
8 - 10 yrs
₹20L - ₹24L / yr
skill iconJava
skill iconPython
API
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
+13 more

Job Details

Job Title: Lead Software Engineer - Java, Python, API Development

Industry: Global digital transformation solutions provider

Domain - Information technology (IT)

Experience Required: 8-10 years

Employment Type: Full Time

Job Location: Pune & Trivandrum/ Thiruvananthapuram

CTC Range: Best in Industry

 

Job Description

Job Summary

We are seeking a Lead Software Engineer with strong hands-on expertise in Java and Python to design, build, and optimize scalable backend applications and APIs. The ideal candidate will bring deep experience in cloud technologies, large-scale data processing, and leading the design of high-performance, reliable backend systems.

 

Key Responsibilities

  • Design, develop, and maintain backend services and APIs using Java and Python
  • Build and optimize Java-based APIs for large-scale data processing
  • Ensure high performance, scalability, and reliability of backend systems
  • Architect and manage backend services on cloud platforms (AWS, GCP, or Azure)
  • Collaborate with cross-functional teams to deliver production-ready solutions
  • Lead technical design discussions and guide best practices

 

Requirements

  • 8+ years of experience in backend software development
  • Strong proficiency in Java and Python
  • Proven experience building scalable APIs and data-driven applications
  • Hands-on experience with cloud services and distributed systems
  • Solid understanding of databases, microservices, and API performance optimization

 

Nice to Have

  • Experience with Spring Boot, Flask, or FastAPI
  • Familiarity with Docker, Kubernetes, and CI/CD pipelines
  • Exposure to Kafka, Spark, or other big data tools

 

Skills

Java, Python, API Development, Data Processing, AWS Backend

 

Skills: Java, API development, Data Processing, AWS backend, Python,

 

Must-Haves

Java (8+ years), Python (8+ years), API Development (8+ years), Cloud Services (AWS/GCP/Azure), Database & Microservices

8+ years of experience in backend software development

Strong proficiency in Java and Python

Proven experience building scalable APIs and data-driven applications

Hands-on experience with cloud services and distributed systems

Solid understanding of databases, microservices, and API performance optimization

Mandatory Skills: Java API AND AWS

 

******

Notice period - 0 to 15 days only

Job stability is mandatory

Location: Pune, Trivandrum

Read more
NeoGenCode Technologies Pvt Ltd
Bengaluru (Bangalore)
3 - 8 yrs
₹10L - ₹25L / yr
skill iconGo Programming (Golang)
skill iconPython
FastAPI
skill iconJava
skill iconSpring Boot
+12 more

🚀 Job Title : Backend Engineer (Go / Python / Java)

Experience : 3+ Years

Location : Bangalore (Client Location – Work From Office)

Notice Period : Immediate to 15 Days

Open Positions : 4

Working Days : 6 Days a Week


🧠 Job Summary :

We are looking for a highly skilled Backend Engineer to build scalable, reliable, and high-performance systems in a fast-paced product environment.

You will own large features end-to-end — from design and development to deployment and monitoring — while collaborating closely with product, frontend, and infrastructure teams.

This role requires strong backend fundamentals, distributed systems exposure, and a mindset of operational ownership.


⭐ Mandatory Skills :

Strong backend development experience in Go / Python (FastAPI) / Java (Spring Boot) with hands-on expertise in Microservices, REST APIs, PostgreSQL, Redis, Kafka/SQS, AWS/GCP, Docker, Kubernetes, CI/CD, and strong DSA & System Design fundamentals.


🔧 Key Responsibilities :

  • Design, develop, test, and deploy backend services end-to-end.
  • Build scalable, modular, and production-grade microservices.
  • Develop and maintain RESTful APIs.
  • Architect reliable distributed systems with performance and fault tolerance in mind.
  • Debug complex cross-system production issues.
  • Implement secure development practices (authentication, authorization, data integrity).
  • Work with monitoring dashboards, alerts, and performance metrics.
  • Participate in code reviews and enforce engineering best practices.
  • Contribute to CI/CD pipelines and release processes.
  • Collaborate with product, frontend, and DevOps teams.

✅ Required Skills :

  • Strong proficiency in Go OR Python (FastAPI) OR Java (Spring Boot).
  • Hands-on experience building Microservices-based architectures.
  • Strong understanding of REST APIs & distributed systems.
  • Experience with PostgreSQL and Redis.
  • Exposure to Kafka / SQS or other messaging systems.
  • Hands-on experience with AWS or GCP.
  • Experience with Docker and Kubernetes.
  • Familiarity with CI/CD pipelines.
  • Strong knowledge of Data Structures & System Design.
  • Ability to independently own features and solve ambiguous engineering problems.

⭐ Preferred Background :

  • Experience in product-based companies.
  • Exposure to high-throughput or event-driven systems.
  • Strong focus on code quality, observability, and reliability.
  • Comfortable working in high-growth, fast-paced environments.

🧑‍💻 Interview Process :

  • 1 Internal Screening Round
  • HR Discussion (Project & Communication Evaluation)
  • 3 Technical Rounds with Client

This is a fresh requirement, and interviews will be scheduled immediately.

Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Trivandrum, Thiruvananthapuram
9 - 12 yrs
₹21L - ₹27L / yr
skill iconJava
Spring
Apache Kafka
SQL
skill iconPostgreSQL
+16 more

JOB DETAILS:

Job Title: Java Lead-Java, MS, Kafka-TVM - Java (Core & Enterprise), Spring/Micronaut, Kafka

Industry: Global Digital Transformation Solutions Provider

Salary: Best in Industry

Experience: 9 to 12 years

Location: Trivandrum, Thiruvananthapuram

 

Job Description

Experience

  • 9+ years of experience in Java-based backend application development
  • Proven experience building and maintaining enterprise-grade, scalable applications
  • Hands-on experience working with microservices and event-driven architectures
  • Experience working in Agile and DevOps-driven development environments

 

Mandatory Skills

  • Advanced proficiency in core Java and enterprise Java concepts
  • Strong hands-on experience with Spring Framework and/or Micronaut for building scalable backend applications
  • Strong expertise in SQL, including database design, query optimization, and performance tuning
  • Hands-on experience with PostgreSQL or other relational database management systems
  • Strong experience with Kafka or similar event-driven messaging and streaming platforms
  • Practical knowledge of CI/CD pipelines using GitLab
  • Experience with Jenkins for build automation and deployment processes
  • Strong understanding of GitLab for source code management and DevOps workflows

 

Responsibilities

  • Design, develop, and maintain robust, scalable, and high-performance backend solutions
  • Develop and deploy microservices using Spring or Micronaut frameworks
  • Implement and integrate event-driven systems using Kafka
  • Optimize SQL queries and manage PostgreSQL databases for performance and reliability
  • Build, implement, and maintain CI/CD pipelines using GitLab and Jenkins
  • Collaborate with cross-functional teams including product, QA, and DevOps to deliver high-quality software solutions
  • Ensure code quality through best practices, reviews, and automated testing

 

Good-to-Have Skills

  • Strong problem-solving and analytical abilities
  • Experience working with Agile development methodologies such as Scrum or Kanban
  • Exposure to cloud platforms such as AWS, Azure, or GCP
  • Familiarity with containerization and orchestration tools such as Docker or Kubernetes

 

Skills: java, spring boot, kafka development, cicd, postgresql, gitlab

 

Must-Haves

Java Backend (9+ years), Spring Framework/Micronaut, SQL/PostgreSQL, Kafka, CI/CD (GitLab/Jenkins)

Advanced proficiency in core Java and enterprise Java concepts

Strong hands-oacn experience with Spring Framework and/or Micronaut for building scalable backend applications

Strong expertise in SQL, including database design, query optimization, and performance tuning

Hands-on experience with PostgreSQL or other relational database management systems

Strong experience with Kafka or similar event-driven messaging and streaming platforms

Practical knowledge of CI/CD pipelines using GitLab

Experience with Jenkins for build automation and deployment processes

Strong understanding of GitLab for source code management and DevOps workflows

 

 

*******

Notice period - 0 to 15 days only

Job stability is mandatory

Location: only Trivandrum

F2F Interview on 21st Feb 2026

 

Read more
AI Industry

AI Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Mumbai, Bengaluru (Bangalore), Hyderabad, Gurugram
5 - 17 yrs
₹34L - ₹45L / yr
Dremio
Data engineering
Business Intelligence (BI)
Tableau
PowerBI
+51 more

Review Criteria:

  • Strong Dremio / Lakehouse Data Architect profile
  • 5+ years of experience in Data Architecture / Data Engineering, with minimum 3+ years hands-on in Dremio
  • Strong expertise in SQL optimization, data modeling, query performance tuning, and designing analytical schemas for large-scale systems
  • Deep experience with cloud object storage (S3 / ADLS / GCS) and file formats such as Parquet, Delta, Iceberg along with distributed query planning concepts
  • Hands-on experience integrating data via APIs, JDBC, Delta/Parquet, object storage, and coordinating with data engineering pipelines (Airflow, DBT, Kafka, Spark, etc.)
  • Proven experience designing and implementing lakehouse architecture including ingestion, curation, semantic modeling, reflections/caching optimization, and enabling governed analytics
  • Strong understanding of data governance, lineage, RBAC-based access control, and enterprise security best practices
  • Excellent communication skills with ability to work closely with BI, data science, and engineering teams; strong documentation discipline
  • Candidates must come from enterprise data modernization, cloud-native, or analytics-driven companies


Preferred:

  • Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) or data catalogs (Collibra, Alation, Purview); familiarity with Snowflake, Databricks, or BigQuery environments


Role & Responsibilities:

You will be responsible for architecting, implementing, and optimizing Dremio-based data lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.


  • Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
  • Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
  • Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
  • Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
  • Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
  • Support self-service analytics by enabling governed data products and semantic layers.
  • Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
  • Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.


Ideal Candidate:

  • Bachelor’s or Master’s in Computer Science, Information Systems, or related field.
  • 5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
  • Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
  • Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
  • Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
  • Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
  • Excellent problem-solving, documentation, and stakeholder communication skills.


Preferred:

  • Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) and data catalogs (Collibra, Alation, Purview).
  • Exposure to Snowflake, Databricks, or BigQuery environments.
  • Experience in high-tech, manufacturing, or enterprise data modernization programs.
Read more
Global digital transformation solutions provider.

Global digital transformation solutions provider.

Agency job
via Peak Hire Solutions by Dhara Thakkar
Hyderabad
5 - 8 yrs
₹11L - ₹20L / yr
PySpark
Apache Kafka
Data architecture
skill iconAmazon Web Services (AWS)
EMR
+32 more

JOB DETAILS:

* Job Title: Lead II - Software Engineering - AWS, Apache Spark (PySpark/Scala), Apache Kafka

* Industry: Global digital transformation solutions provider

* Salary: Best in Industry

* Experience: 5-8 years

* Location: Hyderabad

 

Job Summary

We are seeking a skilled Data Engineer to design, build, and optimize scalable data pipelines and cloud-based data platforms. The role involves working with large-scale batch and real-time data processing systems, collaborating with cross-functional teams, and ensuring data reliability, security, and performance across the data lifecycle.


Key Responsibilities

ETL Pipeline Development & Optimization

  • Design, develop, and maintain complex end-to-end ETL pipelines for large-scale data ingestion and processing.
  • Optimize data pipelines for performance, scalability, fault tolerance, and reliability.

Big Data Processing

  • Develop and optimize batch and real-time data processing solutions using Apache Spark (PySpark/Scala) and Apache Kafka.
  • Ensure fault-tolerant, scalable, and high-performance data processing systems.

Cloud Infrastructure Development

  • Build and manage scalable, cloud-native data infrastructure on AWS.
  • Design resilient and cost-efficient data pipelines adaptable to varying data volume and formats.

Real-Time & Batch Data Integration

  • Enable seamless ingestion and processing of real-time streaming and batch data sources (e.g., AWS MSK).
  • Ensure consistency, data quality, and a unified view across multiple data sources and formats.

Data Analysis & Insights

  • Partner with business teams and data scientists to understand data requirements.
  • Perform in-depth data analysis to identify trends, patterns, and anomalies.
  • Deliver high-quality datasets and present actionable insights to stakeholders.

CI/CD & Automation

  • Implement and maintain CI/CD pipelines using Jenkins or similar tools.
  • Automate testing, deployment, and monitoring to ensure smooth production releases.

Data Security & Compliance

  • Collaborate with security teams to ensure compliance with organizational and regulatory standards (e.g., GDPR, HIPAA).
  • Implement data governance practices ensuring data integrity, security, and traceability.

Troubleshooting & Performance Tuning

  • Identify and resolve performance bottlenecks in data pipelines.
  • Apply best practices for monitoring, tuning, and optimizing data ingestion and storage.

Collaboration & Cross-Functional Work

  • Work closely with engineers, data scientists, product managers, and business stakeholders.
  • Participate in agile ceremonies, sprint planning, and architectural discussions.


Skills & Qualifications

Mandatory (Must-Have) Skills

  1. AWS Expertise
  • Hands-on experience with AWS Big Data services such as EMR, Managed Apache Airflow, Glue, S3, DMS, MSK, and EC2.
  • Strong understanding of cloud-native data architectures.
  1. Big Data Technologies
  • Proficiency in PySpark or Scala Spark and SQL for large-scale data transformation and analysis.
  • Experience with Apache Spark and Apache Kafka in production environments.
  1. Data Frameworks
  • Strong knowledge of Spark DataFrames and Datasets.
  1. ETL Pipeline Development
  • Proven experience in building scalable and reliable ETL pipelines for both batch and real-time data processing.
  1. Database Modeling & Data Warehousing
  • Expertise in designing scalable data models for OLAP and OLTP systems.
  1. Data Analysis & Insights
  • Ability to perform complex data analysis and extract actionable business insights.
  • Strong analytical and problem-solving skills with a data-driven mindset.
  1. CI/CD & Automation
  • Basic to intermediate experience with CI/CD pipelines using Jenkins or similar tools.
  • Familiarity with automated testing and deployment workflows.

 

Good-to-Have (Preferred) Skills

  • Knowledge of Java for data processing applications.
  • Experience with NoSQL databases (e.g., DynamoDB, Cassandra, MongoDB).
  • Familiarity with data governance frameworks and compliance tooling.
  • Experience with monitoring and observability tools such as AWS CloudWatch, Splunk, or Dynatrace.
  • Exposure to cost optimization strategies for large-scale cloud data platforms.

 

Skills: big data, scala spark, apache spark, ETL pipeline development

 

******

Notice period - 0 to 15 days only

Job stability is mandatory

Location: Hyderabad

Note: If a candidate is a short joiner, based in Hyderabad, and fits within the approved budget, we will proceed with an offer

F2F Interview: 14th Feb 2026

3 days in office, Hybrid model.

 


Read more
Trential Technologies

at Trential Technologies

1 candid answer
Garima Jangid
Posted by Garima Jangid
Gurugram
5 - 8 yrs
₹30L - ₹45L / yr
skill iconNodeJS (Node.js)
skill iconJavascript
RabbitMQ
Apache Kafka
skill iconRedis
+14 more

About us:

Trential is engineering the future of digital identity with W3C Verifiable Credentials—secure, decentralized, privacy-first. We make identity and credentials verifiable anywhere, instantly.


We are looking for a Team lead to architect, build, and scale high-performance web applications that power our core products. You will lead the full development lifecycle—from system design to deployment—while mentoring the team and driving best engineering practices across frontend and backend stacks.


 Design & Implement: Lead the design, implementation and management of Trential products.

 Lead by example: Be the most senior and impactful engineer on the team, setting the technical bar through your direct contributions.

 Code Quality & Best Practices: Enforce high standards for code quality, security, and performance through rigorous code reviews, automated testing, and continuous delivery pipelines.

 Standards Adherence: Ensure all solutions comply with relevant open standards like W3C Verifiable Credentials (VCs), Decentralized Identifiers (DIDs) & Privacy Laws, maintaining global interoperability.

 Continuous Improvement: Lead the charge to continuously evaluate and improve the products & processes. Instill a culture of metrics-driven process improvement to boost team efficiency and product quality.

 Cross-Functional Collaboration: Work closely with the Co-Founders & Product Team to translate business requirements and market needs into clear, actionable technical specifications and stories. Represent Trential in interactions with external stakeholders for integrations.


What we're looking for:

 Experience: 5+ years of experience in software development, with at least 2 years as a Technical Lead.

 Technical Depth: Deep proficiency in JavaScript and experience in building and operating distributed, fault-tolerant systems.

 Cloud & Infrastructure: Hands-on experience with cloud platforms (AWS & GCP) and modern DevOps practices (e.g., CI/CD, Infrastructure as Code, Docker).

 Databases: Strong knowledge of SQL/NoSQL databases and data modeling for high-throughput, secure applications.


Preferred Qualifications (Nice to Have)

 Identity & Credentials: Knowledge of decentralized identity principles, Verifiable Credentials (W3C VCs), DIDs, and relevant protocols (e.g., OpenID4VC, DIDComm)

 Familiarity with data privacy and security standards (GDPR, SOC 2, ISO 27001) and designing systems complying to these laws.

 Experience integrating AI/ML models into verification or data extraction workflows

Read more
Remote only
6 - 10 yrs
₹15L - ₹25L / yr
Cloud Architect
data Architect
skill iconData Analytics
Google Cloud Platform (GCP)
Apache Kafka

Role Summary

Provide architectural leadership across a large-scale, multi-cloud data ecosystem, helping design and guide scalable, future-ready data platforms.


Key Responsibilities

• Design and review data architectures across Google Cloud and Microsoft Azure (multi-cloud).

• Guide decisions around data platforms, pipelines, streaming, and integration patterns.

• Advise on abstraction layers, APIs, messaging/streaming (Kafka, MQ), and system interoperability.

• Partner with engineering teams to ensure designs are practical and executable.


Key Skills

• Deep experience with large-scale data platforms and distributed systems.

• Strong background in multi-cloud architectures (GCP + Azure).

• Expertise in data pipelines, streaming, and enterprise integration.

Read more
Technology Industry

Technology Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
5 - 8 yrs
₹38L - ₹50L / yr
skill iconJava
skill iconSpring Boot
CI/CD
Spring
Microservices
+16 more

Job Details

- Job Title: SDE-3

Industry: Technology

Domain - Information technology (IT)

Experience Required: 5-8 years

Employment Type: Full Time

Job Location: Bengaluru

CTC Range: Best in Industry

 

Role & Responsibilities

As a Software Development Engineer - 3, Backend Engineer at company, you will play a critical role in architecting, designing, and delivering robust backend systems that power our platform. You will lead by example, driving technical excellence and mentoring peers while solving complex engineering problems. This position offers the opportunity to work with a highly motivated team in a fast-paced and innovative environment.

 

Key Responsibilities:

Technical Leadership-

  • Design and develop highly scalable, fault-tolerant, and maintainable backend systems using Java and related frameworks.
  • Provide technical guidance and mentorship to junior developers, fostering a culture of learning and growth.
  • Review code and ensure adherence to best practices, coding standards, and security guidelines.

System Architecture and Design-

  • Collaborate with cross-functional teams, including product managers and frontend engineers, to translate business requirements into efficient technical solutions.
  • Own the architecture of core modules and contribute to overall platform scalability and reliability.
  • Advocate for and implement microservices architecture, ensuring modularity and reusability.

Problem Solving and Optimization-

  • Analyze and resolve complex system issues, ensuring high availability and performance of the platform.
  • Optimize database queries and design scalable data storage solutions.
  • Implement robust logging, monitoring, and alerting systems to proactively identify and mitigate issues.

Innovation and Continuous Improvement-

  • Stay updated on emerging backend technologies and incorporate relevant advancements into our systems.
  • Identify and drive initiatives to improve codebase quality, deployment processes, and team productivity.
  • Contribute to an advocate for a DevOps culture, supporting CI/CD pipelines and automated testing.

Collaboration and Communication-

  • Act as a liaison between the backend team and other technical and non-technical teams, ensuring smooth communication and alignment.
  • Document system designs, APIs, and workflows to maintain clarity and knowledge transfer across the team.

 

Ideal Candidate

  • Strong Java Backend Engineer.
  • Must have 5+ years of backend development with strong focus on Java (Spring / Spring Boot)
  • Must have been SDE-2 for at least 2.5 years
  • Hands-on experience with RESTful APIs and microservices architecture
  • Strong understanding of distributed systems, multithreading, and async programming
  • Experience with relational and NoSQL databases
  • Exposure to Kafka/RabbitMQ and Redis/Memcached
  • Experience with AWS / GCP / Azure, Docker, and Kubernetes
  • Familiar with CI/CD pipelines and modern DevOps practices
  • Product companies (B2B SAAS preferred)
  • have stayed for at least 2 years with each of the previous companies
  • (Education): B.Tech in computer science from Tier 1, Tier 2 colleges


Read more
Mumbai
4 - 10 yrs
₹12L - ₹24L / yr
skill iconNodeJS (Node.js)
skill iconAmazon Web Services (AWS)
skill iconMongoDB
skill iconPostgreSQL
NOSQL Databases
+3 more

Job Title : Senior Backend Developer (Node.js + AWS + MongoDB)

Experience : 4+ Years

Location : Andheri, Mumbai (Work From Office)


About the Role :

We are looking for a highly skilled Senior Backend Developer with strong expertise in Node.js (NestJS), AWS, and MongoDB to join our growing engineering team.

This role requires someone who takes ownership, is proactive, and enjoys building scalable, high-performance backend systems in a fast-paced environment.


Key Responsibilities :

  • Architect, design, and develop scalable backend services using Node.js (NestJS).
  • Design and manage cloud infrastructure on AWS Services (EC2, ECS, RDS, Lambda, etc.).
  • Develop and maintain high-performance database solutions using MongoDB.
  • Work with Kafka, Docker, and serverless frameworks (SST) for efficient deployments.
  • Optimize system performance, scalability, and reliability across services.
  • Ensure application security, best practices, and compliance standards.
  • Collaborate with cross-functional teams to deliver robust product features.
  • Take end-to-end ownership of features from design to deployment.

Technical Requirements :

  • 4+ years of backend development experience.
  • 3+ years of hands-on experience with Node.js.
  • 2+ years of hands-on experience with AWS.
  • Strong experience with NestJS framework.
  • Solid experience with MongoDB and database design.
  • Experience with Kafka, Docker, and serverless architecture.
  • Understanding of system design, scalability, and performance optimization.

Good to Have (Bonus Skills) :

  • Experience with Python or other backend languages.
  • Exposure to Agentic AI use cases or implementations.
  • Strong understanding of security best practices.

What We’re Looking For :

  • Curious mindset and eagerness to learn new technologies.
  • Proactive problem solver with strong ownership attitude.
  • Strong team player with effective communication skills.
  • Positive, energetic, and passionate about building great systems.
Read more
Global digital transformation solutions provider

Global digital transformation solutions provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Trivandrum, Thiruvananthapuram
5 - 7 yrs
₹18L - ₹26L / yr
skill iconKotlin
skill iconJava
skill iconAmazon Web Services (AWS)
skill iconSpring Boot
Microservices
+24 more

JOB DETAILS:

* Job Title: Lead I - Software Engineering-Kotlin, Java, Spring Boot, Aws

* Industry: Global digital transformation solutions provide

* Salary: Best in Industry

* Experience: 5 -7 years

* Location: Trivandrum, Thiruvananthapuram

Role Proficiency:

Act creatively to develop applications and select appropriate technical options optimizing application development maintenance and performance by employing design patterns and reusing proven solutions account for others' developmental activities

 

Skill Examples:

  1. Explain and communicate the design / development to the customer
  2. Perform and evaluate test results against product specifications
  3. Break down complex problems into logical components
  4. Develop user interfaces business software components
  5. Use data models
  6. Estimate time and effort required for developing / debugging features / components
  7. Perform and evaluate test in the customer or target environment
  8. Make quick decisions on technical/project related challenges
  9. Manage a Team mentor and handle people related issues in team
  10.  Maintain high motivation levels and positive dynamics in the team.
  11.  Interface with other teams’ designers and other parallel practices
  12.  Set goals for self and team. Provide feedback to team members
  13.  Create and articulate impactful technical presentations
  14.  Follow high level of business etiquette in emails and other business communication
  15.  Drive conference calls with customers addressing customer questions
  16.   Proactively ask for and offer help
  17.  Ability to work under pressure determine dependencies risks facilitate planning; handling multiple tasks.
  18.  Build confidence with customers by meeting the deliverables on time with quality.
  19.  Estimate time and effort resources required for developing / debugging features / components
  20.  Make on appropriate utilization of Software / Hardware’s.
  21.  Strong analytical and problem-solving abilities

 

Knowledge Examples:

  •     Appropriate software programs / modules
  1. Functional and technical designing
  2. Programming languages – proficient in multiple skill clusters
  3. DBMS
  4. Operating Systems and software platforms
  5. Software Development Life Cycle
  6. Agile – Scrum or Kanban Methods
  7. Integrated development environment (IDE)
  8. Rapid application development (RAD)
  9. Modelling technology and languages
  10. Interface definition languages (IDL)
  11. Knowledge of customer domain and deep understanding of sub domain where problem is solved

 

Additional Comments:

We are seeking an experienced Senior Backend Engineer with strong expertise in Kotlin and Java to join our dynamic engineering team.

The ideal candidate will have a deep understanding of backend frameworks, cloud technologies, and scalable microservices architectures, with a passion for clean code, resilience, and system observability.

You will play a critical role in designing, developing, and maintaining core backend services that power our high-availability e-commerce and promotion platforms.

 

Key Responsibilities

Design, develop, and maintain backend services using Kotlin (JVM, Coroutines, Serialization) and Java.

Build robust microservices with Spring Boot and related Spring ecosystem components (Spring Cloud, Spring Security, Spring Kafka, Spring Data).

Implement efficient serialization/deserialization using Jackson and Kotlin Serialization. Develop, maintain, and execute automated tests using JUnit 5, Mockk, and ArchUnit to ensure code quality.

Work with Kafka Streams (Avro), Oracle SQL (JDBC, JPA), DynamoDB, and Redis for data storage and caching needs. Deploy and manage services in AWS environment leveraging DynamoDB, Lambdas, and IAM.

Implement CI/CD pipelines with GitLab CI to automate build, test, and deployment processes.

Containerize applications using Docker and integrate monitoring using Datadog for tracing, metrics, and dashboards.

Define and maintain infrastructure as code using Terraform for services including GitLab, Datadog, Kafka, and Optimizely.

Develop and maintain RESTful APIs with OpenAPI (Swagger) and JSON API standards.

Apply resilience patterns using Resilience4j to build fault-tolerant systems.

Adhere to architectural and design principles such as Domain-Driven Design (DDD), Object-Oriented Programming (OOP), and Contract Testing (Pact).

Collaborate with cross-functional teams in an Agile Scrum environment to deliver high-quality features.

Utilize feature flagging tools like Optimizely to enable controlled rollouts.

 

Mandatory Skills & Technologies Languages:

Kotlin (JVM, Coroutines, Serialization),

Java Frameworks: Spring Boot (Spring Cloud, Spring Security, Spring Kafka, Spring Data)

Serialization: Jackson, Kotlin Serialization

Testing: JUnit 5, Mockk, ArchUnit

Data: Kafka (Avro) Streams Oracle SQL (JDBC, JPA) DynamoDB (NoSQL) Redis (Caching)

Cloud: AWS (DynamoDB, Lambda, IAM)

CI/CD: GitLab CI Containers: Docker

Monitoring & Observability: Datadog (Tracing, Metrics, Dashboards, Monitors)

Infrastructure as Code: Terraform (GitLab, Datadog, Kafka, Optimizely)

API: OpenAPI (Swagger), REST API, JSON API

Resilience: Resilience4j

Architecture & Practices: Domain-Driven Design (DDD) Object-Oriented Programming (OOP) Contract Testing (Pact) Feature Flags (Optimizely)

Platforms: E-Commerce Platform (CommerceTools), Promotion Engine (Talon.One)

Methodologies: Scrum, Agile

 

Skills: Kotlin, Java, Spring Boot, Aws

 

Must-Haves

Kotlin (JVM, Coroutines, Serialization), Java, Spring Boot (Spring Cloud, Spring Security, Spring Kafka, Spring Data), AWS (DynamoDB, Lambda, IAM), Microservices Architecture

 

 

******

Notice period - 0 to 15 days only

Job stability is mandatory

Location: Trivandrum

Virtual Weekend Interview on 7th Feb 2026 - Saturday

Read more
Deqode

at Deqode

1 recruiter
Apoorva Jain
Posted by Apoorva Jain
Hyderabad
6 - 12 yrs
₹5L - ₹20L / yr
skill iconJava
Apache Kafka
skill iconSpring Boot
Spring MVC
Microservices

Job Summary

We are seeking an experienced Java + Kafka Developer to design, develop, and maintain high-performance, event-driven applications. The ideal candidate will have strong backend development skills in Java and hands-on experience with Apache Kafka for building scalable, real-time data processing systems.

Key Responsibilities

  • Design and develop event-driven microservices using Java and Apache Kafka
  • Build and maintain Kafka producers, consumers, topics, and streams
  • Implement message processing, retries, error handling, and idempotency
  • Develop and integrate RESTful APIs
  • Ensure high availability, scalability, and performance of Kafka-based systems
  • Monitor, troubleshoot, and optimize Kafka and application performance
  • Collaborate with architects, DevOps, and QA teams
  • Participate in code reviews and follow best practices
  • Contribute to system design and technical discussions

Required Skills

  • 6+ years of experience in Java backend development
  • Strong hands-on experience with Apache Kafka
  • Experience with event-driven and streaming architectures
  • Solid understanding of microservices architecture
  • Experience with REST APIs
  • Good knowledge of multithreading, concurrency, and JVM concepts
  • Familiarity with SQL / NoSQL databases
  • Strong problem-solving and debugging skills



Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
6 - 10 yrs
Best in industry
Salesforce Apex
Lightning Web Components
Salesforce Visualforce
API
CI/CD
+4 more

Key Responsibilities:-

  • Design, build, and enhance Salesforce applications using Apex, Lightning Web Components (LWC), Visualforce, and SOQL.
  • Implement integrations with external systems using REST APIs and event-driven messaging (e.g., Kafka).
  • Collaborate with architects and business analysts to translate requirements into scalable, maintainable solutions.
  • Establish and follow engineering best practices, including source control (Git), code reviews, branching strategies, CI/CD pipelines, automated testing, and environment management.
  • Establish and maintain Azure DevOps-based workflows (repos, pipelines, automated testing) for Salesforce engineering.
  • Ensure solutions follow Salesforce security, data modeling, and performance guidelines.
  • Participate in Agile ceremonies, providing technical expertise and leadership within sprints and releases.
  • Optimize workflows, automations, and data processes across Sales Cloud, Service Cloud, and custom Salesforce apps.
  • Provide technical mentoring and knowledge sharing when required.
  • Support production environments, troubleshoot issues, and drive root-cause analysis for long-term reliability.
  • Stay current on Salesforce platform updates, releases, and new features, recommending adoption where beneficial.


Required Qualifications:-

  • Bachelor’s degree in Computer Science, Engineering, or related field (or equivalent experience).
  • 6+ years of Salesforce development experience with strong knowledge of Apex, Lightning Web Components, and Salesforce APIs.
  • Proven experience with Salesforce core clouds (Sales Cloud, Service Cloud, or equivalent).
  • Strong hands-on experience with API integrations (REST/SOAP) and event-driven architectures (Kafka, Pub/Sub).
  • Solid understanding of engineering practices: Git-based source control (Salesforce DX/metadata), branching strategies, CI/CD, automated testing, and deployment management.
  • Familiarity with Azure DevOps repositories and pipelines.
  • Strong knowledge of Salesforce data modeling, security, and sharing rules.
  • Excellent problem-solving skills and ability to collaborate across teams.


Preferred Qualifications:-

  • Salesforce Platform Developer II certification (or equivalent advanced credentials).
  • Experience with Health Cloud, Financial Services Cloud, or other industry-specific Salesforce products.
  • Experience implementing logging, monitoring, and observability within Salesforce and integrated systems.
  • Background in Agile/Scrum delivery with strong collaboration skills.
  • Prior experience establishing or enforcing engineering standards across Salesforce teams.


Read more
Remote only
6 - 12 yrs
₹15L - ₹35L / yr
skill icon.NET
ASP.NET MVC
Apache Kafka
Middleware
Microsoft Windows Azure
+5 more

Key Responsibilities 

• Design, develop, and deploy scalable .NET applications and microservices supporting core 

retail operations.

• Build and maintain robust integrations using Kafka, REST APIs, and middleware 

components.

• Work closely with product, POS, merchandising, warehousing, and ERP teams to translate 

business needs into reliable technical solutions.

• Optimize system performance and ensure integrations remain stable across high-volume 

retail cycles.

• Implement best practices in coding, testing, CI/CD, and cloud-native architectures.

• Troubleshoot production issues with speed and a cool head—especially when it impacts 

stores.

• Document technical designs, data flows, and integration contracts clearly.


Required Skills & Experience 

• 6–8+ years of hands-on full stack development using .NET Core / .NET 6+, C#, and 

modern frontend frameworks (React/Angular preferred).

• Strong expertise in Kafka (producers, consumers, topics, partitions, schema registry).

• Solid experience designing RESTful APIs and integrating with enterprise systems.

• Prior experience in retail technology is non-negotiable—POS, inventory, warehousing, or 

ERP exposure is essential.

• Strong knowledge of relational databases (SQL Server, PostgreSQL) and caching 

frameworks.

• Experience building scalable, distributed systems that can handle peak loads.

• Familiarity with containerization (Docker/Kubernetes) is a plus.

• Excellent problem-solving skills and the ability to work independently on a contract basis.


Nice to Have 

• Experience with message brokers beyond Kafka (e.g., RabbitMQ)

• Exposure to cloud platforms (Azure/AWS/GCP)

• Knowledge of CI/CD pipelines (Azure DevOps, GitHub Actions, Jenkins)

• Understanding of retail data models and omnichannel workflows

Read more
Technology Industry

Technology Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
6 - 9 yrs
₹36L - ₹48L / yr
skill iconPython
TypeScript
skill iconNodeJS (Node.js)
ReAct (Reason + Act)
skill iconReact Native
+13 more

Review Criteria:

  • Strong Software Engineer fullstack profile using NodeJS / Python and React
  • 6+ YOE in Software Development using Python OR NodeJS (For backend) & React (For frontend)
  • Must have strong experience in working on Typescript
  • Must have experience in message-based systems like Kafka, RabbitMq, Redis
  • Databases - PostgreSQL & NoSQL databases like MongoDB
  • Product Companies Only
  • Tier 1 Engineering Institutes preferred (IIT, NIT, BITS, IIIT, DTU or equivalent)

 

Preferred:

  • Experience in Fin-Tech, Payment, POS and Retail products is highly preferred
  • Experience in mentoring, coaching the team.


Role & Responsibilities:

We are currently seeking a Senior Engineer to join our Financial Services team, contributing to the design and development of scalable system.

 

The Ideal Candidate Will Be Able To-

  • Take ownership of delivering performant, scalable and high-quality cloud-based software, both frontend and backend side.
  • Mentor team members to develop in line with product requirements.
  • Collaborate with Senior Architect for design and technology choices for product development roadmap.
  • Do code reviews.


Ideal Candidate:

  • Thorough knowledge of developing cloud-based software including backend APIs and react based frontend.
  • Thorough knowledge of scalable design patterns and message-based systems such as Kafka, RabbitMq, Redis, MongoDB, ORM, SQL etc.
  • Experience with AWS services such as S3, IAM, Lambda etc.
  • Expert level coding skills in Python FastAPI/Django, NodeJs, TypeScript, ReactJs.
  • Eye for user responsive designs on the frontend.


Read more
AryuPay Technologies
Bhavana Chaudhari
Posted by Bhavana Chaudhari
Bengaluru (Bangalore)
4 - 8 yrs
₹4L - ₹9L / yr
skill iconDjango
skill iconPython
RESTful APIs
skill iconFlask
skill iconPostgreSQL
+7 more

We are seeking a highly skilled and experienced Python Developer with a strong background in fintech to join our dynamic team. The ideal candidate will have at least 7+ years of professional experience in Python development, with a proven track record of delivering high-quality software solutions in the fintech industry.

Responsibilities:

Design, build, and maintain RESTful APIs using Django and Django Rest Framework.

Integrate AI/ML models into existing applications to enhance functionality and provide data-driven insights.

Collaborate with cross-functional teams, including product managers, designers, and other developers, to define and implement new features and functionalities.

Manage deployment processes, ensuring smooth and efficient delivery of applications.

Implement and maintain payment gateway solutions to facilitate secure transactions.

Conduct code reviews, provide constructive feedback, and mentor junior members of the development team.

Stay up-to-date with emerging technologies and industry trends, and evaluate their potential impact on our products and services.

Maintain clear and comprehensive documentation for all development processes and integrations.

Requirements:

Proficiency in Python and Django/Django Rest Framework.

Experience with REST API development and integration.

Knowledge of AI/ML concepts and practical experience integrating AI/ML models.

Hands-on experience with deployment tools and processes.

Familiarity with payment gateway integration and management.

Strong understanding of database systems (SQL, PostgreSQL, MySQL).

Experience with version control systems (Git).

Strong problem-solving skills and attention to detail.

Excellent communication and teamwork skills.

Job Types: Full-time, Permanent

Work Location: In person

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Archana M
Posted by Archana M
Bengaluru (Bangalore)
8 - 12 yrs
Best in industry
Distributed Systems
skill iconLeadership
LLD
HLD
DDD
+5 more

Location: Hybrid (Bangalore)

Travel: Quarterly travel to Seattle(US)

Education: B.Tech from premium institutes only

Note: Only immediate joiners required/ 0 to 15 Days — no other applications accepted.

Role Summary

We are seeking top-tier Lead Engineers who can design, build, and deliver large-scale distributed systems with high performance, reliability, and operational excellence. The ideal candidate will be a hands-on engineer with expert system design ability, deep understanding of distributed architectures, and strong communication and leadership skills.

The Lead Engineer must be able to convert complex and ambiguous requirements into a fully engineered architecture and implementation plan covering components, data flows, infrastructure, observability, and operations.

Key Responsibilities

1. End-to-End System Architecture

  • Architect scalable, reliable, and secure systems from initial concept through production rollout.
  • Define system boundaries, components, service responsibilities, and integration points.
  • Produce high-level (HLD) and low-level design (LLD) documents.
  • Ensure designs meet performance, reliability, security, and cost objectives.
  • Make informed design trade-offs with solid technical reasoning.

2. Component & Communication Design

  • Break complex systems into independently deployable services.
  • Define APIs, communication contracts, data models, and event schemas.
  • Apply modern architecture patterns such as microservices, event-driven design, DDD, CQRS, and hexagonal architecture.
  • Ensure component clarity, maintainability, and extensibility.

3. Communication Protocol & Middleware

  • Design both sync and async communication layers: REST, RPC, gRPC, message queues, event streams (Kafka/Kinesis/Pulsar).
  • Define retry/timeout strategies, circuit breakers, rate limiting, and versioning strategies.
  • Handle backpressure, partitioning, delivery semantics (at-least/at-most/exactly once).

4. Data Architecture & Storage Strategy

  • Architect data models and storage strategies for SQL and NoSQL databases, distributed caches, blob stores, and search indexes.
  • Define sharding/partitioning, replication, consistency, indexing, backup/restore, and schema evolution strategies.
  • Design real-time and batch data processing pipelines.

5. Operational Readiness

  • Define observability (metrics, logs, traces) requirements.
  • Collaborate with DevOps to ensure deployment, monitoring, alerts, and incident management readiness.
  • Provide production support as a senior technical owner.

6. Leadership & Influence

  • Lead technical discussions, design reviews, and cross-team collaboration.
  • Mentor engineers and help elevate team practices.
  • Influence technology direction and architectural standards.

Required Qualifications

  • 10+ years of professional software engineering experience with strong backend and distributed systems background.
  • Proven track record of leading large-scale architecture and delivery of production systems.
  • Expert in system design with the ability to simplify ambiguity and craft robust solutions.
  • Strong programming experience in one or more languages (Java, Go, Python, C++).
  • Deep understanding of distributed systems, message streaming, queues, RPC/REST, and event-driven architecture.
  • Experience with cloud platforms (AWS/Azure/GCP) and container technologies (Kubernetes/Docker).
  • Strong communication, documentation, and leadership skills.

Preferred Skills

  • Experience with large-scale messaging/streaming (Kafka/Pulsar), caching, and NoSQL.
  • Experience designing for high availability, fault tolerance, and performance at scale.
  • Mentoring and leading global engineering teams.
  • Familiarity with observability tooling (Grafana, Prometheus, Jaeger). 


Read more
Ride-hailing Industry

Ride-hailing Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
5 - 7 yrs
₹42L - ₹45L / yr
DevOps
skill iconPython
Shell Scripting
Infrastructure
Terraform
+16 more

JOB DETAILS:

- Job Title: Senior Devops Engineer 2

- Industry: Ride-hailing

- Experience: 5-7 years

- Working Days: 5 days/week

- Work Mode: ONSITE

- Job Location: Bangalore

- CTC Range: Best in Industry


Required Skills: Cloud & Infrastructure Operations, Kubernetes & Container Orchestration, Monitoring, Reliability & Observability, Proficiency with Terraform, Ansible etc., Strong problem-solving skills with scripting (Python/Go/Shell)

 

Criteria:

1.   Candidate must be from a product-based or scalable app-based start-ups company with experience handling large-scale production traffic.

2.   Minimum 5 yrs of experience working as a DevOps/Infrastructure Consultant

3.   Own end-to-end infrastructure right from non-prod to prod environment including self-managed

4.   Candidate must have experience in database migration from scratch 

5.   Must have a firm hold on the container orchestration tool Kubernetes

6.   Must have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

7.   Understanding programming languages like GO/Python, and Java

8.   Working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

9.   Working experience on Cloud platform - AWS

10. Candidate should have Minimum 1.5 years stability per organization, and a clear reason for relocation.

 

Description 

Job Summary:

As a DevOps Engineer at company, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.

 

Job Responsibilities:

● Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs

● Codify our infrastructure

● Do what it takes to keep the uptime above 99.99%

● Understand the bigger picture and sail through the ambiguities

● Scale technology considering cost and observability and manage end-to-end processes

● Understand DevOps philosophy and evangelize the principles across the organization

● Strong communication and collaboration skills to break down the silos

 

Job Requirements:

● B.Tech. / B.E. degree in Computer Science or equivalent software engineering degree/experience

● Minimum 5 yrs of experience working as a DevOps/Infrastructure Consultant

● Must have a firm hold on the container orchestration tool Kubernetes

● Must have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

● Strong problem-solving skills, and ability to write scripts using any scripting language

● Understanding programming languages like GO/Python, and Java

● Comfortable working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

 

What’s there for you?

Company’s team handles everything – infra, tooling, and self-manages a bunch of databases, such as

● 150+ microservices with event-driven architecture across different tech stacks Golang/ java/ node

● More than 100,000 Request per second on our edge gateways

● ~20,000 events per second on self-managed Kafka

● 100s of TB of data on self-managed databases

● 100s of real-time continuous deployment to production

● Self-managed infra supporting

● 100% OSS

Read more
Ride-hailing Industry

Ride-hailing Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
6 - 9 yrs
₹47L - ₹50L / yr
DevOps
skill iconPython
Shell Scripting
skill iconKubernetes
Terraform
+15 more

JOB DETAILS:

- Job Title: Lead DevOps Engineer

- Industry: Ride-hailing

- Experience: 6-9 years

- Working Days: 5 days/week

- Work Mode: ONSITE

- Job Location: Bangalore

- CTC Range: Best in Industry


Required Skills: Cloud & Infrastructure Operations, Kubernetes & Container Orchestration, Monitoring, Reliability & Observability, Proficiency with Terraform, Ansible etc., Strong problem-solving skills with scripting (Python/Go/Shell)

 

Criteria:

1.   Candidate must be from a product-based or scalable app-based start-ups company with experience handling large-scale production traffic.

2.   Minimum 6 yrs of experience working as a DevOps/Infrastructure Consultant

3.   Candidate must have 2 years of experience as an lead (handling team of 3 to 4 members at least)

4.   Own end-to-end infrastructure right from non-prod to prod environment including self-managed

5.   Candidate must have Self experience in database migration from scratch 

6.   Must have a firm hold on the container orchestration tool Kubernetes

7.   Should have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

8.   Understanding programming languages like GO/Python, and Java

9.   Working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

10.   Working experience on Cloud platform -AWS

11. Candidate should have Minimum 1.5 years stability per organization, and a clear reason for relocation.

 

Description

Job Summary:

As a DevOps Engineer at company, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.

 

Job Responsibilities:

● Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs

● Codify our infrastructure

● Do what it takes to keep the uptime above 99.99%

● Understand the bigger picture and sail through the ambiguities

● Scale technology considering cost and observability and manage end-to-end processes

● Understand DevOps philosophy and evangelize the principles across the organization

● Strong communication and collaboration skills to break down the silos

 

Job Requirements:

● B.Tech. / B.E. degree in Computer Science or equivalent software engineering degree/experience

● Minimum 6 yrs of experience working as a DevOps/Infrastructure Consultant

● Must have a firm hold on the container orchestration tool Kubernetes

● Must have expertise in configuration management tools like Ansible, Terraform, Chef / Puppet

● Strong problem-solving skills, and ability to write scripts using any scripting language

● Understanding programming languages like GO/Python, and Java

● Comfortable working on databases like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

 

What’s there for you?

Company’s team handles everything – infra, tooling, and self-manages a bunch of databases, such as

● 150+ microservices with event-driven architecture across different tech stacks Golang/ java/ node

● More than 100,000 Request per second on our edge gateways

● ~20,000 events per second on self-managed Kafka

● 100s of TB of data on self-managed databases

● 100s of real-time continuous deployment to production

● Self-managed infra supporting

● 100% OSS

Read more
Ride-hailing Industry

Ride-hailing Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
4 - 6 yrs
₹34L - ₹37L / yr
DevOps
skill iconPython
Shell Scripting
skill iconKubernetes
Monitoring
+18 more

JOB DETAILS:

- Job Title: Senior Devops Engineer 1

- Industry: Ride-hailing

- Experience: 4-6 years

- Working Days: 5 days/week

- Work Mode: ONSITE

- Job Location: Bangalore

- CTC Range: Best in Industry


Required Skills: Cloud & Infrastructure Operations, Kubernetes & Container Orchestration, Monitoring, Reliability & Observability, Proficiency with Terraform, Ansible etc., Strong problem-solving skills with scripting (Python/Go/Shell)

 

Criteria:

1. Candidate must be from a product-based or scalable app-based startups company with experience handling large-scale production traffic.

2. Candidate must have strong Linux expertise with hands-on production troubleshooting and working knowledge of databases and middleware (Mongo, Redis, Cassandra, Elasticsearch, Kafka).

3. Candidate must have solid experience with Kubernetes.

4. Candidate should have strong knowledge of configuration management tools like Ansible, Terraform, and Chef / Puppet. Add on- Prometheus & Grafana etc.

5. Candidate must be an individual contributor with strong ownership.

6. Candidate must have hands-on experience with DATABASE MIGRATIONS and observability tools such as Prometheus and Grafana.

7. Candidate must have working knowledge of Go/Python and Java.

8. Candidate should have working experience on Cloud platform - AWS

9. Candidate should have Minimum 1.5 years stability per organization, and a clear reason for relocation.

 

Description 

Job Summary:

As a DevOps Engineer at company, you will be working on building and operating infrastructure at scale, designing and implementing a variety of tools to enable product teams to build and deploy their services independently, improving observability across the board, and designing for security, resiliency, availability, and stability. If the prospect of ensuring system reliability at scale and exploring cutting-edge technology to solve problems, excites you, then this is your fit.

 

Job Responsibilities:

- Own end-to-end infrastructure right from non-prod to prod environment including self-managed DBs.

- Understanding the needs of stakeholders and conveying this to developers.

- Working on ways to automate and improve development and release processes.

- Identifying technical problems and developing software updates and ‘fixes’.

- Working with software developers to ensure that development follows established processes and works as intended.

- Do what it takes to keep the uptime above 99.99%.

- Understand DevOps philosophy and evangelize the principles across the organization.

- Strong communication and collaboration skills to break down the silos

 

Job Requirements:

- B.Tech. / B.E. degree in Computer Science or equivalent software engineering degree/experience.

- Minimum 4 yrs of experience working as a DevOps/Infrastructure Consultant.

- Strong background in operating systems like Linux.

- Understands the container orchestration tool Kubernetes.

- Proficient Knowledge of configuration management tools like Ansible, Terraform, and Chef / Puppet. Add on- Prometheus & Grafana etc.

- Problem-solving attitude, and ability to write scripts using any scripting language.

- Understanding programming languages like GO/Python, and Java.

- Basic understanding of databases and middlewares like Mongo/Redis/Cassandra/Elasticsearch/Kafka.

- Should be able to take ownership of tasks, and must be responsible. - Good communication skills

 

Read more
Bengaluru (Bangalore)
5 - 10 yrs
₹25L - ₹50L / yr
skill iconNodeJS (Node.js)
skill iconReact.js
skill iconPython
skill iconJava
Data engineering
+10 more

Job Title : Senior Software Engineer (Full Stack — AI/ML & Data Applications)

Experience : 5 to 10 Years

Location : Bengaluru, India

Employment Type : Full-Time | Onsite


Role Overview :

We are seeking a Senior Full Stack Software Engineer with strong technical leadership and hands-on expertise in AI/ML, data-centric applications, and scalable full-stack architectures.

In this role, you will design and implement complex applications integrating ML/AI models, lead full-cycle development, and mentor engineering teams.


Mandatory Skills :

Full Stack Development (React/Angular/Vue + Node.js/Python/Java), Data Engineering (Spark/Kafka/ETL), ML/AI Model Integration (TensorFlow/PyTorch/scikit-learn), Cloud & DevOps (AWS/GCP/Azure, Docker, Kubernetes, CI/CD), SQL/NoSQL Databases (PostgreSQL/MongoDB).


Key Responsibilities :

  • Architect, design, and develop scalable full-stack applications for data and AI-driven products.
  • Build and optimize data ingestion, processing, and pipeline frameworks for large datasets.
  • Deploy, integrate, and scale ML/AI models in production environments.
  • Drive system design, architecture discussions, and API/interface standards.
  • Ensure engineering best practices across code quality, testing, performance, and security.
  • Mentor and guide junior developers through reviews and technical decision-making.
  • Collaborate cross-functionally with product, design, and data teams to align solutions with business needs.
  • Monitor, diagnose, and optimize performance issues across the application stack.
  • Maintain comprehensive technical documentation for scalability and knowledge-sharing.

Required Skills & Experience :

  • Education : B.E./B.Tech/M.E./M.Tech in Computer Science, Data Science, or equivalent fields.
  • Experience : 5+ years in software development with at least 2+ years in a senior or lead role.
  • Full Stack Proficiency :
  • Front-end : React / Angular / Vue.js
  • Back-end : Node.js / Python / Java
  • Data Engineering : Experience with data frameworks such as Apache Spark, Kafka, and ETL pipeline development.
  • AI/ML Expertise : Practical exposure to TensorFlow, PyTorch, or scikit-learn and deploying ML models at scale.
  • Databases : Strong knowledge of SQL & NoSQL systems (PostgreSQL, MongoDB) and warehousing tools (Snowflake, BigQuery).
  • Cloud & DevOps : Working knowledge of AWS, GCP, or Azure; containerization & orchestration (Docker, Kubernetes); CI/CD; MLflow/SageMaker is a plus.
  • Visualization : Familiarity with modern data visualization tools (D3.js, Tableau, Power BI).

Soft Skills :

  • Excellent communication and cross-functional collaboration skills.
  • Strong analytical mindset with structured problem-solving ability.
  • Self-driven with ownership mentality and adaptability in fast-paced environments.

Preferred Qualifications (Bonus) :

  • Experience deploying distributed, large-scale ML or data-driven platforms.
  • Understanding of data governance, privacy, and security compliance.
  • Exposure to domain-driven data/AI use cases in fintech, healthcare, retail, or e-commerce.
  • Experience working in Agile environments (Scrum/Kanban).
  • Active open-source contributions or a strong GitHub technical portfolio.
Read more
Technology Industry

Technology Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
6 - 9 yrs
₹30L - ₹48L / yr
skill iconPython
skill iconReact.js
skill iconNodeJS (Node.js)
TypeScript
ReAct (Reason + Act)
+13 more

Review Criteria:

  • Strong Software Engineer fullstack profile using NodeJS / Python and React
  • 6+ YOE in Software Development using Python OR NodeJS (For backend) & React (For frontend)
  • Must have strong experience in working on Typescript
  • Must have experience in message-based systems like Kafka, RabbitMq, Redis
  • Databases - PostgreSQL & NoSQL databases like MongoDB
  • Product Companies Only
  • Tier 1 Engineering Institutes (IIT, NIT, BITS, IIIT, DTU or equivalent)

 

Preferred:

  • Experience in Fin-Tech, Payment, POS and Retail products is highly preferred
  • Experience in mentoring, coaching the team.


Role & Responsibilities:

We are currently seeking a Senior Engineer to join our Financial Services team, contributing to the design and development of scalable system.

 

The Ideal Candidate Will Be Able To-

  • Take ownership of delivering performant, scalable and high-quality cloud-based software, both frontend and backend side.
  • Mentor team members to develop in line with product requirements.
  • Collaborate with Senior Architect for design and technology choices for product development roadmap.
  • Do code reviews.


Ideal Candidate:

  • Thorough knowledge of developing cloud-based software including backend APIs and react based frontend.
  • Thorough knowledge of scalable design patterns and message-based systems such as Kafka, RabbitMq, Redis, MongoDB, ORM, SQL etc.
  • Experience with AWS services such as S3, IAM, Lambda etc.
  • Expert level coding skills in Python FastAPI/Django, NodeJs, TypeScript, ReactJs.
  • Eye for user responsive designs on the frontend.


Perks, Benefits and Work Culture:

  • We prioritize people above all else. While we're recognized for our innovative technology solutions, it's our people who drive our success. That’s why we offer a comprehensive and competitive benefits package designed to support your well-being and growth:
  • Medical Insurance with coverage up to INR 8,00,000 for the employee and their family
Read more
Global digital transformation solutions provider

Global digital transformation solutions provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
6 - 9 yrs
₹15L - ₹25L / yr
Data engineering
Apache Kafka
skill iconPython
skill iconAmazon Web Services (AWS)
AWS Lambda
+11 more

Job Details

- Job Title: Lead I - Data Engineering 

- Industry: Global digital transformation solutions provider

- Domain - Information technology (IT)

- Experience Required: 6-9 years

- Employment Type: Full Time

- Job Location: Pune

- CTC Range: Best in Industry


Job Description

Job Title: Senior Data Engineer (Kafka & AWS)

Responsibilities:

  • Develop and maintain real-time data pipelines using Apache Kafka (MSK or Confluent) and AWS services.
  • Configure and manage Kafka connectors, ensuring seamless data flow and integration across systems.
  • Demonstrate strong expertise in the Kafka ecosystem, including producers, consumers, brokers, topics, and schema registry.
  • Design and implement scalable ETL/ELT workflows to efficiently process large volumes of data.
  • Optimize data lake and data warehouse solutions using AWS services such as Lambda, S3, and Glue.
  • Implement robust monitoring, testing, and observability practices to ensure reliability and performance of data platforms.
  • Uphold data security, governance, and compliance standards across all data operations.

 

Requirements:

  • Minimum of 5 years of experience in Data Engineering or related roles.
  • Proven expertise with Apache Kafka and the AWS data stack (MSK, Glue, Lambda, S3, etc.).
  • Proficient in coding with Python, SQL, and Java — with Java strongly preferred.
  • Experience with Infrastructure-as-Code (IaC) tools (e.g., CloudFormation) and CI/CD pipelines.
  • Excellent problem-solvingcommunication, and collaboration skills.
  • Flexibility to write production-quality code in both Python and Java as required.

 

Skills: Aws, Kafka, Python


Must-Haves

Minimum of 5 years of experience in Data Engineering or related roles.

Proven expertise with Apache Kafka and the AWS data stack (MSK, Glue, Lambda, S3, etc.).

Proficient in coding with Python, SQL, and Java — with Java strongly preferred.

Experience with Infrastructure-as-Code (IaC) tools (e.g., CloudFormation) and CI/CD pipelines.

Excellent problem-solving, communication, and collaboration skills.

Flexibility to write production-quality code in both Python and Java as required.

Skills: Aws, Kafka, Python

Notice period - 0 to 15days only

Read more
AI-First Company

AI-First Company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore), Mumbai, Hyderabad, Gurugram
5 - 17 yrs
₹30L - ₹45L / yr
Data engineering
Data architecture
SQL
Data modeling
GCS
+47 more

ROLES AND RESPONSIBILITIES:

You will be responsible for architecting, implementing, and optimizing Dremio-based data Lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.


  • Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
  • Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
  • Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
  • Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
  • Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
  • Support self-service analytics by enabling governed data products and semantic layers.
  • Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
  • Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.


IDEAL CANDIDATE:

  • Bachelor’s or Master’s in Computer Science, Information Systems, or related field.
  • 5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
  • Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
  • Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
  • Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
  • Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
  • Excellent problem-solving, documentation, and stakeholder communication skills.


PREFERRED:

  • Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) and data catalogs (Collibra, Alation, Purview).
  • Exposure to Snowflake, Databricks, or BigQuery environments.
  • Experience in high-tech, manufacturing, or enterprise data modernization programs.
Read more
iMerit
Bengaluru (Bangalore)
6 - 9 yrs
₹10L - ₹15L / yr
DevOps
Terraform
Apache Kafka
skill iconPython
skill iconGo Programming (Golang)
+4 more

Exp: 7- 10 Years

CTC: up to 35 LPA


Skills:

  • 6–10 years DevOps / SRE / Cloud Infrastructure experience
  • Expert-level Kubernetes (networking, security, scaling, controllers)
  • Terraform Infrastructure-as-Code mastery
  • Hands-on Kafka production experience
  • AWS cloud architecture and networking expertise
  • Strong scripting in Python, Go, or Bash
  • GitOps and CI/CD tooling experience


Key Responsibilities:

  • Design highly available, secure cloud infrastructure supporting distributed microservices at scale
  • Lead multi-cluster Kubernetes strategy optimized for GPU and multi-tenant workloads
  • Implement Infrastructure-as-Code using Terraform across full infrastructure lifecycle
  • Optimize Kafka-based data pipelines for throughput, fault tolerance, and low latency
  • Deliver zero-downtime CI/CD pipelines using GitOps-driven deployment models
  • Establish SRE practices with SLOs, p95 and p99 monitoring, and FinOps discipline
  • Ensure production-ready disaster recovery and business continuity testing



If interested Kindly share your updated resume at 82008 31681

Read more
wwwwebnyayai
Ishita Jindal
Posted by Ishita Jindal
Noida
2 - 4 yrs
₹3L - ₹9L / yr
skill iconPython
skill iconDjango
Google Cloud Platform (GCP)
Apache Kafka

About the role

Webnyay is looking for an experienced Backend Developer to build and scale reliable backend systems for our legal tech platform. You will work on core product architecture, high-performance APIs, and cloud-native services that support AI-driven workflows and large-scale data processing.


Responsibilities

  • Develop backend services using Python, Django, and FastAPI
  • Build scalable APIs and microservices for product features
  • Implement event-driven and asynchronous workflows using Kafka
  • Design and maintain backend integrations and data pipelines
  • Deploy and manage services on Google Cloud Platform (GCP)
  • Ensure performance, security, and reliability of backend systems
  • Collaborate with product and engineering teams to deliver production-ready features


Requirements

  • 4+ years of backend development experience
  • Strong proficiency in Python
  • Hands-on experience with Django and FastAPI
  • Experience working with Kafka or similar messaging systems
  • Working knowledge of GCP and cloud-based deployments
  • Solid understanding of backend architecture and API design
  • Experience with databases and production systems
  • Experience building SaaS or platform-based products
  • Exposure to AI-driven or data-intensive applications


Why Webnyay

  • Build technology that improves access to justice
  • Work on real-world, high-impact legal tech problems
  • Collaborative and ownership-driven work culture
  • Opportunity to grow with a fast-scaling startup
Read more
Capace Software Private Limited
Bengaluru (Bangalore), Bhopal
5 - 10 yrs
₹4L - ₹10L / yr
skill iconDjango
CI/CD
Software deployment
RESTful APIs
skill iconFlask
+8 more

Senior Python Django Developer 

Experience: Back-end development: 6 years (Required)


Location:  Bangalore/ Bhopal

Job Description:

We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the start-up environment.

This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.

Responsibilities:

  • Design and develop scalable, secure, and high-performance applications using Python (Django framework).
  • Architect system components, define database schemas, and optimize backend services for speed and efficiency.
  • Lead and implement design patterns and software architecture best practices.
  • Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
  • Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
  • Drive performance improvements, monitor system health, and troubleshoot production issues.
  • Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
  • Contribute to technical decision-making and mentor junior developers.

Requirements:

  • 6 to 10 years of professional backend development experience with Python and Django.
  • Strong background in payments/financial systems or FinTech applications.
  • Proven experience in designing software architecture in a microservices or modular monolith environment.
  • Experience working in fast-paced startup environments with agile practices.
  • Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
  • Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
  • Hands-on experience with test-driven development (TDD) and frameworks like pytest, unittest, or factory_boy.
  • Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).

Preferred Skills:

  • Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
  • Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
  • Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
  • Contributions to open-source or personal finance-related projects.

Job Types: Full-time, Permanent


Schedule:

  • Day shift

Supplemental Pay:

  • Performance bonus
  • Yearly bonus

Ability to commute/relocate:

  • JP Nagar, 5th Phase, Bangalore, Karnataka or Indrapuri, Bhopal, Madhya Pradesh: Reliably commute or willing to relocate with an employer-provided relocation package (Preferred)


Read more
Tradelab Software Private Limited
Mumbai
3 - 5 yrs
₹10L - ₹15L / yr
skill iconPython
FastAPI
RabbitMQ
Apache Kafka
skill iconRedis
+1 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.



What You Will Do:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What We Expect:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Experience with GoLang, C/C++, Erlang, or Elixir.

• Exposure to trading, fintech, or low-latency systems.

• Familiarity with microservices and CI/CD pipelines.

Read more
Virtana

at Virtana

3 candid answers
2 recruiters
Krutika Devadiga
Posted by Krutika Devadiga
Pune
4 - 10 yrs
Best in industry
skill iconJava
skill iconKubernetes
skill iconGo Programming (Golang)
skill iconPython
Apache Kafka
+13 more

Senior Software Engineer 

Challenge convention and work on cutting edge technology that is transforming the way our customers manage their physical, virtual and cloud computing environments. Virtual Instruments seeks highly talented people to join our growing team, where your contributions will impact the development and delivery of our product roadmap. Our award-winning Virtana Platform provides the only real-time, system-wide, enterprise scale solution for providing visibility into performance, health and utilization metrics, translating into improved performance and availability while lowering the total cost of the infrastructure supporting mission-critical applications.  

We are seeking an individual with expert knowledge in Systems Management and/or Systems Monitoring Software, Observability platforms and/or Performance Management Software and Solutions with insight into integrated infrastructure platforms like Cisco UCS, infrastructure providers like Nutanix, VMware, EMC & NetApp and public cloud platforms like Google Cloud and AWS to expand the depth and breadth of Virtana Products. 


Work Location: Pune/ Chennai


Job Type: Hybrid

 

Role Responsibilities: 

  • The engineer will be primarily responsible for architecture, design and development of software solutions for the Virtana Platform 
  • Partner and work closely with cross functional teams and with other engineers and product managers to architect, design and implement new features and solutions for the Virtana Platform. 
  • Communicate effectively across the departments and R&D organization having differing levels of technical knowledge.  
  • Work closely with UX Design, Quality Assurance, DevOps and Documentation teams. Assist with functional and system test design and deployment automation 
  • Provide customers with complex and end-to-end application support, problem diagnosis and problem resolution 
  • Learn new technologies quickly and leverage 3rd party libraries and tools as necessary to expedite delivery 

 

Required Qualifications:    

  • Minimum of 7+ years of progressive experience with back-end development in a Client Server Application development environment focused on Systems Management, Systems Monitoring and Performance Management Software. 
  • Deep experience in public cloud environment using Kubernetes and other distributed managed services like Kafka etc (Google Cloud and/or AWS) 
  • Experience with CI/CD and cloud-based software development and delivery 
  • Deep experience with integrated infrastructure platforms and experience working with one or more data collection technologies like SNMP, REST, OTEL, WMI, WBEM. 
  • Minimum of 6 years of development experience with one or more of these high level languages like GO, Python, Java. Deep experience with one of these languages is required. 
  • Bachelor’s or Master’s degree in computer science, Computer Engineering or equivalent 
  • Highly effective verbal and written communication skills and ability to lead and participate in multiple projects 
  • Well versed with identifying opportunities and risks in a fast-paced environment and ability to adjust to changing business priorities 
  • Must be results-focused, team-oriented and with a strong work ethic 

 

Desired Qualifications: 

  • Prior experience with other virtualization platforms like OpenShift is a plus 
  • Prior experience as a contributor to engineering and integration efforts with strong attention to detail and exposure to Open-Source software is a plus 
  • Demonstrated ability as a lead engineer who can architect, design and code with strong communication and teaming skills 
  • Deep development experience with the development of Systems, Network and performance Management Software and/or Solutions is a plus 

  

About Virtana:  Virtana delivers the industry’s only broadest and deepest Observability Platform that allows organizations to monitor infrastructure, de-risk cloud migrations, and reduce cloud costs by 25% or more. 

  

Over 200 Global 2000 enterprise customers, such as AstraZeneca, Dell, Salesforce, Geico, Costco, Nasdaq, and Boeing, have valued Virtana’s software solutions for over a decade. 

  

Our modular platform for hybrid IT digital operations includes Infrastructure Performance Monitoring and Management (IPM), Artificial Intelligence for IT Operations (AIOps), Cloud Cost Management (Fin Ops), and Workload Placement Readiness Solutions. Virtana is simplifying the complexity of hybrid IT environments with a single cloud-agnostic platform across all the categories listed above. The $30B IT Operations Management (ITOM) Software market is ripe for disruption, and Virtana is uniquely positioned for success. 

Read more
DigitalSprint AI Solutions

at DigitalSprint AI Solutions

1 candid answer
Nalini Sanka
Posted by Nalini Sanka
Bengaluru (Bangalore)
6 - 12 yrs
₹15L - ₹25L / yr
skill iconJava
skill iconSpring Boot
skill iconAmazon Web Services (AWS)
Apache Kafka
skill iconJenkins

Requirements

  • 6–12 years of backend development experience.
  • Strong expertise in Java 11+, Spring Boot, REST APIs, AWS.
  • Solid experience with distributed, high-volume systems.
  • Strong knowledge of RDBMS (e.g., MySQL, Oracle) and NoSQL databases (e.g., DynamoDB, MongoDB, Cassandra).
  • Hands-on with CI/CD (Jenkins) and caching technologies Redis or Similar.
  • Strong debugging and system troubleshooting skills.
  • Experience in payments system is a must.


Read more
Tradelab Technologies

at Tradelab Technologies

1 candid answer
Aakanksha Yadav
Posted by Aakanksha Yadav
Bengaluru (Bangalore), Mumbai
2 - 4 yrs
₹10L - ₹15L / yr
MySQL
skill iconRedis
Apache Kafka

We're Hiring for Golang Developer (2–4 years experience)

Company: Tradelab Technologies

Location: Bangalore (Preferred candidates from Bangalore or nearby locations only)


We are looking for a Golang Developer with strong experience in backend development, microservices, and system-level programming. The role involves contributing to high-performance trading systems, low-latency architecture, and scalable backend solutions.


Key Responsibilities

• Develop and maintain backend services using Golang

• Build scalable, secure and high-performance microservices

• Work with REST APIs, WebSockets, message queues and distributed systems

• Collaborate with DevOps, frontend and product teams for smooth delivery

• Optimize performance, troubleshoot issues and ensure system stability


Skills & Experience Required

• 2–4 years of experience in Golang development

• Strong fundamentals in data structures, concurrency and networking

• Experience with MySQL/Redis/Kafka or similar technologies

• Understanding of microservices, APIs and cloud environments

• Experience in fintech/trading systems is a plus


If you are passionate about backend engineering and want to work on fast, scalable trading systems.

Read more
Tradelab Technologies

at Tradelab Technologies

1 candid answer
Aakanksha Yadav
Posted by Aakanksha Yadav
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
skill iconPython
RestAPI
FastAPI
RabbitMQ
Apache Kafka
+3 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What We Expect:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and

a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex

performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What You Will Do:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Experience with GoLang, C/C++, Erlang, or Elixir.

• Exposure to trading, fintech, or low-latency systems.

• Familiarity with microservices and CI/CD pipelines.



Read more
AI company

AI company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore), Mumbai, Hyderabad, Gurugram
5 - 17 yrs
₹30L - ₹45L / yr
Data architecture
Data engineering
SQL
Data modeling
GCS
+21 more

Review Criteria

  • Strong Dremio / Lakehouse Data Architect profile
  • 5+ years of experience in Data Architecture / Data Engineering, with minimum 3+ years hands-on in Dremio
  • Strong expertise in SQL optimization, data modeling, query performance tuning, and designing analytical schemas for large-scale systems
  • Deep experience with cloud object storage (S3 / ADLS / GCS) and file formats such as Parquet, Delta, Iceberg along with distributed query planning concepts
  • Hands-on experience integrating data via APIs, JDBC, Delta/Parquet, object storage, and coordinating with data engineering pipelines (Airflow, DBT, Kafka, Spark, etc.)
  • Proven experience designing and implementing lakehouse architecture including ingestion, curation, semantic modeling, reflections/caching optimization, and enabling governed analytics
  • Strong understanding of data governance, lineage, RBAC-based access control, and enterprise security best practices
  • Excellent communication skills with ability to work closely with BI, data science, and engineering teams; strong documentation discipline
  • Candidates must come from enterprise data modernization, cloud-native, or analytics-driven companies


Preferred

  • Preferred (Nice-to-have) – Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) or data catalogs (Collibra, Alation, Purview); familiarity with Snowflake, Databricks, or BigQuery environments


Job Specific Criteria

  • CV Attachment is mandatory
  • How many years of experience you have with Dremio?
  • Which is your preferred job location (Mumbai / Bengaluru / Hyderabad / Gurgaon)?
  • Are you okay with 3 Days WFO?
  • Virtual Interview requires video to be on, are you okay with it?


Role & Responsibilities

You will be responsible for architecting, implementing, and optimizing Dremio-based data lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.

  • Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
  • Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
  • Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
  • Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
  • Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
  • Support self-service analytics by enabling governed data products and semantic layers.
  • Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
  • Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.


Ideal Candidate

  • Bachelor’s or master’s in computer science, Information Systems, or related field.
  • 5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
  • Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
  • Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
  • Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
  • Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
  • Excellent problem-solving, documentation, and stakeholder communication skills.
Read more
Appiness Interactive
Remote only
6 - 10 yrs
₹10L - ₹14L / yr
skill iconPython
skill iconDjango
FastAPI
skill iconFlask
pandas
+9 more

Position Overview: The Lead Software Architect - Python & Data Engineering is a senior technical leadership role responsible for designing and owning end-to-end architecture for data-intensive, AI/ML, and analytics platforms, while mentoring developers and ensuring technical excellence across the organization. 


Key Responsibilities: 

  • Design end-to-end software architecture for data-intensive applications, AI/ML pipelines, and analytics platforms
  • Evaluate trade-offs between competing technical approaches 
  • Define data models, API approach, and integration patterns across systems 
  • Create technical specifications and architecture documentation 
  • Lead by example through production-grade Python code and mentor developers on engineering fundamentals 
  • Conduct design and code reviews focused on architectural soundness 
  • Establish engineering standards, coding practices, and design patterns for the team 
  • Translate business requirements into technical architecture 
  • Collaborate with data scientists, analysts, and other teams to design integrated solutions 
  • Whiteboard and defend system design and architectural choices 
  • Take responsibility for system performance, reliability, and maintainability 
  • Identify and resolve architectural bottlenecks proactively 


Required Skills:  

  • 8+ years of experience in software architecture and development  
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field 
  • Strong foundations in data structures, algorithms, and computational complexity 
  • Experience in system design for scale, including caching strategies, load balancing, and asynchronous processing  
  • 6+ years of Python development experience 
  • Deep knowledge of Django, Flask, or FastAPI 
  • Expert understanding of Python internals including GIL and memory management 
  • Experience with RESTful API design and event-driven architectures (Kafka, RabbitMQ) 
  • Proficiency in data processing frameworks such as Pandas, Apache Spark, and Airflow 
  • Strong SQL optimization and database design experience (PostgreSQL, MySQL, MongoDB)  Experience with AWS, GCP, or Azure cloud platforms 
  • Knowledge of containerization (Docker) and orchestration (Kubernetes) 
  • Hands-on experience designing CI/CD pipelines Preferred (Bonus) 


Skills

  • Experience deploying ML models to production (MLOps, model serving, monitoring) Understanding of ML system design including feature stores and model versioning 
  • Familiarity with ML frameworks such as scikit-learn, TensorFlow, and PyTorch  
  • Open-source contributions or technical blogging demonstrating architectural depth 
  • Experience with modern front-end frameworks for full-stack perspective


Read more
Tradelab Technologies

at Tradelab Technologies

1 candid answer
Aakanksha Yadav
Posted by Aakanksha Yadav
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
skill iconGo Programming (Golang)
CI/CD
Apache Kafka
RabbitMQ
skill iconDocker
+1 more

Location: Bengalore, India, Exp: 3-5 Yrs

Backend Developer (Golang) - Trading & Fintech


About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What we expect:

 You should already be exceptional at Golang. If you need hand-holding, this isn’t the place for you.

 You thrive on challenges, not on perks or financial rewards.

 You measure success by your own growth, not external validation.

 Taking calculated risks excites you—you’re here to build, break, and learn.

 You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading environment.

 You understand the stakes—milliseconds can make or break trades, and precision is everything.



What you will do:

 Develop and optimize high-performance backend systems in Golang for trading platforms and financial services.

 Architect low-latency, high-throughput microservices that push the boundaries ofspeed and efficiency.

 Build event-driven, fault-tolerant systems that can handle massive real-time data streams.

 Own your work—no babysitting, no micromanagement.

 Work alongside equally driven engineers who expect nothing less than brilliance.


Must have skills:

 Learn faster than you ever thought possible.

Proven expertise in Golang (if you need to prove yourself, this isn’t the role for you).

 Deep understanding of concurrency, memory management, and system design.

 Experience with Trading, market data processing, or low-latency systems.

 Strong knowledge of distributed systems, message queues (Kafka, RabbitMQ), and real-time processing.

 Hands-on with Docker, Kubernetes, and CI/CD pipelines.

 A portfolio of work that speaks louder than a resume.


Nice-to-Have Skills:

 Past experience in fintech, trading systems, or algorithmic trading.

 Contributions to open-source Golang projects.

 A history of building something impactful from scratch.

 Understanding of FIX protocol, WebSockets, and streaming APIs.


Why Join Us?

 Work with a team that expects and delivers excellence.

 A culture where risk-taking is rewarded, and complacency is not.

 Limitless opportunities for growth—if you can handle the pace.

 A place where learning is currency, and outperformance is the only metric that matters.


The opportunity to build systems that move markets, execute trades in microseconds, and redefine fintech. This isn’t just a job—it’s a proving ground. Ready to take the leap? Apply now.



Read more
prep study
Pooja Sharma
Posted by Pooja Sharma
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
FastAPI
RESTful APIs
RabbitMQ
Apache Kafka
skill iconElastic Search
+2 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.

What We Expect:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.

Read more
Bengaluru (Bangalore)
4 - 7 yrs
₹15L - ₹35L / yr
skill iconJava
skill iconSpring Boot
Apache Kafka

SE/SSE - Backend (Java)



JAVA: Software Engineer / Senior Software Engineer


Job Overview: RevSure.AI is building a world-class team of Engineers with a mandate to architect, design, build, scale, and maintain our cutting-edge Revops platform and derive insights from the customer data. If you are a great JavaScript developer (or, for that matter, in any major programming language) with experience in building SaaS web applications and looking for an opportunity to build world-class products using cutting-edge technologies, please read on -


Responsibilities:

● Contribute to the growth and development of our product by writing and maintaining code, focusing on quality and maintainability. We believe that our teams own the code they build.

● Share your knowledge, both inside and outside your own team.

● Design, implement, test, and monitor valuable solutions that achieve the teamʼs goals while keeping a smooth delivery flow and balancing tradeoffs between scope, time, and effort.

● Continuously improve your and your teamʼs way of working, expanding the boundaries of your teamʼs autonomy and self-organisation.


Requirements:


Must-Have:

● Good interpersonal and communication skills. An analytical mind and an eye for detail

● Experience of 4-7 years in building large applications, preferably from the heavy data processing & analytics side

● Experience in architecting, designing, and building end-to-end System Design of multi tenant SAAS applications

● Experience architecting interactive configuration experiences around DAGs, configuration-heavy reporting feature, interactive data pipeline design interfaces, and charting.

● Experience in Java, Spring Boot, and GCP

● In-depth understanding of design patterns, OOPs, and Functional programming.

● Passionate programmer focused on Backend Programming.


Good to have:

● Working knowledge of GCP managed services like Kubernetes, PubSub


Please feel free to share your resume to chandana at revsure.ai


**Based on your experience, the title could vary. We care more about mindset , ownership, and craft than just titles.

Read more
Whiz IT Services
Sheeba Harish
Posted by Sheeba Harish
Remote only
10 - 15 yrs
₹20L - ₹20L / yr
skill iconJava
skill iconSpring Boot
Microservices
API
Apache Kafka
+5 more

We are looking for highly experienced Senior Java Developers who can architect, design, and deliver high-performance enterprise applications using Spring Boot and Microservices . The role requires a strong understanding of distributed systems, scalability, and data consistency.

Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
6 - 12 yrs
₹25L - ₹30L / yr
skill iconMachine Learning (ML)
AWS CloudFormation
Online machine learning
skill iconAmazon Web Services (AWS)
ECS
+20 more

MUST-HAVES: 

  • Machine Learning + Aws + (EKS OR ECS OR Kubernetes) + (Redshift AND Glue) + Sage maker
  • Notice period - 0 to 15 days only 
  • Hybrid work mode- 3 days office, 2 days at home


SKILLS: AWS, AWS CLOUD, AMAZON REDSHIFT, EKS


ADDITIONAL GUIDELINES:

  • Interview process: - 2 Technical round + 1 Client round
  • 3 days in office, Hybrid model. 


CORE RESPONSIBILITIES:

  • The MLE will design, build, test, and deploy scalable machine learning systems, optimizing model accuracy and efficiency
  • Model Development: Algorithms and architectures span traditional statistical methods to deep learning along with employing LLMs in modern frameworks.
  • Data Preparation: Prepare, cleanse, and transform data for model training and evaluation.
  • Algorithm Implementation: Implement and optimize machine learning algorithms and statistical models.
  • System Integration: Integrate models into existing systems and workflows.
  • Model Deployment: Deploy models to production environments and monitor performance.
  • Collaboration: Work closely with data scientists, software engineers, and other stakeholders.
  • Continuous Improvement: Identify areas for improvement in model performance and systems.


SKILLS:

  • Programming and Software Engineering: Knowledge of software engineering best practices (version control, testing, CI/CD).
  • Data Engineering: Ability to handle data pipelines, data cleaning, and feature engineering. Proficiency in SQL for data manipulation + Kafka, Chaos search logs, etc. for troubleshooting; Other tech touch points are Scylla DB (like BigTable), OpenSearch, Neo4J graph
  • Model Deployment and Monitoring: MLOps Experience in deploying ML models to production environments.
  • Knowledge of model monitoring and performance evaluation.


REQUIRED EXPERIENCE:

  • Amazon SageMaker: Deep understanding of SageMaker's capabilities for building, training, and deploying ML models; understanding of the Sage maker pipeline with ability to analyze gaps and recommend/implement improvements
  • AWS Cloud Infrastructure: Familiarity with S3, EC2, Lambda and using these services in ML workflows
  • AWS data: Redshift, Glue
  • Containerization and Orchestration: Understanding of Docker and Kubernetes, and their implementation within AWS (EKS, ECS)
Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort