Cutshort logo
Apache kafka jobs

50+ Apache Kafka Jobs in India

Apply to 50+ Apache Kafka Jobs on CutShort.io. Find your next job, effortlessly. Browse Apache Kafka Jobs and apply today!

icon
Capace Software Private Limited
Bengaluru (Bangalore), Bhopal
5 - 10 yrs
₹4L - ₹10L / yr
skill iconDjango
CI/CD
Software deployment
RESTful APIs
skill iconFlask
+8 more

Senior Python Django Developer 

Experience: Back-end development: 6 years (Required)


Location:  Bangalore/ Bhopal

Job Description:

We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the start-up environment.

This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.

Responsibilities:

  • Design and develop scalable, secure, and high-performance applications using Python (Django framework).
  • Architect system components, define database schemas, and optimize backend services for speed and efficiency.
  • Lead and implement design patterns and software architecture best practices.
  • Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
  • Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
  • Drive performance improvements, monitor system health, and troubleshoot production issues.
  • Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
  • Contribute to technical decision-making and mentor junior developers.

Requirements:

  • 6 to 10 years of professional backend development experience with Python and Django.
  • Strong background in payments/financial systems or FinTech applications.
  • Proven experience in designing software architecture in a microservices or modular monolith environment.
  • Experience working in fast-paced startup environments with agile practices.
  • Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
  • Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
  • Hands-on experience with test-driven development (TDD) and frameworks like pytest, unittest, or factory_boy.
  • Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).

Preferred Skills:

  • Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
  • Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
  • Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
  • Contributions to open-source or personal finance-related projects.

Job Types: Full-time, Permanent


Schedule:

  • Day shift

Supplemental Pay:

  • Performance bonus
  • Yearly bonus

Ability to commute/relocate:

  • JP Nagar, 5th Phase, Bangalore, Karnataka or Indrapuri, Bhopal, Madhya Pradesh: Reliably commute or willing to relocate with an employer-provided relocation package (Preferred)


Read more
Tradelab Software Private Limited
Mumbai
3 - 5 yrs
₹10L - ₹15L / yr
skill iconPython
FastAPI
RabbitMQ
Apache Kafka
skill iconRedis
+1 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.



What You Will Do:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What We Expect:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Experience with GoLang, C/C++, Erlang, or Elixir.

• Exposure to trading, fintech, or low-latency systems.

• Familiarity with microservices and CI/CD pipelines.

Read more
Virtana

at Virtana

2 candid answers
Krutika Devadiga
Posted by Krutika Devadiga
Pune
4 - 10 yrs
Best in industry
skill iconJava
skill iconKubernetes
skill iconGo Programming (Golang)
skill iconPython
Apache Kafka
+13 more

Senior Software Engineer 

Challenge convention and work on cutting edge technology that is transforming the way our customers manage their physical, virtual and cloud computing environments. Virtual Instruments seeks highly talented people to join our growing team, where your contributions will impact the development and delivery of our product roadmap. Our award-winning Virtana Platform provides the only real-time, system-wide, enterprise scale solution for providing visibility into performance, health and utilization metrics, translating into improved performance and availability while lowering the total cost of the infrastructure supporting mission-critical applications.  

We are seeking an individual with expert knowledge in Systems Management and/or Systems Monitoring Software, Observability platforms and/or Performance Management Software and Solutions with insight into integrated infrastructure platforms like Cisco UCS, infrastructure providers like Nutanix, VMware, EMC & NetApp and public cloud platforms like Google Cloud and AWS to expand the depth and breadth of Virtana Products. 


Work Location: Pune/ Chennai


Job Type: Hybrid

 

Role Responsibilities: 

  • The engineer will be primarily responsible for architecture, design and development of software solutions for the Virtana Platform 
  • Partner and work closely with cross functional teams and with other engineers and product managers to architect, design and implement new features and solutions for the Virtana Platform. 
  • Communicate effectively across the departments and R&D organization having differing levels of technical knowledge.  
  • Work closely with UX Design, Quality Assurance, DevOps and Documentation teams. Assist with functional and system test design and deployment automation 
  • Provide customers with complex and end-to-end application support, problem diagnosis and problem resolution 
  • Learn new technologies quickly and leverage 3rd party libraries and tools as necessary to expedite delivery 

 

Required Qualifications:    

  • Minimum of 7+ years of progressive experience with back-end development in a Client Server Application development environment focused on Systems Management, Systems Monitoring and Performance Management Software. 
  • Deep experience in public cloud environment using Kubernetes and other distributed managed services like Kafka etc (Google Cloud and/or AWS) 
  • Experience with CI/CD and cloud-based software development and delivery 
  • Deep experience with integrated infrastructure platforms and experience working with one or more data collection technologies like SNMP, REST, OTEL, WMI, WBEM. 
  • Minimum of 6 years of development experience with one or more of these high level languages like GO, Python, Java. Deep experience with one of these languages is required. 
  • Bachelor’s or Master’s degree in computer science, Computer Engineering or equivalent 
  • Highly effective verbal and written communication skills and ability to lead and participate in multiple projects 
  • Well versed with identifying opportunities and risks in a fast-paced environment and ability to adjust to changing business priorities 
  • Must be results-focused, team-oriented and with a strong work ethic 

 

Desired Qualifications: 

  • Prior experience with other virtualization platforms like OpenShift is a plus 
  • Prior experience as a contributor to engineering and integration efforts with strong attention to detail and exposure to Open-Source software is a plus 
  • Demonstrated ability as a lead engineer who can architect, design and code with strong communication and teaming skills 
  • Deep development experience with the development of Systems, Network and performance Management Software and/or Solutions is a plus 

  

About Virtana:  Virtana delivers the industry’s only broadest and deepest Observability Platform that allows organizations to monitor infrastructure, de-risk cloud migrations, and reduce cloud costs by 25% or more. 

  

Over 200 Global 2000 enterprise customers, such as AstraZeneca, Dell, Salesforce, Geico, Costco, Nasdaq, and Boeing, have valued Virtana’s software solutions for over a decade. 

  

Our modular platform for hybrid IT digital operations includes Infrastructure Performance Monitoring and Management (IPM), Artificial Intelligence for IT Operations (AIOps), Cloud Cost Management (Fin Ops), and Workload Placement Readiness Solutions. Virtana is simplifying the complexity of hybrid IT environments with a single cloud-agnostic platform across all the categories listed above. The $30B IT Operations Management (ITOM) Software market is ripe for disruption, and Virtana is uniquely positioned for success. 

Read more
DigitalSprint AI Solutions

at DigitalSprint AI Solutions

1 candid answer
Nalini Sanka
Posted by Nalini Sanka
Bengaluru (Bangalore)
6 - 12 yrs
₹15L - ₹25L / yr
skill iconJava
skill iconSpring Boot
skill iconAmazon Web Services (AWS)
Apache Kafka
skill iconJenkins

Requirements

  • 6–12 years of backend development experience.
  • Strong expertise in Java 11+, Spring Boot, REST APIs, AWS.
  • Solid experience with distributed, high-volume systems.
  • Strong knowledge of RDBMS (e.g., MySQL, Oracle) and NoSQL databases (e.g., DynamoDB, MongoDB, Cassandra).
  • Hands-on with CI/CD (Jenkins) and caching technologies Redis or Similar.
  • Strong debugging and system troubleshooting skills.
  • Experience in payments system is a must.


Read more
Tradelab Technologies
Bengaluru (Bangalore)
2 - 4 yrs
₹10L - ₹15L / yr
MySQL
skill iconRedis
Apache Kafka

We're Hiring for Golang Developer (2–4 years experience)

Company: Tradelab Technologies

Location: Bangalore (Preferred candidates from Bangalore or nearby locations only)


We are looking for a Golang Developer with strong experience in backend development, microservices, and system-level programming. The role involves contributing to high-performance trading systems, low-latency architecture, and scalable backend solutions.


Key Responsibilities

• Develop and maintain backend services using Golang

• Build scalable, secure and high-performance microservices

• Work with REST APIs, WebSockets, message queues and distributed systems

• Collaborate with DevOps, frontend and product teams for smooth delivery

• Optimize performance, troubleshoot issues and ensure system stability


Skills & Experience Required

• 2–4 years of experience in Golang development

• Strong fundamentals in data structures, concurrency and networking

• Experience with MySQL/Redis/Kafka or similar technologies

• Understanding of microservices, APIs and cloud environments

• Experience in fintech/trading systems is a plus


If you are passionate about backend engineering and want to work on fast, scalable trading systems.


Read more
Tradelab Technologies
Aakanksha Yadav
Posted by Aakanksha Yadav
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
skill iconPython
RestAPI
FastAPI
RabbitMQ
Apache Kafka
+3 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What We Expect:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and

a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex

performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What You Will Do:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Experience with GoLang, C/C++, Erlang, or Elixir.

• Exposure to trading, fintech, or low-latency systems.

• Familiarity with microservices and CI/CD pipelines.



Read more
Inflectionio
Renu Philip
Posted by Renu Philip
Remote only
5 - 8 yrs
₹48L - ₹60L / yr
skill iconJava
skill iconSpring Boot
Microservices
skill iconPython
skill iconGo Programming (Golang)
+4 more

Inflection.io is a venture-backed B2B marketing automation company, enabling to communicate with their customers and prospects from one platform. We're used by leading SaaS companies like Sauce Labs, Sigma Computing, BILL, Mural, and Elastic, many of which pay more than $100K/yr (1 crore rupee).

And,... it’s working! We have world-class stats: our largest deal is over 3 crore, we have a 5 star rating on G2, over 100% NRR, and constantly break sales and customer records. We’ve raised $14M in total since 2021 with $7.6M of fresh funding in 2024, giving us many years of runway.

However, we’re still in startup mode with approximately 30 employees and looking for the next SDE3 to help propel Inflection forward. Do you want to join a fast growing startup that is aiming to build a very large company?

Key Responsibilities:

  • Lead the design, development, and deployment of complex software systems and applications.
  • Collaborate with engineers and product managers to define and implement innovative solutions
  • Provide technical leadership and mentorship to junior engineers, promoting best practices and fostering a culture of continuous improvement.
  • Write clean, maintainable and efficient code, ensuring high performance and scalability of the software.
  • Conduct code reviews and provide constructive feedback to ensure code quality and adherence to coding standards.
  • Troubleshoot and resolve complex technical issues, optimizing system performance and reliability.
  • Stay updated with the latest industry trends and technologies, evaluating their potential for adoption in our projects.
  • Participate in the full software development lifecycle, from requirements gathering to deployment and monitoring.

Qualifications:

  • 5+ years of professional software development experience, with a strong focus on backend development.
  • Proficiency in one or more programming languages such as Java, Python, Golang or C#
  • Strong understanding of database systems, both relational (e.g., MySQL, PostgreSQL) and NoSQL (e.g., MongoDB, Cassandra).
  • Hands-on experience with message brokers such as Kafka, RabbitMQ, or Amazon SQS.
  • Experience with cloud platforms (AWS or Azure or Google Cloud) and containerization technologies (Docker, Kubernetes).
  • Proven track record of designing and implementing scalable, high-performance systems.
  • Excellent problem-solving skills and the ability to think critically and creatively.
  • Strong communication and collaboration skills, with the ability to work effectively in a fast-paced, team-oriented environment.


Read more
AI company

AI company

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore)
5 - 10 yrs
₹20L - ₹45L / yr
Data architecture
Data engineering
SQL
Data modeling
GCS
+21 more

Review Criteria

  • Strong Dremio / Lakehouse Data Architect profile
  • 5+ years of experience in Data Architecture / Data Engineering, with minimum 3+ years hands-on in Dremio
  • Strong expertise in SQL optimization, data modeling, query performance tuning, and designing analytical schemas for large-scale systems
  • Deep experience with cloud object storage (S3 / ADLS / GCS) and file formats such as Parquet, Delta, Iceberg along with distributed query planning concepts
  • Hands-on experience integrating data via APIs, JDBC, Delta/Parquet, object storage, and coordinating with data engineering pipelines (Airflow, DBT, Kafka, Spark, etc.)
  • Proven experience designing and implementing lakehouse architecture including ingestion, curation, semantic modeling, reflections/caching optimization, and enabling governed analytics
  • Strong understanding of data governance, lineage, RBAC-based access control, and enterprise security best practices
  • Excellent communication skills with ability to work closely with BI, data science, and engineering teams; strong documentation discipline
  • Candidates must come from enterprise data modernization, cloud-native, or analytics-driven companies


Preferred

  • Preferred (Nice-to-have) – Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) or data catalogs (Collibra, Alation, Purview); familiarity with Snowflake, Databricks, or BigQuery environments


Job Specific Criteria

  • CV Attachment is mandatory
  • How many years of experience you have with Dremio?
  • Which is your preferred job location (Mumbai / Bengaluru / Hyderabad / Gurgaon)?
  • Are you okay with 3 Days WFO?
  • Virtual Interview requires video to be on, are you okay with it?


Role & Responsibilities

You will be responsible for architecting, implementing, and optimizing Dremio-based data lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.

  • Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
  • Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
  • Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
  • Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
  • Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
  • Support self-service analytics by enabling governed data products and semantic layers.
  • Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
  • Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.


Ideal Candidate

  • Bachelor’s or master’s in computer science, Information Systems, or related field.
  • 5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
  • Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
  • Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
  • Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
  • Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
  • Excellent problem-solving, documentation, and stakeholder communication skills.


Read more
Appiness Interactive
Remote only
6 - 10 yrs
₹10L - ₹14L / yr
skill iconPython
skill iconDjango
FastAPI
skill iconFlask
pandas
+9 more

Position Overview: The Lead Software Architect - Python & Data Engineering is a senior technical leadership role responsible for designing and owning end-to-end architecture for data-intensive, AI/ML, and analytics platforms, while mentoring developers and ensuring technical excellence across the organization. 


Key Responsibilities: 

  • Design end-to-end software architecture for data-intensive applications, AI/ML pipelines, and analytics platforms
  • Evaluate trade-offs between competing technical approaches 
  • Define data models, API approach, and integration patterns across systems 
  • Create technical specifications and architecture documentation 
  • Lead by example through production-grade Python code and mentor developers on engineering fundamentals 
  • Conduct design and code reviews focused on architectural soundness 
  • Establish engineering standards, coding practices, and design patterns for the team 
  • Translate business requirements into technical architecture 
  • Collaborate with data scientists, analysts, and other teams to design integrated solutions 
  • Whiteboard and defend system design and architectural choices 
  • Take responsibility for system performance, reliability, and maintainability 
  • Identify and resolve architectural bottlenecks proactively 


Required Skills:  

  • 8+ years of experience in software architecture and development  
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field 
  • Strong foundations in data structures, algorithms, and computational complexity 
  • Experience in system design for scale, including caching strategies, load balancing, and asynchronous processing  
  • 6+ years of Python development experience 
  • Deep knowledge of Django, Flask, or FastAPI 
  • Expert understanding of Python internals including GIL and memory management 
  • Experience with RESTful API design and event-driven architectures (Kafka, RabbitMQ) 
  • Proficiency in data processing frameworks such as Pandas, Apache Spark, and Airflow 
  • Strong SQL optimization and database design experience (PostgreSQL, MySQL, MongoDB)  Experience with AWS, GCP, or Azure cloud platforms 
  • Knowledge of containerization (Docker) and orchestration (Kubernetes) 
  • Hands-on experience designing CI/CD pipelines Preferred (Bonus) 


Skills

  • Experience deploying ML models to production (MLOps, model serving, monitoring) Understanding of ML system design including feature stores and model versioning 
  • Familiarity with ML frameworks such as scikit-learn, TensorFlow, and PyTorch  
  • Open-source contributions or technical blogging demonstrating architectural depth 
  • Experience with modern front-end frameworks for full-stack perspective


Read more
ADTRAN

at ADTRAN

1 recruiter
Bisman Gill
Posted by Bisman Gill
Bengaluru (Bangalore), Hyderabad
2yrs+
Upto ₹24L / yr (Varies
)
skill iconScala
Apache Kafka
skill iconPostgreSQL
Monix
Cats Effect
+1 more

Key Responsibilities

  • Contribute to the design, development, and maintenance of scalable software systems.
  • Write clean, efficient, and testable code following best practices.
  • Participate in code reviews, technical discussions, and design sessions.
  • Collaborate with cross-functional teams to deliver high-quality features on time.
  • Troubleshoot, debug, and optimize application performance.
  • Continuously learn and apply new tools, frameworks, and technologies.

Required Skills & Qualifications

  • Strong foundation in computer science fundamentals: data structures, algorithms.
  • Hands-on experience in a Scalafunctional programming.
  • Experience in software design, implementation, and debugging, including multi-threading and inter-process communication.
  • Familiarity with Apache Kafka (messaging and persistence) and relational databases (Postgres preferred).
  • Basic knowledge of Python (for unit/integration testing) and Ansible.

Nice-to-Have / Bonus Skills

  • Familiarity with modeling in YANG.
  • Exposure to Scala libraries such as Cats Effect, Monix, or Akka.
  • Experience with Agile/Scrum methodologies.
Read more
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
skill iconGo Programming (Golang)
CI/CD
Apache Kafka
RabbitMQ
skill iconDocker
+1 more

Location: Bengalore, India, Exp: 3-5 Yrs

Backend Developer (Golang) - Trading & Fintech


About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What we expect:

 You should already be exceptional at Golang. If you need hand-holding, this isn’t the place for you.

 You thrive on challenges, not on perks or financial rewards.

 You measure success by your own growth, not external validation.

 Taking calculated risks excites you—you’re here to build, break, and learn.

 You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading environment.

 You understand the stakes—milliseconds can make or break trades, and precision is everything.



What you will do:

 Develop and optimize high-performance backend systems in Golang for trading platforms and financial services.

 Architect low-latency, high-throughput microservices that push the boundaries ofspeed and efficiency.

 Build event-driven, fault-tolerant systems that can handle massive real-time data streams.

 Own your work—no babysitting, no micromanagement.

 Work alongside equally driven engineers who expect nothing less than brilliance.


Must have skills:

 Learn faster than you ever thought possible.

Proven expertise in Golang (if you need to prove yourself, this isn’t the role for you).

 Deep understanding of concurrency, memory management, and system design.

 Experience with Trading, market data processing, or low-latency systems.

 Strong knowledge of distributed systems, message queues (Kafka, RabbitMQ), and real-time processing.

 Hands-on with Docker, Kubernetes, and CI/CD pipelines.

 A portfolio of work that speaks louder than a resume.


Nice-to-Have Skills:

 Past experience in fintech, trading systems, or algorithmic trading.

 Contributions to open-source Golang projects.

 A history of building something impactful from scratch.

 Understanding of FIX protocol, WebSockets, and streaming APIs.


Why Join Us?

 Work with a team that expects and delivers excellence.

 A culture where risk-taking is rewarded, and complacency is not.

 Limitless opportunities for growth—if you can handle the pace.

 A place where learning is currency, and outperformance is the only metric that matters.


The opportunity to build systems that move markets, execute trades in microseconds, and redefine fintech. This isn’t just a job—it’s a proving ground. Ready to take the leap? Apply now.



Read more
HighLevel Inc.
Reshika Mendiratta
Posted by Reshika Mendiratta
Remote only
6yrs+
Upto ₹50L / yr (Varies
)
skill iconNodeJS (Node.js)
skill iconReact.js
skill iconMongoDB
skill iconExpress
MERN Stack
+2 more

About Us

HighLevel is an AI powered, all-in-one white-label sales & marketing platform that empowers agencies, entrepreneurs, and businesses to elevate their digital presence and drive growth. We are proud to support a global and growing community of over 2 million businesses, comprised of agencies, consultants, and businesses of all sizes and industries. HighLevel empowers users with all the tools needed to capture, nurture, and close new leads into repeat customers. As of mid 2025, HighLevel processes over 15 billion API hits and handles more than 2.5 billion message events every day. Our platform manages over 470 terabytes of data distributed across five databases, operates with a network of over 250 microservices, and supports over 1 million domain names.


About the Role

  • We’re looking for a Lead Software Engineer with deep expertise in Node.js, MongoDB, and modern frontend frameworks, along with a bold vision for how AI-assisted development can accelerate engineering workflows.
  • You’ll join our Snapshots Team, the engine behind how agencies and SaaS creators templatise and clone complete automation systems—workflows, funnels, forms, calendars, settings, and more—instantly across thousands of client accounts.
  • Snapshots are now being cloned to over 200,000+ businesses every month, driving rapid onboarding and platform extensibility. The system saves businesses countless hours by replicating complex structures and business processes seamlessly across accounts, making automation scalable, repeatable, and fast.
  • Under the hood, Snapshots tackles engineering challenges like asset graph traversal, dependency resolution, real-time syncing across multi-tenant environments, and copying gigabytes of data across accounts—all while managing CPU and memory intensive operations at scale.
  • As a Lead Engineer, you’ll own architectural decisions and solve deep product-platform problems around high-performance snapshot cloning, rollback systems, and version management. You’ll collaborate across product, infra, and platform teams while mentoring engineers and driving the long-term roadmap for one of HighLevel’s most strategic surface areas. If you enjoy building technically rigorous systems that scale, this team is where your work will have immediate and visible impact.

Responsibilities

  • Lead the design and development of scalable, high-performance systems to improve reliability, latency, and throughput across time-sensitive APIs and workflows
  • Own features end-to-end — from architecture and implementation to testing, deployment, and ongoing optimization.
  • Work hands-on with technologies like NestJS, Firestore, MongoDB, PostgreSQL, Redis, Queuing Systems and Service Mesh-based microservices.
  • Drive technical direction across both product features and platform layers to ensure stability, scalability, and maintainability.Integrate and optimize AI-assisted development tools, improving developer productivity while ensuring accuracy and minimizing hallucinations.
  • Collaborate closely with Product, Design, and AI teams to deliver impactful, user-facing features and backend systems.Mentor and support other engineers, fostering a culture of technical excellence, learning, and ownership.Proactively identify and solve performance bottlenecks, scalability challenges, and security concerns in a multi-tenant environment.

Requirements

  • 6+ years of backend engineering experience, including designing fault-tolerant systems and working on high-scale platforms.Deep expertise in distributed systems, event-driven architectures, and asynchronous job processing
  • Strong experience with relational and NoSQL data models (especially with complex temporal data)
  • Experience with modern front-end frameworks (e.g., React, Vue, Angular) and building full-stack web applications.Proven track record of architecting complex systems and delivering scalable, high-performance web apps in production.
  • Strong understanding of software design patterns, API design, and microservices architecture in a multi-tenant environment.
  • Skilled in guiding technical architecture, making high-impact engineering decisions, and mentoring fellow engineers.Experience with code quality practices, automated testing, CI/CD pipelines, and dev tooling optimization.Excellent problem-solving skills, with clear and collaborative communication across cross-functional teams.

Our People

With over 1,500 team members across 15+ countries, we operate in a global, remote-first environment. We are building more than software; we are building a global community rooted in creativity, collaboration, and impact. We take pride in cultivating a culture where innovation thrives, ideas are celebrated, and people come first, no matter where they call home.


Our Impact

As of mid 2025, our platform powers over 1.5 billion messages, helps generate over 200 million leads, and facilitates over 20 million conversations for the more than 2 million businesses we serve each month. Behind those numbers are real people growing their companies, connecting with customers, and making their mark - and we get to help make that happen.

Read more
prep study
Pooja Sharma
Posted by Pooja Sharma
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
FastAPI
RESTful APIs
RabbitMQ
Apache Kafka
skill iconElastic Search
+2 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.

What We Expect:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field.

Read more
Bengaluru (Bangalore)
4 - 7 yrs
₹15L - ₹35L / yr
skill iconJava
skill iconSpring Boot
Apache Kafka

SE/SSE - Backend (Java)



JAVA: Software Engineer / Senior Software Engineer


Job Overview: RevSure.AI is building a world-class team of Engineers with a mandate to architect, design, build, scale, and maintain our cutting-edge Revops platform and derive insights from the customer data. If you are a great JavaScript developer (or, for that matter, in any major programming language) with experience in building SaaS web applications and looking for an opportunity to build world-class products using cutting-edge technologies, please read on -


Responsibilities:

● Contribute to the growth and development of our product by writing and maintaining code, focusing on quality and maintainability. We believe that our teams own the code they build.

● Share your knowledge, both inside and outside your own team.

● Design, implement, test, and monitor valuable solutions that achieve the teamʼs goals while keeping a smooth delivery flow and balancing tradeoffs between scope, time, and effort.

● Continuously improve your and your teamʼs way of working, expanding the boundaries of your teamʼs autonomy and self-organisation.


Requirements:


Must-Have:

● Good interpersonal and communication skills. An analytical mind and an eye for detail

● Experience of 4-7 years in building large applications, preferably from the heavy data processing & analytics side

● Experience in architecting, designing, and building end-to-end System Design of multi tenant SAAS applications

● Experience architecting interactive configuration experiences around DAGs, configuration-heavy reporting feature, interactive data pipeline design interfaces, and charting.

● Experience in Java, Spring Boot, and GCP

● In-depth understanding of design patterns, OOPs, and Functional programming.

● Passionate programmer focused on Backend Programming.


Good to have:

● Working knowledge of GCP managed services like Kubernetes, PubSub


Please feel free to share your resume to chandana at revsure.ai


**Based on your experience, the title could vary. We care more about mindset , ownership, and craft than just titles.

Read more
Whiz IT Services
Sheeba Harish
Posted by Sheeba Harish
Remote only
10 - 15 yrs
₹20L - ₹20L / yr
skill iconJava
skill iconSpring Boot
Microservices
API
Apache Kafka
+5 more

We are looking for highly experienced Senior Java Developers who can architect, design, and deliver high-performance enterprise applications using Spring Boot and Microservices . The role requires a strong understanding of distributed systems, scalability, and data consistency.

Read more
Sonatype

at Sonatype

5 candid answers
Reshika Mendiratta
Posted by Reshika Mendiratta
Hyderabad
5 - 8 yrs
Upto ₹28L / yr (Varies
)
ETL
skill iconPython
databricks
SQL
Data modeling
+6 more

Who We Are

At Sonatype, we help organizations build better, more secure software by enabling them to understand and control their software supply chains. Our products are trusted by thousands of engineering teams globally, providing critical insights into dependency health, license risk, and software security. We’re passionate about empowering developers—and we back it with data.


The Opportunity

We’re looking for a Data Engineer with full stack expertise to join our growing Data Platform team. This role blends data engineering, microservices, and full-stack development to deliver end-to-end services that power analytics, machine learning, and advanced search across Sonatype.

You will design and build data-driven microservices and workflows using Java, Python, and Spring Batch, implement frontends for data workflows, and deploy everything through CI/CD pipelines into AWS ECS/Fargate. You’ll also ensure services are monitorable, debuggable, and reliable at scale, while clearly documenting designs with Mermaid-based sequence and dataflow diagrams.

This is a hands-on engineering role for someone who thrives at the intersection of data systems, fullstack development, ML, and cloud-native platforms.


What You’ll Do

  • Design, build, and maintain data pipelines, ETL/ELT workflows, and scalable microservices.
  • Development of complex web scraping (Playwright) and realtime pipelines (Kafka/Queues/Flink).
  • Develop end-to-end microservices with backend (Java 5+, Python 5+, Spring Batch 2+) and frontend (React or any).
  • Deploy, publish, and operate services in AWS ECS/Fargate using CI/CD pipelines (Jenkins, GitOps).
  • Architect and optimize data storage models in SQL (MySQL, PostgreSQL) and NoSQL stores.
  • Implement web scraping and external data ingestion pipelines.
  • Enable Databricks and PySpark-based workflows for large-scale analytics.
  • Build advanced data search capabilities (fuzzy matching, vector similarity search, semantic retrieval).
  • Apply ML techniques (scikit-learn, classification algorithms, predictive modeling) to data-driven solutions.
  • Implement observability, debugging, monitoring, and alerting for deployed services.
  • Create Mermaid sequence diagrams, flowcharts, and dataflow diagrams to document system architecture and workflows.
  • Drive best practices in fullstack data service development, including architecture, testing, and documentation.


What We’re Looking For


Minimum Qualifications

  • 2+ years of experience as a Data Engineer or a Software Backend engineering role
  • Strong programming skills in Python, Scala, or Java
  • Hands-on experience with HBase or similar NoSQL columnar stores
  • Hands-on experience with distributed data systems like Spark, Kafka, or Flink
  • Proficient in writing complex SQL and optimizing queries for performance
  • Experience building and maintaining robust ETL/ELT pipelines in production
  • Familiarity with workflow orchestration tools (Airflow, Dagster, or similar)
  • Understanding of data modeling techniques (star schema, dimensional modeling, etc.)
  • Familiarity with CI/CD pipelines (Jenkins or similar)
  • Ability to visualize and communicate architectures using Mermaid diagrams

Bonus Points

  • Experience working with Databricks, dbt, Terraform, or Kubernetes
  • Familiarity with streaming data pipelines or real-time processing
  • Exposure to data governance frameworks and tools
  • Experience supporting data products or ML pipelines in production
  • Strong understanding of data privacy, security, and compliance best practices


Why You’ll Love Working Here

  • Data with purpose: Work on problems that directly impact how the world builds secure software
  • Modern tooling: Leverage the best of open-source and cloud-native technologies
  • Collaborative culture: Join a passionate team that values learning, autonomy, and impact
Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Pune
6 - 12 yrs
₹25L - ₹30L / yr
skill iconMachine Learning (ML)
AWS CloudFormation
Online machine learning
skill iconAmazon Web Services (AWS)
ECS
+20 more

MUST-HAVES: 

  • Machine Learning + Aws + (EKS OR ECS OR Kubernetes) + (Redshift AND Glue) + Sage maker
  • Notice period - 0 to 15 days only 
  • Hybrid work mode- 3 days office, 2 days at home


SKILLS: AWS, AWS CLOUD, AMAZON REDSHIFT, EKS


ADDITIONAL GUIDELINES:

  • Interview process: - 2 Technical round + 1 Client round
  • 3 days in office, Hybrid model. 


CORE RESPONSIBILITIES:

  • The MLE will design, build, test, and deploy scalable machine learning systems, optimizing model accuracy and efficiency
  • Model Development: Algorithms and architectures span traditional statistical methods to deep learning along with employing LLMs in modern frameworks.
  • Data Preparation: Prepare, cleanse, and transform data for model training and evaluation.
  • Algorithm Implementation: Implement and optimize machine learning algorithms and statistical models.
  • System Integration: Integrate models into existing systems and workflows.
  • Model Deployment: Deploy models to production environments and monitor performance.
  • Collaboration: Work closely with data scientists, software engineers, and other stakeholders.
  • Continuous Improvement: Identify areas for improvement in model performance and systems.


SKILLS:

  • Programming and Software Engineering: Knowledge of software engineering best practices (version control, testing, CI/CD).
  • Data Engineering: Ability to handle data pipelines, data cleaning, and feature engineering. Proficiency in SQL for data manipulation + Kafka, Chaos search logs, etc. for troubleshooting; Other tech touch points are Scylla DB (like BigTable), OpenSearch, Neo4J graph
  • Model Deployment and Monitoring: MLOps Experience in deploying ML models to production environments.
  • Knowledge of model monitoring and performance evaluation.


REQUIRED EXPERIENCE:

  • Amazon SageMaker: Deep understanding of SageMaker's capabilities for building, training, and deploying ML models; understanding of the Sage maker pipeline with ability to analyze gaps and recommend/implement improvements
  • AWS Cloud Infrastructure: Familiarity with S3, EC2, Lambda and using these services in ML workflows
  • AWS data: Redshift, Glue
  • Containerization and Orchestration: Understanding of Docker and Kubernetes, and their implementation within AWS (EKS, ECS)
Read more
Tradelab Software Private Limited
Pooja Sharma
Posted by Pooja Sharma
Bengaluru (Bangalore)
5 - 8 yrs
₹10L - ₹15L / yr
concurrency
Apache Kafka
RabbitMQ
skill iconDocker
skill iconKubernetes

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What We Expect:

• You should already be exceptional at Backend. If you need hand-holding, this isn’t the place for you.

• You thrive on challenges, not on perks or financial rewards.

• You measure success by your own growth, not external validation.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a pay check; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What You Will Do:

• Develop and optimize high-performance backend systems in for trading platforms and financial services.

• Architect low-latency, high-throughput microservices that push the boundaries of speed and efficiency.

• Build event-driven, fault-tolerant systems that can handle massive real-time data streams.

• Own your work—no babysitting, no micromanagement.

• Work alongside equally driven engineers who expect nothing less than brilliance.

• Learn faster than you ever thought possible.


Must-Have Skills:

Proven expertise in Backend (if you need to prove yourself, this isn’t the role for you).

• Deep understanding of concurrency, memory management, and system design.

• Experience with Trading, market data processing, or low-latency systems.

• Strong knowledge of distributed systems, message queues (Kafka, RabbitMQ), and real-time processing.

• Hands-on with Docker, Kubernetes, and CI/CD pipelines.

• A portfolio of work that speaks louder than a resume.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Contributions to open-source projects.

• A history of building something impactful from scratch.

• Understanding of FIX protocol, WebSockets, and streaming APIs.


Why Join Us?

• Work with a team that expects and delivers excellence.

• A culture where risk-taking is rewarded, and complacency is not.

• Limitless opportunities for growth—if you can handle the pace.

• A place where learning is currency, and outperformance is the only metric that matters.

• The opportunity to build systems that move markets, execute trades in microseconds, and redefine fintech.

Read more
Tradelab Software Private Limited
Pooja Sharma
Posted by Pooja Sharma
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
RESTful APIs
RabbitMQ
Apache Kafka
skill iconRedis
skill iconElastic Search
+2 more

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits every day. If you thrive in high-stakes environments and have a deep passion for performance. driven backend systems, we want you.


What We Expect:

• We’re looking for a Backend Developer (Python) with a strong foundation in backend technologies and

a deep interest in scalable, low-latency systems.

• You should have 3–4 years of experience in Python-based development and be eager to solve complex

performance and scalability challenges in trading and fintech applications.

• You measure success by your own growth, not external validation.

• You thrive on challenges, not on perks or financial rewards.

• Taking calculated risks excites you—you’re here to build, break, and learn.

• You don’t clock in for a paycheck; you clock in to outperform yourself in a high-frequency trading

environment.

• You understand the stakes—milliseconds can make or break trades, and precision is everything.


What You Will Do:

• Develop and maintain scalable backend systems using Python.

• Design and implement REST APIs and socket-based communication.

• Optimize code for speed, performance, and reliability.

• Collaborate with frontend teams to integrate server-side logic.

• Work with RabbitMQ, Kafka, Redis, and Elasticsearch for robust backend design.

• Build fault-tolerant, multi-producer/consumer systems.


Must-Have Skills:

• 3–4 years of experience in Python and backend development.

• Strong understanding of REST APIs, sockets, and network protocols (TCP/UDP/HTTP).

• Experience with RabbitMQ/Kafka, SQL & NoSQL databases, Redis, and Elasticsearch.

• Bachelor’s degree in Computer Science or related field

Read more
ADTRAN

at ADTRAN

1 recruiter
Reema Meshram
Posted by Reema Meshram
Bengaluru (Bangalore)
3 - 7 yrs
₹12L - ₹35L / yr
skill iconScala
Functional programming
Akka
Apache Kafka
skill iconPostgreSQL

Key Responsibilities

  • Provide technical leadership in the design, development, and delivery of scalable, high-performance software systems.
  • Partner with product managers, architects, and cross-functional teams to define technical strategy and ensure alignment with business objectives.
  • Lead by example in writing high-quality, testable, and maintainable code.
  • Drive best practices in software engineering, including code reviews, system design, and performance optimization.
  • Mentor and guide engineers across teams, fostering a culture of technical excellence and continuous learning.
  • Evaluate and introduce new technologies, tools, and frameworks to improve productivity, scale and system robustness.

Required Skills & Qualifications

  • Strong foundation in computer science fundamentals: data structures, algorithms, and functional programming techniques.
  • Expertise in Scala, with strong preference for functional programming.
  • Solid experience in software design, implementation, and debugging, including inter-process communication and multi-threading.
  • Hands-on experience with distributed systems and event-driven architectures.
  • Familiarity with databases (Postgres preferred).
  • Proficiency with Apache Kafka for messaging and persistence.
  • Working knowledge of Python for unit and integration testing.
  • Basic to intermediate experience with Ansible for automation.
  • Strong problem-solving, analytical, and communication skills.

Nice-to-Have / Bonus Skills

  • Experience with modeling in YANG.
  • Experience with Scala libraries such as Cats Effect (2/3), Monix, and Akka.
  • Experience working in Agile/Scrum environments.

What We Offer

  • Opportunity to work on cutting-edge technologies in a collaborative environment.
  • A role with strong ownership, technical influence, and visibility across teams.
  • Competitive compensation and benefits.


Read more
Zeliot

at Zeliot

4 candid answers
2 recruiters
Bisman Gill
Posted by Bisman Gill
Remote, Bengaluru (Bangalore)
5yrs+
Upto ₹25L / yr (Varies
)
skill iconNodeJS (Node.js)
skill iconReact.js
Systems design
Apache Kafka
skill iconKubernetes

We are seeking a highly skilled Full Stack Developer with strong proficiencyi n Node.js, React.js, and modern database design principles. The ideal candidate will be responsible for designing, developing, testing, and deploying scalable web applications that deliver exceptional user experiences and robust backend performance.


Key Responsibilities


  • Participate in end-to-end product development — from

conceptualization to deployment and support.

  • Design and develop RESTful and GraphQL APIs using Node.js and frameworks such as Express or NestJS.
  • Build and maintain responsive front-end interfaces using React.js, HTML5, CSS3, and modern libraries like Redux or React Query.
  • Architect, create, and maintain database schemas, relationships, and efficient data models in PostgreSQL or MySQL.
  • Optimize queries, indexes, and data access patterns for performance and scalability.
  • Implement and maintain observability and monitoring using tools such as Prometheus and Grafana.
  • Troubleshoot and debug API issues, backend bottlenecks, and production incidents effectively.
  • Write automated unit, integration, and end-to-end tests to ensure high code quality.
  • Collaborate with cross-functional teams (Product, QA, DevOps, Design, Data) to ship high-impact features.
  • Deploy and manage applications on AWS, Azure, or GCP using CI/CD pipelines and container technologies such as Docker and Kubernetes.
  • Contribute to code reviews, maintain best practices in software design, and ensure adherence to coding standards.
  • Continuously explore and learn emerging technologies to improve efficiency and scalability.
  • Demonstrate problem-solving skills through debugging, performance tuning, and design optimization.


Technical Skills & Requirements

  • 5+ years of hands-on experience in full stack development.
  • Strong proficiency in:
  • Backend: Node.js (Express.js, NestJS)
  • Frontend: React.js, Next.js, Redux/Context API
  • Database: PostgreSQL / MySQL (schema design, indexing, joins, and data type optimization)
  • API: REST and GraphQL design, implementation, and security.
  • Proficient in JavaScript (ES6+), TypeScript, and asynchronous programming.
  • Experience with Git, Yarn/NPM, and Visual Studio Code or equivalent tools.
  • Familiarity with observability, logging, and tracing concepts — using tools such as Grafana & Prometheus.
  • Knowledge of cloud services (AWS, GCP, Azure) including compute, storage, and managed databases.
  • Understanding of CI/CD pipelines, Docker, and Kubernetes for modern deployments.
  • Strong grasp of software design patterns, system architecture, and scalability principles.
  • Excellent problem-solving, debugging, and analytical abilities.


Preferred Skills

  • Experience with microservices architecture or event-driven systems.
  • Exposure to Kafka, Redis, or message queues.
  • Familiarity with automated testing frameworks (Jest, Cypress, Mocha).
  • Understanding of security best practices (JWT, OAuth, HTTPS, CORS).

Education & Experience

  • Bachelor’s or Master’s degree in Computer Science, Information Systems, or a related technical field.
  • Typically 5–7 years of relevant experience in software development.
  • Demonstrated experience working in an Agile or Scrum environment.

Soft Skills

  • Strong communication and collaboration skills.
  • Self-motivated, quick learner, and capable of working autonomously.
  • Passionate about writing clean, maintainable, and efficient code.
  • Curious about new technologies and approaches to problem-solving.



Read more
Lookup

at Lookup

2 candid answers
1 recruiter
Ajay Kumar
Posted by Ajay Kumar
Bengaluru (Bangalore)
3 - 8 yrs
₹30L - ₹35L / yr
skill iconPython
skill iconPostgreSQL
FastAPI
skill iconRedis
skill iconAmazon Web Services (AWS)
+2 more

Our Mission

To make video as accessible to machines as text and voice are today.


At lookup, we believe the world's most valuable asset is trapped. Video is everywhere, but it's unsearchable—a black box of insight that no one can open or atleast open affordably. We’re changing that. We're building the search engine for the visual world, so anyone can find or do anything with video just by asking.


Text is queryable. Voice is transcribed. Video, the largest and richest data source of all, is still a black box. A computer can't understand it, and so its value remains trapped.


Our mission at lookup is to fix this.


About the Role

We are looking for founding Backend Engineers to build a highly performant, reliable, and scalable API platform that makes enterprise video knowledge readily available for video search, summarization, and natural‑language Q&A. You will partner closely with our ML team working on vision‑language models to productionize research and deliver fast, trustworthy APIs for customers.


Examples of technical challenges you will work on include: distributed video storage, a unified application framework and data model for indexing large video libraries, low‑latency clip retrieval, vector search at scale, and end‑to‑end build, test, deploy, and observability in cloud environments.


What You’ll Do:

  • Design and build robust backend services and APIs (REST, gRPC) for vector search, video summarization, and video Q&A.
  • Own API performance and reliability, including low‑latency retrieval, pagination, rate limiting, and backwards‑compatible versioning.
  • Design schemas and tune queries in Postgres, and integrate with unstructured storage.
  • Implement observability across metrics, logs, and traces. Set error budgets and SLOs.
  • Write clear design docs and ship high‑quality, well‑tested code.
  • Collaborate with ML engineers to integrate and productionize VLMs and retrieval pipelines.
  • Take ownership of architecture from inception to production launch.


Who You Are:

  • 3+ years of professional experience in backend development.
  • Proven experience building and scaling polished WebSocket, gRPC, and REST APIs.
  • Exposure to distributed systems and container orchestration (Docker and Kubernetes).
  • Hands‑on experience with AWS.
  • Strong knowledge of SQL (Postgres) and NoSQL (e.g., Cassandra), including schema design, query optimization, and scaling.
  • Familiarity with our stack is a plus, but not mandatory: Python (FastAPI), Celery, Kafka, Postgres, Redis, Weaviate, React.
  • Ability to diagnose complex issues, identify root causes, and implement effective fixes.
  • Comfortable working in a fast‑paced startup environment.


Nice to have:

  • Hands-on work with LLM agents, vector embeddings, or RAG applications.
  • Building video streaming pipelines and storage systems at scale (FFmpeg, RTSP, WebRTC).
  • Proficiency with modern frontend frameworks (React, TypeScript, Tailwind CSS) and responsive UI design.


Location & Culture

  • Full-time, in-office role in Bangalore (we’re building fast and hands-on).
  • Must be comfortable with a high-paced environment and collaboration across PST time zones for our US customers and investors.
  • Expect startup speed — daily founder syncs, rapid design-to-prototype cycles, and a culture of deep ownership.


Why You’ll Love This Role

  • Work on the frontier of video understanding and real-world AI — products that can redefine trust and automation.
  • Build core APIs that make video queryable and power real customer use.
  • Own systems end to end: performance, reliability, and developer experience.
  • Work closely with founders and collaborate in person in Bangalore.
  • Competitive salary with meaningful early equity.
Read more
AryuPay Technologies
Bengaluru (Bangalore), Bhopal
4 - 8 yrs
₹5L - ₹10L / yr
skill iconDjango
RESTful APIs
skill iconFlask
skill iconPostgreSQL
CI/CD
+7 more

Senior Python Django Developer 

Experience: Back-end development: 6 years (Required)


Location:  Bangalore/ Bhopal

Job Description:

We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the startup environment.

This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.

Responsibilities:

  • Design and develop scalable, secure, and high-performance applications using Python (Django framework).
  • Architect system components, define database schemas, and optimize backend services for speed and efficiency.
  • Lead and implement design patterns and software architecture best practices.
  • Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
  • Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
  • Drive performance improvements, monitor system health, and troubleshoot production issues.
  • Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
  • Contribute to technical decision-making and mentor junior developers.

Requirements:

  • 6 to 10 years of professional backend development experience with Python and Django.
  • Strong background in payments/financial systems or FinTech applications.
  • Proven experience in designing software architecture in a microservices or modular monolith environment.
  • Experience working in fast-paced startup environments with agile practices.
  • Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
  • Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
  • Hands-on experience with test-driven development (TDD) and frameworks like pytest, unittest, or factory_boy.
  • Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).

Preferred Skills:

  • Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
  • Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
  • Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
  • Contributions to open-source or personal finance-related projects.

Job Types: Full-time, Permanent


Schedule:

  • Day shift

Supplemental Pay:

  • Performance bonus
  • Yearly bonus

Ability to commute/relocate:

  • JP Nagar, 5th Phase, Bangalore, Karnataka or Indrapuri, Bhopal, Madhya Pradesh: Reliably commute or willing to relocate with an employer-provided relocation package (Preferred)

Read less


Read more
Beyond Seek Technologies Pvt Ltd
Remote only
4 - 7 yrs
₹14L - ₹25L / yr
Process automation
RPA
skill iconPython
AWS Lambda
AWS Simple Queuing Service (SQS)
+6 more

Responsibilities:

  • Develop and maintain RPA workflows using Selenium, AWS Lambda, and message queues (SQS/RabbitMQ/Kafka).
  • Build and evolve internal automation frameworks, reusable libraries, and CI-integrated test suites to accelerate developer productivity.
  • Develop comprehensive test strategies (unit, integration, end-to-end), optimize performance, handle exceptions, and ensure high reliability of automation scripts.
  • Monitor automation health and maintain dashboards/logging via cloud tools (CloudWatch, ELK, etc. ).
  • Champion automation standards workshops, write documentation, and coach other engineers on test-driven development and behavior-driven automation.


Requirements:

  • 4-5 years of experience in automation engineering with deep/hands-on experience in Python and modern browser automation frameworks (Selenium/PythonRPA).
  • Solid background with desktop-automation solutions (UiPath, PythonRPA) and scripting legacy applications.
  • Strong debugging skills, with an eye for edge cases and race conditions in distributed, asynchronous systems.
  • Hands-on experience with AWS services like Lambda, S3 and API Gateway.
  • Familiarity with REST APIs, webhooks, and queue-based async processing.
  • Experience integrating with third-party platforms or enterprise systems.
  • Ability to translate business workflows into technical automation logic.
  • Able to evangelize automation best practices, present complex ideas clearly, and drive cross-team alignment.


Nice to Have:

  • Experience with RPA frameworks (UiPath, BluePrism, etc. ).
  • Familiarity with building LLM-based workflows (LangChain, LlamaIndex) or custom agent loops to automate cognitive tasks.
  • Exposure to automotive dealer management or VMS platforms.
  • Understanding of cloud security and IAM practices.
Read more
Cspar Enterprises Private Limited
Bhopal, Bengaluru (Bangalore)
4 - 10 yrs
₹3L - ₹8L / yr
skill iconDjango
RESTful APIs
deployment tools
RabbitMQ
Apache Kafka
+11 more

Designation: Senior Python Django Developer 

Position: Senior Python Developer

Job Types: Full-time, Permanent

Pay: Up to ₹800,000.00 per year

Schedule: Day shift

Ability to commute/relocate: Bhopal Indrapuri (MP) And Bangalore JP Nagar

 

Experience: Back-end development: 4 years (Required)

 

Job Description:

We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the startup environment.

This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.

 

Responsibilities:

  • Design and develop scalable, secure, and high-performance applications using Python (Django framework).
  • Architect system components, define database schemas, and optimize backend services for speed and efficiency.
  • Lead and implement design patterns and software architecture best practices.
  • Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
  • Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
  • Drive performance improvements, monitor system health, and troubleshoot production issues.
  • Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
  • Contribute to technical decision-making and mentor junior developers.

 

Requirements:

  • 4 to 10 years of professional backend development experience with Python and Django.
  • Strong background in payments/financial systems or FinTech applications.
  • Proven experience in designing software architecture in a microservices or modular monolith environment.
  • Experience working in fast-paced startup environments with agile practices.
  • Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
  • Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
  • Hands-on experience with test-driven development (TDD) and frameworks like pytest, unit test, or factory boy.
  • Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).

 

Preferred Skills:

  • Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
  • Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
  • Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
  • Contributions to open-source or personal finance-related projects.


Read more
Bengaluru (Bangalore)
2 - 5 yrs
₹10L - ₹15L / yr
Apache Kafka
RabbitMQ
skill iconDocker
skill iconKubernetes
CI/CD

About Us:

Tradelab Technologies Pvt Ltd is not for those seeking comfort—we are for those hungry to make a mark in the trading and fintech industry. If you are looking for just another backend role, this isn’t it. We want risk-takers, relentless learners, and those who find joy in pushing their limits

every day. If you thrive in high-stakes environments and have a deep passion for performance driven backend systems, we want you.


What You Will Do:

• Develop and optimize high-performance backend systems in Golang for trading platforms and financial

services.

• Architect low-latency, high-throughput microservices that push the boundaries of speed and efficiency.

• Build event-driven, fault-tolerant systems that can handle massive real-time data streams.

• Own your work—no babysitting, no micromanagement.

• Work alongside equally driven engineers who expect nothing less than brilliance.

• Learn faster than you ever thought possible.


Must-Have Skills:

Proven expertise in Golang (if you need to prove yourself, this isn’t the role for you).

• Deep understanding of concurrency, memory management, and system design.

• Experience with Trading, market data processing, or low-latency systems.

• Strong knowledge of distributed systems, message queues (Kafka, RabbitMQ), and real-time processing.

• Hands-on with Docker, Kubernetes, and CI/CD pipelines.

• A portfolio of work that speaks louder than a resume.


Nice-to-Have Skills:

• Past experience in fintech, trading systems, or algorithmic trading.

• Contributions to open-source Golang projects.

• A history of building something impactful from scratch.

• Understanding of FIX protocol, WebSockets, and streaming APIs.

Read more
Inncircles
Gangadhar M
Posted by Gangadhar M
Hyderabad
3 - 5 yrs
Best in industry
PySpark
Spark
skill iconPython
ETL
Amazon EMR
+7 more


We are looking for a highly skilled Sr. Big Data Engineer with 3-5 years of experience in

building large-scale data pipelines, real-time streaming solutions, and batch/stream

processing systems. The ideal candidate should be proficient in Spark, Kafka, Python, and

AWS Big Data services, with hands-on experience in implementing CDC (Change Data

Capture) pipelines and integrating multiple data sources and sinks.


Responsibilities

  • Design, develop, and optimize batch and streaming data pipelines using Apache Spark and Python.
  • Build and maintain real-time data ingestion pipelines leveraging Kafka and AWS Kinesis.
  • Implement CDC (Change Data Capture) pipelines using Kafka Connect, Debezium or similar frameworks.
  • Integrate data from multiple sources and sinks (databases, APIs, message queues, file systems, cloud storage).
  • Work with AWS Big Data ecosystem: Glue, EMR, Kinesis, Athena, S3, Lambda, Step Functions.
  • Ensure pipeline scalability, reliability, and performance tuning of Spark jobs and EMR clusters.
  • Develop data transformation and ETL workflows in AWS Glue and manage schema evolution.
  • Collaborate with data scientists, analysts, and product teams to deliver reliable and high-quality data solutions.
  • Implement monitoring, logging, and alerting for critical data pipelines.
  • Follow best practices for data security, compliance, and cost optimization in cloud environments.


Required Skills & Experience

  • Programming: Strong proficiency in Python (PySpark, data frameworks, automation).
  • Big Data Processing: Hands-on experience with Apache Spark (batch & streaming).
  • Messaging & Streaming: Proficient in Kafka (brokers, topics, partitions, consumer groups) and AWS Kinesis.
  • CDC Pipelines: Experience with Debezium / Kafka Connect / custom CDC frameworks.
  • AWS Services: AWS Glue, EMR, S3, Athena, Lambda, IAM, CloudWatch.
  • ETL/ELT Workflows: Strong knowledge of data ingestion, transformation, partitioning, schema management.
  • Databases: Experience with relational databases (MySQL, Postgres, Oracle) and NoSQL (MongoDB, DynamoDB, Cassandra).
  • Data Formats: JSON, Parquet, Avro, ORC, Delta/Iceberg/Hudi.
  • Version Control & CI/CD: Git, GitHub/GitLab, Jenkins, or CodePipeline.
  • Monitoring/Logging: CloudWatch, Prometheus, ELK/Opensearch.
  • Containers & Orchestration (nice-to-have): Docker, Kubernetes, Airflow/Step
  • Functions for workflow orchestration.


Preferred Qualifications

  • Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field.
  • Experience in large-scale data lake / lake house architectures.
  • Knowledge of data warehousing concepts and query optimisation.
  • Familiarity with data governance, lineage, and cataloging tools (Glue Data Catalog, Apache Atlas).
  • Exposure to ML/AI data pipelines is a plus.


Tools & Technologies (must-have exposure)

  • Big Data & Processing: Apache Spark, PySpark, AWS EMR, AWS Glue
  • Streaming & Messaging: Apache Kafka, Kafka Connect, Debezium, AWS Kinesis
  • Cloud & Storage: AWS (S3, Athena, Lambda, IAM, CloudWatch)
  • Programming & Scripting: Python, SQL, Bash
  • Orchestration: Airflow / Step Functions
  • Version Control & CI/CD: Git, Jenkins/CodePipeline
  • Data Formats: Parquet, Avro, ORC, JSON, Delta, Iceberg, Hudi
Read more
Technology Industry

Technology Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Delhi
10 - 15 yrs
₹105L - ₹140L / yr
Data engineering
Apache Spark
Apache
Apache Kafka
skill iconJava
+25 more

MANDATORY:

  • Super Quality Data Architect, Data Engineering Manager / Director Profile
  • Must have 12+ YOE in Data Engineering roles, with at least 2+ years in a Leadership role
  • Must have 7+ YOE in hands-on Tech development with Java (Highly preferred) or Python, Node.JS, GoLang
  • Must have strong experience in large data technologies, tools like HDFS, YARN, Map-Reduce, Hive, Kafka, Spark, Airflow, Presto etc.
  • Strong expertise in HLD and LLD, to design scalable, maintainable data architectures.
  • Must have managed a team of at least 5+ Data Engineers (Read Leadership role in CV)
  • Product Companies (Prefers high-scale, data-heavy companies)


PREFERRED:

  • Must be from Tier - 1 Colleges, preferred IIT
  • Candidates must have spent a minimum 3 yrs in each company.
  • Must have recent 4+ YOE with high-growth Product startups, and should have implemented Data Engineering systems from an early stage in the Company


ROLES & RESPONSIBILITIES:

  • Lead and mentor a team of data engineers, ensuring high performance and career growth.
  • Architect and optimize scalable data infrastructure, ensuring high availability and reliability.
  • Drive the development and implementation of data governance frameworks and best practices.
  • Work closely with cross-functional teams to define and execute a data roadmap.
  • Optimize data processing workflows for performance and cost efficiency.
  • Ensure data security, compliance, and quality across all data platforms.
  • Foster a culture of innovation and technical excellence within the data team.


IDEAL CANDIDATE:

  • 10+ years of experience in software/data engineering, with at least 3+ years in a leadership role.
  • Expertise in backend development with programming languages such as Java, PHP, Python, Node.JS, GoLang, JavaScript, HTML, and CSS.
  • Proficiency in SQL, Python, and Scala for data processing and analytics.
  • Strong understanding of cloud platforms (AWS, GCP, or Azure) and their data services.
  • Strong foundation and expertise in HLD and LLD, as well as design patterns, preferably using Spring Boot or Google Guice
  • Experience in big data technologies such as Spark, Hadoop, Kafka, and distributed computing frameworks.
  • Hands-on experience with data warehousing solutions such as Snowflake, Redshift, or BigQuery
  • Deep knowledge of data governance, security, and compliance (GDPR, SOC2, etc.).
  • Experience in NoSQL databases like Redis, Cassandra, MongoDB, and TiDB.
  • Familiarity with automation and DevOps tools like Jenkins, Ansible, Docker, Kubernetes, Chef, Grafana, and ELK.
  • Proven ability to drive technical strategy and align it with business objectives.
  • Strong leadership, communication, and stakeholder management skills.


PREFERRED QUALIFICATIONS:

  • Experience in machine learning infrastructure or MLOps is a plus.
  • Exposure to real-time data processing and analytics.
  • Interest in data structures, algorithm analysis and design, multicore programming, and scalable architecture.
  • Prior experience in a SaaS or high-growth tech company.
Read more
Wissen Technology

at Wissen Technology

4 recruiters
Robin Silverster
Posted by Robin Silverster
Bengaluru (Bangalore)
5 - 11 yrs
₹10L - ₹35L / yr
skill iconPython
Spark
Apache Kafka
Snow flake schema
databricks
+1 more

Required Skills:

· 8+ years of being a practitioner in data engineering or a related field.

· Proficiency in programming skills in Python

· Experience with data processing frameworks like Apache Spark or Hadoop.

· Experience working on Databricks.

· Familiarity with cloud platforms (AWS, Azure) and their data services.

· Experience with data warehousing concepts and technologies.

· Experience with message queues and streaming platforms (e.g., Kafka).

· Excellent communication and collaboration skills.

· Ability to work independently and as part of a geographically distributed team.

Read more
Gurugram
3 - 8 yrs
₹8L - ₹20L / yr
skill iconPython
FastAPI
LangChain
Web Realtime Communication (WebRTC)
WebSocket
+6 more

Job Title : Full Stack Engineer (Real-Time Audio Systems) – Voice AI

Experience : 4+ Years

Location : Gurgaon (Hybrid)


About the Role :

We’re looking for a Voice AI / Full Stack Engineer to build our real-time Voice AI platform for low-latency, intelligent voice-driven agents in healthcare and beyond.

You’ll work closely with the founding team, combining audio infrastructure, AI, and full stack development to deliver natural, production-grade voice experiences.


Hands-on experience with WebRTC, WebSocket, and streaming services is required.

Experience with TTS (Text-to-Speech) and STT (Speech-to-Text) modules is a strong plus.


Mandatory Skills :

Python (FastAPI, Async frameworks, LangChain), WebRTC, WebSockets, Redis, Kafka, Docker, AWS, real-time streaming systems, TTS/STT modules.


Responsibilities :

  • Build and optimize voice-driven AI systems integrating ASR, TTS, and LLM inference with WebRTC & WebSocket.
  • Develop scalable backend APIs and streaming pipelines for real-time communication.
  • Translate AI audio models into reliable, production-ready services.
  • Collaborate across teams for rapid prototyping and deployment.
  • Monitor and improve system performance, latency, and reliability.

Requirements :

  • 4+ years of experience in real-time systems, streaming, or conversational AI.
  • Strong in Python (FastAPI, Async frameworks, LangChain).
  • Hands-on with WebRTC, WebSockets, Redis, Kafka, Docker, AWS.
  • Familiarity with Next.js or similar frontend frameworks is a plus.
  • Experience in healthcare tech or regulated domains preferred.

Bonus Skills :

  • Contributions to open-source audio/media projects.
  • Background in DSP, live streaming, or media infrastructure.
  • Familiarity with Grafana, Prometheus, or other observability tools.

Why Join Us :

Be part of a team working at the intersection of AI research and product engineering, shaping next-gen voice intelligence for real-world applications.

Own your systems end-to-end, innovate fast, and make a direct impact in healthcare AI.


Interview Process :

  1. Screening & Technical Task
  2. Technical Discussion
  3. HR/Leadership Round
Read more
Tata Consultancy Services
Agency job
via Risk Resources LLP hyd by Jhansi Padiy
Hyderabad
4 - 10 yrs
₹5L - ₹30L / yr
Kafka tester
Partitions
Consumer
producer
security
+2 more

Desired Competencies (Technical/Behavioral Competency)

Must-Have

·       Strong understanding of Kafka concepts, including topics, partitions, consumers, producers, and security.

·       Experience with testing Kafka Connect, Kafka Streams, and other Kafka ecosystem components.

·       API Testing Experience

·       X-RAY and Test Automation Experience

·       Expertise with Postman/SOAP

·       Agile/JIRA/Confluence

·       Strong familiarity such as XML, JSON, CSV, Avro, etc.

·       Strong hands-on SQL, Mongo.

·       Continuous integration and automated testing.

·       Working knowledge and experience of Git.

Good-to-Have

·       Troubleshoot Kafka related issues, Strong in Kafka client configuration and troubleshooting

 

SN

Responsibility of / Expectations from the Role

1

Engage with the customer to understand the requirements, provide technical solutions, provide value added suggestions

2

Help build and manage the team of Kafka and Java developers in the near future.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Manisha Gouda
Posted by Manisha Gouda
Bengaluru (Bangalore)
3 - 6 yrs
₹5L - ₹18L / yr
DevOps
Red Hat Linux
Apache Kafka
IBM WebSphere MQ
skill iconKubernetes
+1 more

- Experience comparable to DevOps SIRE providing SME-tevel application or platform support with responsibility for designing and automating operational procedures and best practices 


-Experience writing python and shell scripts to perform health checks and automations 


- Experience with Linux System Administration (preferably Red Hat)


- Hands-on experience with multi-tenant hosting environments for middleware applications (for example: centrally managed platform or infrastructure as a service) 


- Experience with implementing observabitity, monitoring and alerting tools 


- Excellent written and oral English communication skills. The candidate must write user-facing documentation, prepare and deliver presentations to an internal audience and effectively interact with upper management, colleagues, and customers 


- Independent problem-solving skills, self-motivated, and a mindset for taking ownership


- A minimum of 5 years of infrastructure production support or DevOps experience 




Additional Technical Skills 


Experience with broker-based messaging infrastructure such as Apache Kafka, IBM MQ (or similar technology like ActiveMQ, Azure Service Bus) including configuration and performance tuning 

Experience with public/private cloud and containerization technologies (e.g. Kubernetes) 

Experience with Agile development methodology, Cl/CD and automated build pipelines 

Experience with DevOps methodology (e.g. Phoenix Project) 

Experience with tools such as Jira, Confluence and ServiceNow 

Experience working with JSON, XML, Google Protocol Buffers, Avro, FIX 

Experience with troubleshooting tools such as TCPdump and Wireshark

Experience with NoSQL databases such as MongoDB and Redis interest and understanding of emerging IT trends 

Experience with system architecture design


Read more
Noida
6 - 10 yrs
₹25L - ₹35L / yr
skill iconPostgreSQL
Apache Kafka
CI/CD
Apache Airflow
Slowly changing dimensions
+2 more

Title: Data Platform / Database Architect (Postgres + Kafka) — AI‑Ready Data Infrastructure

Location: Noida (Hybrid). Remote within IST±3 considered for exceptional candidates.

Employment: Full‑time


About Us

We are building a high‑throughput, audit‑friendly data platform that powers a SaaS for financial data automation and reconciliation. The stack blends OLTP (Postgres), streaming (Kafka/Debezium), and OLAP (ClickHouse/Snowflake/BigQuery), with hooks for AI use‑cases (vector search, feature store, RAG).


Role Summary

Own the end‑to‑end design and performance of our data platform—from multi‑tenant Postgres schemas to CDC pipelines and analytics stores—while laying the groundwork for AI‑powered product features.


What You’ll Do

• Design multi‑tenant Postgres schemas (partitioning, indexing, normalization, RLS), and define retention/archival strategies.  

• Make Postgres fast and reliable: EXPLAIN/ANALYZE, connection pooling, vacuum/bloat control, query/index tuning, replication.  

• Build event‑streaming/CDC with Kafka/Debezium (topics, partitions, schema registry), and deliver data to ClickHouse/Snowflake/BigQuery.  

• Model analytics layers (star/snowflake), orchestrate jobs (Airflow/Dagster), and implement dbt‑based transformations.  

• Establish observability and SLOs for data: query/queue metrics, tracing, alerting, capacity planning.  

• Implement data security: encryption, masking, tokenization of PII, IAM boundaries; contribute to PCI‑like audit posture.  

• Integrate AI plumbing: vector embeddings (pgvector/Milvus), basic feature‑store patterns (Feast), retrieval pipelines and metadata lineage.  

• Collaborate with backend/ML/product to review designs, coach engineers, write docs/runbooks, and lead migrations.


Must‑Have Qualifications

• 6+ years building high‑scale data platforms with deep PostgreSQL experience (partitioning, advanced indexing, query planning, replication/HA).  

• Hands‑on with Kafka (or equivalent) and Debezium/CDC patterns; schema registry (Avro/Protobuf) and exactly‑once/at‑least‑once tradeoffs.  

• One or more analytics engines at scale: ClickHouse, Snowflake, or BigQuery, plus strong SQL.  

• Python for data tooling (pydantic, SQLAlchemy, or similar); orchestration with Airflow or Dagster; transformations with dbt.  

• Solid cloud experience (AWS/GCP/Azure)—networking, security groups/IAM, secrets management, cost controls.  

• Pragmatic performance engineering mindset; excellent communication and documentation.


Nice‑to‑Have

• Vector/semantic search (pgvector/Milvus/Pinecone), feature store (Feast), or RAG data pipelines.  

• Experience in fintech‑style domains (reconciliation, ledgers, payments) and SOX/PCI‑like controls.  

• Infra‑as‑Code (Terraform), containerized services (Docker/K8s), and observability stacks (Prometheus/Grafana/OpenTelemetry).  

• Exposure to Go/Java for stream processors/consumers.  

• Lakehouse formats (Delta/Iceberg/Hudi).

Read more
FiftyFive Technologies Pvt Ltd
Nishant Gandhi
Posted by Nishant Gandhi
Gurugram
5 - 9 yrs
₹10L - ₹35L / yr
skill iconJava
Microservices
skill iconSpring Boot
Apache Kafka
Maven
+9 more

Job Title: Senior Java Developer – Multi-SaaS / Microservices / Pub-Sub


About the Role

We are seeking a highly skilled Senior Java Developer with strong leadership abilities to join our team. The ideal candidate will have deep expertise in Java, micro services architecture, multi-tenant SaaS systems, pub/sub messaging, and cloud-based deployments on AWS EKS. This role requires not only technical mastery but also the ability to mentor teams, influence architecture decisions, and ensure best practices through Test-Driven Development (TDD).


Key Responsibilities

  • Lead design and development of Java-based multi-tenant SaaS applications using microservices architecture.
  • Implement pub/sub messaging systems for event-driven communication.
  • Deploy, monitor, and optimize services on AWS EKS.
  • Ensure system reliability, scalability, and security using AWS CloudWatch and other observability tools.
  • Apply TDD principles to drive high-quality, maintainable code.
  • Collaborate with product managers, architects, and other developers to translate requirements into technical solutions.
  • Lead code reviews, mentor junior engineers, and promote engineering best practices.
  • Take ownership of end-to-end delivery, from architecture design to production rollout.


Required Skills & Qualifications

  • Proficiency in Java with 6+ years of professional development experience.
  • Strong background in microservices architecture and multi-tenant SaaS systems.
  • Expertise in pub/sub messaging patterns (e.g., Kafka, RabbitMQ, or AWS SNS/SQS).
  • Hands-on experience with AWS EKS and container orchestration (Kubernetes).
  • Strong understanding of monitoring and logging using AWS CloudWatch or similar tools.
  • Proven experience applying Test-Driven Development (TDD) in real-world projects.
  • Demonstrated leadership skills – mentoring, decision-making, and driving engineering excellence.
  • Solid understanding of software engineering best practices, CI/CD, and version control (Git).


Read more
Publicis Sapient

at Publicis Sapient

10 recruiters
Dipika
Posted by Dipika
Bengaluru (Bangalore), Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Hyderabad, Pune
5 - 7 yrs
₹5L - ₹20L / yr
skill iconJava
Microservices
06692
Apache Kafka
Apache ActiveMQ
+3 more

1 Senior Associate Technology L1 – Java Microservices


Company Description

Publicis Sapient is a digital transformation partner helping established organizations get to their future, digitally-enabled state, both in the way they work and the way they serve their customers. We help unlock value through a start-up mindset and modern methods, fusing strategy, consulting and customer experience with agile engineering and problem-solving creativity. United by our core values and our purpose of helping people thrive in the brave pursuit of next, our 20,000+ people in 53 offices around the world combine experience across technology, data sciences, consulting and customer obsession to accelerate our clients’ businesses through designing the products and services their customers truly value.


Job Description

We are looking for a Senior Associate Technology Level 1 - Java Microservices Developer to join our team of bright thinkers and doers. You’ll use your problem-solving creativity to design, architect, and develop high-end technology solutions that solve our clients’ most complex and challenging problems across different industries.

We are on a mission to transform the world, and you will be instrumental in shaping how we do it with your ideas, thoughts, and solutions.


Your Impact:

• Drive the design, planning, and implementation of multifaceted applications, giving you breadth and depth of knowledge across the entire project lifecycle.

• Combine your technical expertise and problem-solving passion to work closely with clients, turning • complex ideas into end-to-end solutions that transform our clients’ business

• Constantly innovate and evaluate emerging technologies and methods to provide scalable and elegant solutions that help clients achieve their business goals.


Qualifications

➢ 5 to 7 Years of software development experience

➢ Strong development skills in Java JDK 1.8 or above

➢ Java fundamentals like Exceptional handling, Serialization/Deserialization and Immutability concepts

➢ Good fundamental knowledge in Enums, Collections, Annotations, Generics, Auto boxing and Data Structure

➢ Database RDBMS/No SQL (SQL, Joins, Indexing)

➢ Multithreading (Re-entrant Lock, Fork & Join, Sync, Executor Framework)

➢ Spring Core & Spring Boot, security, transactions ➢ Hands-on experience with JMS (ActiveMQ, RabbitMQ, Kafka etc)

➢ Memory Mgmt (JVM configuration, Profiling, GC), profiling, Perf tunning, Testing, Jmeter/similar tool)

➢ Devops (CI/CD: Maven/Gradle, Jenkins, Quality plugins, Docker and containersization)

➢ Logical/Analytical skills. Thorough understanding of OOPS concepts, Design principles and implementation of

➢ different type of Design patterns. ➢ Hands-on experience with any of the logging frameworks (SLF4J/LogBack/Log4j) ➢ Experience of writing Junit test cases using Mockito / Powermock frameworks.

➢ Should have practical experience with Maven/Gradle and knowledge of version control systems like Git/SVN etc.

➢ Good communication skills and ability to work with global teams to define and deliver on projects.

➢ Sound understanding/experience in software development process, test-driven development.

➢ Cloud – AWS / AZURE / GCP / PCF or any private cloud would also be fine

➢ Experience in Microservices

Read more
Victrix Systems  Labs

at Victrix Systems Labs

1 recruiter
Vijayalaxmi Yadav
Posted by Vijayalaxmi Yadav
Pune
6 - 9 yrs
₹20L - ₹30L / yr
Apache Kafka
skill iconElastic Search
skill iconAmazon Web Services (AWS)
skill iconJava
skill iconSpring Boot
+1 more

Role & Responsibilities :


- Lead the design, analysis, and implementation of technical solutions.


- Take full ownership of product features.


- Participate in detailed discussions with the product management team regarding requirements.


- Work closely with the engineering team to design and implement scalable solutions.


- Create detailed functional and technical specifications.


- Follow Test-Driven Development (TDD) and deliver high-quality code.


- Communicate proactively with your manager regarding risks and progress.


- Mentor junior team members and provide technical guidance.


- Troubleshoot and resolve production issues with RCA and long-term solutions


Required Skills & Experience :


- Bachelors/Masters degree in Computer Science or related field with a solid academic track record.


- 6+ years of hands-on experience in backend development for large-scale enterprise products.


- Strong programming skills in Java; familiarity with Python is a plus.


- Deep understanding of data structures, algorithms, and problem-solving.


- Proficient in Spring Boot and RESTful APIs.


- Experience with cloud technologies like ElasticSearch, Kafka, MongoDB, Hazelcast, Ceph, etc.


- Strong experience in building scalable, concurrent applications.


- Exposure to Service-Oriented Architecture (SOA) and Test-Driven Development (TDD).


- Excellent communication and collaboration skills.


Preferred Technologies :


- Java


- Spring Boot, J2EE


- ElasticSearch


- Kafka


- MongoDB, Ceph


- AWS


- Storm, Hazelcast


- TDD, SOA



Read more
iRage

at iRage

3 recruiters
Jyosana Jadhav
Posted by Jyosana Jadhav
Mumbai
4 - 7 yrs
₹25L - ₹40L / yr
skill iconReact.js
skill iconJavascript
DOM
WebSocket
Chart.js
+20 more

We are seeking a highly skilled React JS Developer with exceptional DOM manipulation expertise and real-time data handling experience to join our team. You'll be building and optimizing high-performance user interfaces for stock market trading applications where milliseconds matter and data flows continuously.


The ideal candidate thrives in fast-paced environments, understands the intricacies of browser performance, and has hands-on experience with WebSockets and real-time data streaming architectures.


Key Responsibilities


Core Development

  • Advanced DOM Operations: Implement complex, performance-optimized DOM manipulations for real-time trading interfaces
  • Real-time Data Management: Build robust WebSocket connections and handle high-frequency data streams with minimal latency
  • Performance Engineering: Create lightning-fast, scalable front-end applications that process thousands of market updates per second
  • Custom Component Architecture: Design and build reusable, high-performance React components optimized for trading workflows


Collaboration & Integration

  • Work closely with traders, quants, and backend developers to translate complex trading requirements into intuitive interfaces
  • Collaborate with UX/UI designers and product managers to create responsive, trader-focused experiences
  • Integrate with real-time market data APIs and trading execution systems


Technical Excellence

  • Implement sophisticated data visualizations and interactive charts using libraries like Chartjs, TradingView, or custom D3.js solutions
  • Ensure cross-browser compatibility and responsiveness across multiple devices and screen sizes
  • Debug and resolve complex performance issues, particularly in real-time data processing and rendering
  • Maintain high-quality code through reviews, testing, and comprehensive documentation


Required Skills & Experience


React & JavaScript Mastery

  • 5+ years of professional React.js development with deep understanding of React internals, hooks, and advanced patterns
  • Expert-level JavaScript (ES6+) with strong proficiency in asynchronous programming, closures, and memory management
  • Advanced HTML5 & CSS3 skills with focus on performance and cross-browser compatibility


Real-time & Performance Expertise

  • Proven experience with WebSockets and real-time data streaming protocols
  • Strong DOM manipulation skills - direct DOM access, virtual scrolling, efficient updates, and performance optimization
  • RESTful API integration with experience in handling high-frequency data feeds
  • Browser performance optimization - understanding of rendering pipeline, memory management, and profiling tools


Development Tools & Practices

  • Proficiency with modern build tools: Webpack, Babel, Vite, or similar
  • Experience with Git version control and collaborative development workflows
  • Agile/Scrum development environment experience
  • Understanding of testing frameworks (Jest, React Testing Library)


Financial Data Visualization

  • Experience with financial charting libraries: Chartjs, TradingView, D3.js, or custom visualization solutions
  • Understanding of market data structures, order books, and trading terminology
  • Knowledge of data streaming optimization techniques for financial applications


Nice-to-Have Skills


Domain Expertise

  • Prior experience in stock market, trading, or financial services - understanding of trading workflows, order management, risk systems
  • Algorithmic trading knowledge or exposure to quantitative trading systems
  • Financial market understanding - equities, derivatives, commodities


Technical Plus Points

  • Backend development experience with GoLang, Python, or Node.js
  • Database knowledge: SQL, NoSQL, time-series databases (InfluxDB, TimescaleDB)
  • Cloud platform experience: AWS, Azure, GCP for deploying scalable applications
  • Message queue systems: Redis, RabbitMQ, Kafka, NATS for real-time data processing
  • Microservices architecture understanding and API design principles


Advanced Skills

  • Service Worker implementation for offline-first applications
  • Progressive Web App (PWA) development
  • Mobile-first responsive design expertise


Qualifications

  • Bachelor's degree in Computer Science, Engineering, or related field (or equivalent professional experience)
  • 5+ years of professional React.js development with demonstrable experience in performance-critical applications
  • Portfolio or examples of complex real-time applications you've built
  • Financial services experience strongly preferred


Why You'll Love Working Here


We're a team that hustles—plain and simple. But we also believe life outside work matters. No cubicles, no suits—just great people doing great work in a space built for comfort and creativity.


What We Offer

💰 Competitive salary – Get paid what you're worth

🌴 Generous paid time off – Recharge and come back sharper

🌍 Work with the best – Collaborate with top-tier global talent

✈️ Adventure together – Annual offsites (mostly outside India) and regular team outings

🎯 Performance rewards – Multiple bonuses for those who go above and beyond

🏥 Health covered – Comprehensive insurance so you're always protected

Fun, not just work – On-site sports, games, and a lively workspace

🧠 Learn and lead – Regular knowledge-sharing sessions led by your peers

📚 Annual Education Stipend – Take any external course, bootcamp, or certification that makes you better at your craft

🏋️ Stay fit – Gym memberships with equal employer contribution to keep you at your best

🚚 Relocation support – Smooth move? We've got your back

🏆 Friendly competition – Work challenges and extracurricular contests to keep things exciting


We work hard, play hard, and grow together. Join us.



Read more
NonStop io Technologies Pvt Ltd
Kalyani Wadnere
Posted by Kalyani Wadnere
Pune
6 - 8 yrs
Best in industry
Salesforce
Salesforce Apex
Lightning Web Component
Salesforce API
Sales Cloud
+7 more

Key Responsibilities:-

  • Design, build, and enhance Salesforce applications using Apex, Lightning Web Components (LWC), Visualforce, and SOQL.
  • Implement integrations with external systems using REST APIs and event-driven messaging (e.g., Kafka).
  • Collaborate with architects and business analysts to translate requirements into scalable, maintainable solutions.
  • Establish and follow engineering best practices, including source control (Git), code reviews, branching strategies, CI/CD pipelines, automated testing, and environment management.
  • Establish and maintain Azure DevOps-based workflows (repos, pipelines, automated testing) for Salesforce engineering.
  • Ensure solutions follow Salesforce security, data modeling, and performance guidelines.
  • Participate in Agile ceremonies, providing technical expertise and leadership within sprints and releases.
  • Optimize workflows, automations, and data processes across Sales Cloud, Service Cloud, and custom Salesforce apps.
  • Provide technical mentoring and knowledge sharing when required.
  • Support production environments, troubleshoot issues, and drive root-cause analysis for long-term reliability.
  • Stay current on Salesforce platform updates, releases, and new features, recommending adoption where beneficial.


Required Qualifications:-

  • Bachelor’s degree in Computer Science, Engineering, or related field (or equivalent experience).
  • 6+ years of Salesforce development experience with strong knowledge of Apex, Lightning Web Components, and Salesforce APIs.
  • Proven experience with Salesforce core clouds (Sales Cloud, Service Cloud, or equivalent).
  • Strong hands-on experience with API integrations (REST/SOAP) and event-driven architectures (Kafka, Pub/Sub).
  • Solid understanding of engineering practices: Git-based source control (Salesforce DX/metadata), branching strategies, CI/CD, automated testing, and deployment management.
  • Familiarity with Azure DevOps repositories and pipelines.
  • Strong knowledge of Salesforce data modeling, security, and sharing rules.
  • Excellent problem-solving skills and ability to collaborate across teams.


Preferred Qualifications:-

  • Salesforce Platform Developer II certification (or equivalent advanced credentials).
  • Experience with Health Cloud, Financial Services Cloud, or other industry-specific Salesforce products.
  • Experience implementing logging, monitoring, and observability within Salesforce and integrated systems.
  • Background in Agile/Scrum delivery with strong collaboration skills.
  • Prior experience establishing or enforcing engineering standards across Salesforce teams.
Read more
NeoGenCode Technologies Pvt Ltd
Akshay Patil
Posted by Akshay Patil
Gurugram
2 - 6 yrs
₹4L - ₹9L / yr
skill iconPython
skill iconDjango
skill iconRedis
RabbitMQ
Celery
+5 more

Job Title : Python Django Developer

Experience : 3+ Years

Location : Gurgaon (Work from Office)


Job Summary :

We are looking for an experienced Python Django Developer with strong expertise in building scalable web applications and distributed systems. The ideal candidate must have hands-on experience with Django, Redis, Celery, RabbitMQ, PostgreSQL, and Kafka to design and optimize high-performance applications.


Mandatory Skills :

Python, Django, Redis, Celery, RabbitMQ, PostgreSQL, Kafka


Key Responsibilities :

  • Design, develop, and maintain web applications using Python & Django.
  • Implement asynchronous tasks and background job processing using Celery with RabbitMQ/Redis.
  • Work with PostgreSQL for database design, optimization, and complex queries.
  • Integrate and optimize messaging/streaming systems using Kafka.
  • Write clean, scalable, and efficient code following best practices.
  • Troubleshoot, debug, and optimize application performance.
  • Collaborate with cross-functional teams (frontend, DevOps, QA) for end-to-end delivery.
  • Stay updated with the latest backend development trends and technologies.

Requirements :

  • Minimum 3+ years of experience in backend development using Python & Django.
  • Hands-on experience with Redis, Celery, RabbitMQ, Kafka, and PostgreSQL.
  • Strong understanding of REST APIs, microservices architecture, and asynchronous task management.
  • Knowledge of performance tuning, caching strategies, and scalable system design.
  • Familiarity with Git, CI/CD pipelines, and cloud deployment (AWS/GCP/Azure) is a plus.
  • Excellent problem-solving and communication skills.
Read more
NeoGenCode Technologies Pvt Ltd
Ritika Verma
Posted by Ritika Verma
Gurugram
3 - 6 yrs
₹6L - ₹9L / yr
skill iconPython
skill iconDjango
skill iconRedis
Celery
RabbitMQ
+2 more

Job Title: Python Developer - Django (Full Time)

Location: Gurgaon, Onsite

Interview: Virtual Interview

Experience Required: 3+ Years

About the Role

We are looking for a skilled Python Developer with hands-on experience in building scalable backend systems. The ideal candidate should have strong expertise in Python, Django, distributed task queues using Celery, Redis, RabbitMQ, and experience working with event streaming platforms like Kafka.

Key Responsibilities

  • Design, develop, and maintain backend services using Python and Django.
  • Implement and optimize task queues using Celery with Redis/RabbitMQ as brokers.
  • Develop and integrate event-driven systems using Apache Kafka.
  • Write clean, reusable, and efficient code following best practices.
  • Build RESTful APIs and integrate with external services.
  • Ensure performance, scalability, and security of applications.
  • Collaborate with frontend developers, DevOps, and product teams to deliver high-quality solutions.
  • Troubleshoot and debug issues in production and staging environments.

Required Skills & Experience

  • 2+ years of professional experience in Python backend development.
  • Strong knowledge of Django Framework.
  • Hands-on experience with Celery, Redis, RabbitMQ, and Kafka.
  • Good understanding of REST API design principles.
  • Experience with relational databases (PostgreSQL/MySQL).
  • Familiarity with version control (Git) and Agile development.
  • Strong problem-solving skills and ability to work in a fast-paced environment.


Read more
MindCrew Technologies

at MindCrew Technologies

3 recruiters
Agency job
via TIGI HR Solution Pvt. Ltd. by Vaidehi Sarkar
Mumbai
6 - 8 yrs
₹10L - ₹15L / yr
skill iconC#
ADO.NET
Entity Framework
skill iconReact.js
skill iconPostgreSQL
+6 more

Role: Dot net +React Developer

Experience- 6+Years

Location- Andheri (Navi Mumbai)

Budget- 18 LPA

Opportunity - Contract 


Technical Expertise:

* Proficiency in OOP concepts, C#, .NET Core, Entity Framework, React, SQL Server, PostgreSQL, Dapper, ADO.NET, LINQ, and Web API Development.

* Experience with Kafka or Rabbit MQ for event-driven architecture and messaging systems.

* Debugging and troubleshooting skills with an understanding of performance optimization.

* Strong knowledge of database development, including tables, views, stored procedures, triggers, and functions.

* Familiarity with unit testing frameworks such as XUnit.

* Experience with JWT services, Git, and third-party API integration.

* Experience in code review of Jr. developer.


Read more
Geektrust
Agency job
via Scaling Theory by Sriraagavi Baskaran
Bengaluru (Bangalore)
2.5 - 5 yrs
₹10L - ₹30L / yr
skill iconJava
Microsoft SharePoint
Apache Kafka
skill iconElastic Search

Role Overview


We are seeking a motivated and technically versatile CMS Engineer (IC2) to support our transition from SharePoint to Contentful, while also contributing to the broader CMS ecosystem. This is an excellent opportunity for an early-career engineer to work on enterprise-grade platforms and microservice-based architectures.


Key Responsibilities

·         3-5 years of experience with SharePoint Online and/or enterprise CMS platforms.

·         Familiarity with Contentful or other headless CMS solutions is a strong plus.

·         Hands-on experience with JavaSpring Boot, and relational databases (e.g., PostgreSQL).

·         Exposure to KafkaElasticsearch, or similar distributed technologies is desirable.

·         Solid problem-solving and communication skills with an eagerness to learn.

 

What would you do here

SharePoint to Contentful Migration | Backend + CMS Integration

 

·         Assist in maintaining and enhancing SharePoint Online content and features during the transition period.

·         Support the migration of pages, documents, and metadata from SharePoint to Contentful.

·         Contribute to the design and development of backend services that integrate with Contentful using Java, Spring Boot, and REST APIs.

·         Write reusable services for content delivery, search indexing (via Elasticsearch), and event processing (via Kafka).

·         Help develop APIs for CMS-based applications that interact with PostgreSQL databases.

·         Troubleshoot CMS-related issues and support testing efforts during platform migration.


Read more
empowers digital transformation for innovative and high grow

empowers digital transformation for innovative and high grow

Agency job
via Hirebound by Jebin Joy
Pune
4 - 12 yrs
₹12L - ₹30L / yr
Hadoop
Spark
Apache Kafka
ETL
skill iconJava
+2 more

To be successful in this role, you should possess

• Collaborate closely with Product Management and Engineering leadership to devise and build the

right solution.

• Participate in Design discussions and brainstorming sessions to select, integrate, and maintain Big

Data tools and frameworks required to solve Big Data problems at scale.

• Design and implement systems to cleanse, process, and analyze large data sets using distributed

processing tools like Akka and Spark.

• Understanding and critically reviewing existing data pipelines, and coming up with ideas in

collaboration with Technical Leaders and Architects to improve upon current bottlenecks

• Take initiatives, and show the drive to pick up new stuff proactively, and work as a Senior

Individual contributor on the multiple products and features we have.

• 3+ years of experience in developing highly scalable Big Data pipelines.

• In-depth understanding of the Big Data ecosystem including processing frameworks like Spark,

Akka, Storm, and Hadoop, and the file types they deal with.

• Experience with ETL and Data pipeline tools like Apache NiFi, Airflow etc.

• Excellent coding skills in Java or Scala, including the understanding to apply appropriate Design

Patterns when required.

• Experience with Git and build tools like Gradle/Maven/SBT.

• Strong understanding of object-oriented design, data structures, algorithms, profiling, and

optimization.

• Have elegant, readable, maintainable and extensible code style.


You are someone who would easily be able to

• Work closely with the US and India engineering teams to help build the Java/Scala based data

pipelines

• Lead the India engineering team in technical excellence and ownership of critical modules; own

the development of new modules and features

• Troubleshoot live production server issues.

• Handle client coordination and be able to work as a part of a team, be able to contribute

independently and drive the team to exceptional contributions with minimal team supervision

• Follow Agile methodology, JIRA for work planning, issue management/tracking


Additional Project/Soft Skills:

• Should be able to work independently with India & US based team members.

• Strong verbal and written communication with ability to articulate problems and solutions over phone and emails.

• Strong sense of urgency, with a passion for accuracy and timeliness.

• Ability to work calmly in high pressure situations and manage multiple projects/tasks.

• Ability to work independently and possess superior skills in issue resolution.

• Should have the passion to learn and implement, analyze and troubleshoot issues

Read more
Genspark

at Genspark

2 candid answers
Agency job
via hirezyai by HR Hirezyai
Bengaluru (Bangalore), Chennai, Coimbatore
5 - 9 yrs
₹9L - ₹25L / yr
Apache Kafka
Apache
MLOps
skill iconAmazon Web Services (AWS)

The candidate should have extensive experience in designing and developing scalable data pipelines and real-time data processing solutions. As a key member of the team, the Senior Data Engineer will play a critical role in building end-to-end data workflows, supporting machine learning model deployment, and driving MLOps practices in a fast-paced, agile environment. Strong expertise in Apache Kafka, Apache Flink, AWS SageMaker, and Terraform is essential. Additional experience with infrastructure automation and CI/CD for ML models is a significant advantage.

Key Responsibilities

  1. Design, develop, and maintain high-performance ETL and real-time data pipelines using Apache Kafka and Apache Flink.
  2. Build scalable and automated MLOps pipelines for training, validation, and deployment of models using AWS SageMaker and associated services.
  3. Implement and manage Infrastructure as Code (IaC) using Terraform to provision and manage AWS environments.
  4. Collaborate with data scientists, ML engineers, and DevOps teams to streamline model deployment workflows and ensure reliable production delivery.
  5. Optimize data storage and retrieval strategies for large-scale structured and unstructured datasets.
  6. Develop data transformation logic and integrate data from various internal and external sources into data lakes and warehouses.
  7. Monitor, troubleshoot, and enhance performance of data systems in a cloud-native, fast-evolving production setup.
  8. Ensure adherence to data governance, privacy, and security standards across all data handling activities.
  9. Document data engineering solutions and workflows to facilitate cross-functional understanding and ongoing maintenance.


Read more
Gurugram, Bangarmau
5 - 8 yrs
₹12L - ₹15L / yr
skill iconJava
06692
Microservices
SQL
RESTful APIs
+3 more
  • Strong proficiency in Java programming language.
  • Experience with Java frameworks like Spring and Spring Boot.
  • Understanding of RESTful APIs and web services.
  • Experience with databases and data storage technologies (e.g., SQL, NoSQL).
  • Knowledge of software development best practices, including testing and code quality.
  • Experience with version control systems (e.g., Git).
  • Familiarity with cloud platforms (e.g., AWS, Azure, GCP).
  • Strong problem-solving and debugging skills.
  • Excellent communication and collaboration skills. 


Read more
Deqode

at Deqode

1 recruiter
purvisha Bhavsar
Posted by purvisha Bhavsar
Mumbai
2 - 4 yrs
₹4L - ₹15L / yr
skill iconMongoDB
skill iconExpress
skill iconReact.js
skill iconNodeJS (Node.js)
skill iconDocker
+3 more

🚀 Hiring: MERN Stack at Deqode

⭐ Experience: 2+ Years

📍 Location: Mumbai

⭐ Work Mode:- 5 Days Work From Office

⏱️ Notice Period: Immediate Joiners or 15 Days

(Only immediate joiners & candidates serving notice period)


MERN Stack (2+ Years of Experience) - Mumbai

🔹 Experience: 2 to 4 Years

🔹Skills: MongoDB, Express, React, Node, Docker, Kubernetes, Kafka


Read more
IT service Based
Gurugram, Bengaluru (Bangalore)
5 - 8 yrs
₹7L - ₹15L / yr
skill iconJava
Multithreading
skill iconSpring Boot
Microservices
Apache Kafka
+1 more

Looking for Java Developer for Gurugram and Bangalore Location with 5+ years of experience win Java + Microservices , Multithreading , Spring boot , Kafka and any MQ Series

Read more
AI Powered Logistics Company

AI Powered Logistics Company

Agency job
via Recruiting Bond by Pavan Kumar
Bengaluru (Bangalore)
5 - 8 yrs
₹25L - ₹32L / yr
skill iconNodeJS (Node.js)
NOSQL Databases
SQL
skill iconMongoDB
RabbitMQ
+19 more

Job Title: Backend Engineer - NodeJS, NestJS, and Python

Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)


About the role:

We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.


What You'll Do 🛠️ 

  • Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
  • Develop and integrate RESTful APIs and microservices to support scalable systems.
  • Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
  • Collaborate with cross-functional teams including frontend, DevOps, and QA.
  • Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
  • Monitor and troubleshoot production systems, ensuring high availability and performance.
  • Implement security and data protection best practices.


What You'll Bring 💼 

  • 4 to 6 years of professional experience as a backend developer.
  • Strong proficiency in Node.js and NestJS framework.
  • Good hands-on experience with Python (Django/Flask experience is a plus).
  • Solid understanding of relational and non-relational databases.
  • Proficient in writing complex NoSQL queries and SQL queries
  • Experience with microservices architecture and distributed systems.
  • Familiarity with version control systems like Git.
  • Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
  • Excellent problem-solving skills and a collaborative mindset.

 

Bonus Points ➕ 

  • Experience with CI/CD pipelines.
  • Exposure to cloud platforms like AWS, GCP or Azure.
  • Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)


Why this role matters

You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Read more
AI Powered Logistics Company

AI Powered Logistics Company

Agency job
via Recruiting Bond by Pavan Kumar
Bengaluru (Bangalore)
2 - 4 yrs
₹12L - ₹15L / yr
skill iconNodeJS (Node.js)
skill iconFlask
RESTful APIs
skill iconMongoDB
SQL
+13 more

Job Title: Backend Engineer - NodeJS, NestJS, and Python

Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)


About the role:

We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.


What You'll Do 🛠️ 

  • Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
  • Develop and integrate RESTful APIs and microservices to support scalable systems.
  • Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
  • Collaborate with cross-functional teams including frontend, DevOps, and QA.
  • Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
  • Monitor and troubleshoot production systems, ensuring high availability and performance.
  • Implement security and data protection best practices.


What You'll Bring 💼 

  • 4 to 6 years of professional experience as a backend developer.
  • Strong proficiency in Node.js and NestJS framework.
  • Good hands-on experience with Python (Django/Flask experience is a plus).
  • Solid understanding of relational and non-relational databases.
  • Proficient in writing complex NoSQL queries and SQL queries
  • Experience with microservices architecture and distributed systems.
  • Familiarity with version control systems like Git.
  • Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
  • Excellent problem-solving skills and a collaborative mindset.

 

Bonus Points ➕ 

  • Experience with CI/CD pipelines.
  • Exposure to cloud platforms like AWS, GCP or Azure.
  • Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)


Why this role matters

You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Yamuna Wissen
Posted by Yamuna Wissen
Mumbai
4 - 9 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka

Experience in Core Java and Spring Boot.

 • Extensive experience in developing enterprise-scale applications and systems. Should possess good architectural knowledge and be aware of enterprise application design patterns. 

• Should have the ability to analyze, design, develop and test complex, low-latency clientfacing applications. 

• Good development experience with RDBMS.

 • Good knowledge of multi-threading and high-performance server-side development. • Basic working knowledge of Unix/Linux. 

• Excellent problem solving and coding skills. 

• Strong interpersonal, communication and analytical skills. 

• Should have the ability to express their design ideas and thoughts

Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort