Cutshort logo
Apache Kafka Jobs in Bangalore (Bengaluru)

50+ Apache Kafka Jobs in Bangalore (Bengaluru) | Apache Kafka Job openings in Bangalore (Bengaluru)

Apply to 50+ Apache Kafka Jobs in Bangalore (Bengaluru) on CutShort.io. Explore the latest Apache Kafka Job opportunities across top companies like Google, Amazon & Adobe.

icon
Capace Software Private Limited
Bengaluru (Bangalore), Bhopal
4 - 8 yrs
₹4L - ₹8L / yr
skill iconDjango
skill iconPython
RESTful APIs
skill iconPostgreSQL
CI/CD
+7 more

Senior Python Django Developer 

Experience: Back-end development: 6 years (Required)


Location:  Bangalore/ Bhopal

Job Description:

We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the startup environment.

This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.

Responsibilities:

  • Design and develop scalable, secure, and high-performance applications using Python (Django framework).
  • Architect system components, define database schemas, and optimize backend services for speed and efficiency.
  • Lead and implement design patterns and software architecture best practices.
  • Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
  • Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
  • Drive performance improvements, monitor system health, and troubleshoot production issues.
  • Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
  • Contribute to technical decision-making and mentor junior developers.

Requirements:

  • 6 to 10 years of professional backend development experience with Python and Django.
  • Strong background in payments/financial systems or FinTech applications.
  • Proven experience in designing software architecture in a microservices or modular monolith environment.
  • Experience working in fast-paced startup environments with agile practices.
  • Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
  • Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
  • Hands-on experience with test-driven development (TDD) and frameworks like pytest, unittest, or factory_boy.
  • Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).

Preferred Skills:

  • Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
  • Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
  • Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
  • Contributions to open-source or personal finance-related projects.

Job Types: Full-time, Permanent


Schedule:

  • Day shift

Supplemental Pay:

  • Performance bonus
  • Yearly bonus

Ability to commute/relocate:

  • JP Nagar, 5th Phase, Bangalore, Karnataka or Indrapuri, Bhopal, Madhya Pradesh: Reliably commute or willing to relocate with an employer-provided relocation package (Preferred)


Read more
Tech Prescient

at Tech Prescient

3 candid answers
3 recruiters
Ashwini Kulkarni
Posted by Ashwini Kulkarni
Bengaluru (Bangalore)
8 - 10 yrs
₹15L - ₹38L / yr
skill iconJava
Microservices
skill iconSpring Boot
skill iconAmazon Web Services (AWS)
Apache Kafka

Job Title- Senior Java Developer

Required Experience- 8-10 years

Location- Bangalore-Hybrid

Desired Skills- Java, microservices, Docker, Kubernetes

Good to Have - Kotlin, Kafka

Job Type- Full time


Your key responsibilities:

● Build systems to add features across our messaging platform

● Create scalable microservices that will help support our impressive growth

● Propose scalable, maintainable and cost-effective solutions

● Estimate the effort required to develop and implement new features

● Develop and execute effective testing strategies, including unit tests, integration tests,

and end-to-end tests, to ensure software quality and reliability

● Collaborate with cross-functional teams to ensure seamless integration and functionality.


What you’ll bring

Required:

● Excellent software engineering skills in Java for backend services

● A strong focus on testing, with a customer-first mindset

● Experience with multi-threaded architectures, web services, caching, and event-driven

pipelines.

● Practical experience in deploying code in a modern programming environment using

tools such as Docker and Kubernetes

● Strong emphasis on testing and familiarity with Object-Oriented Design and design

patterns.

Read more
Auxo AI
kusuma Gullamajji
Posted by kusuma Gullamajji
Bengaluru (Bangalore), Gurugram, Hyderabad, Mumbai
8 - 12 yrs
₹20L - ₹35L / yr
skill iconPython
SQL
skill iconPostgreSQL
Apache Kafka

Responsibilities • Design, develop, and maintain backend systems and RESTful APIs using Python (Django, FastAPI, or Flask)• Build real-time communication features using WebSockets, SSE, and async IO • Implement event-driven architectures using messaging systems like Kafka, RabbitMQ, Redis Streams, or NATS • Develop and maintain microservices that interact over messaging and streaming protocols • Ensure high scalability and availability of backend services • Collaborate with frontend developers, DevOps engineers, and product managers to deliver end-to-end solutions • Write clean, maintainable code with unit/integration tests • Lead technical discussions, review code, and mentor junior engineers


Requirements • 8+ years of backend development experience, with at least 8 years in Python • Strong experience with asynchronous programming in Python (e.g., asyncio, aiohttp, FastAPI) • Production experience with WebSockets and Server-Sent Events • Hands-on experience with at least one messaging system: Kafka, RabbitMQ, Redis Pub/Sub, or similar • Proficient in RESTful API design and microservices architecture • Solid experience with relational and NoSQL databases • Familiarity with Docker and container-based deployment • Strong understanding of API security, authentication, and performance optimization


Nice to Have • Experience with GraphQL or gRPC • Familiarity with stream processing frameworks (e.g., Apache Flink, Spark Streaming) • Cloud experience (AWS, GCP, Azure), particularly with managed messaging or pub/sub services • Knowledge of CI/CD and infrastructure as code • Exposure to AI engineering workflows and tools • Interest or experience in building Agentic AI systems or integrating backends with AI agents

Read more
PGAGI
Javeriya Shaik
Posted by Javeriya Shaik
Bengaluru (Bangalore)
2 - 4 yrs
₹12L - ₹14L / yr
skill iconGo Programming (Golang)
skill iconPython
FastAPI
skill iconRedis
Apache Kafka

Backend Engineer Python / Golang / Rust


Location : Bangalore, India


Experience Required : 2-3 years minimum


Job Overview


We are seeking a skilled Backend Engineer with expertise in Python, Golang, or Rust to join our engineering team. The ideal candidate will have hands-on experience in building and maintaining enterprise-level, scalable backend services using microservices architecture.


Key Requirements :


Technical Skills :


- Programming Expertise : Advanced proficiency in Python with strong knowledge of Django, FastAPI, or Flask, OR expertise in Golang or Rust for backend development.


- Microservices Architecture : Solid experience in designing, developing, and maintaining microservices-based systems.


- Database Management : Hands-on experience with PostgreSQL, MySQL, MongoDB, including database design and optimization.


- API Development : Strong experience in designing and implementing RESTful APIs and GraphQL services.


- Cloud Platforms : Proficiency with AWS, GCP, or Azure services for deployment and scaling.


- Containerization & Orchestration : Strong knowledge of Docker and Kubernetes for scalable deployments.


- Messaging & Caching : Experience with Redis, RabbitMQ, Apache Kafka, and caching strategies (Redis, Memcached).


- Version Control : Advanced Git workflows and team collaboration best practices.


Experience Requirements :


- Minimum 2-3 years of backend development experience.


- Proven track record of working on enterprise-level, production-grade applications.


- Strong background in microservices architecture and distributed systems.


- Experience in building scalable systems capable of handling high traffic loads.


- Familiarity with CI/CD pipelines, DevOps practices, and cloud-native deployments.


Responsibilities :


- Design, develop, and maintain robust, scalable backend services and APIs.


- Architect and implement microservices solutions ensuring modularity and resilience.


- Optimize application performance, database queries, and service scalability.


- Collaborate closely with frontend teams, product managers, and DevOps engineers.


- Implement security best practices and data protection measures.


- Write and maintain comprehensive unit and integration tests.


- Participate actively in code reviews and architectural discussions.


- Monitor, debug, and optimize system performance in production environments.


Preferred Qualifications :


- Strong understanding of software architecture patterns (event-driven, CQRS, hexagonal, etc.).


- Experience with Agile/Scrum methodologies.


- Contributions to open-source projects or personal backend projects.


- Experience with observability tools (Prometheus, Grafana, ELK, Jaeger).


Read more
NeoGenCode Technologies Pvt Ltd
Shivank Bhardwaj
Posted by Shivank Bhardwaj
Remote, Bengaluru (Bangalore), Pune, Chennai
10 - 15 yrs
₹10L - ₹55L / yr
skill iconJava
skill iconSpring Boot
skill iconAmazon Web Services (AWS)
Microservices
High-level design
+5 more

We are looking for a highly skilled Solution Architect with a passion for software engineering and deep experience in backend technologies, cloud, and DevOps. This role will be central in managing, designing, and delivering large-scale, scalable solutions.


Core Skills

  • Strong coding and software engineering fundamentals.
  • Experience in large-scale custom-built applications and platforms.
  • Champion of SOLID principles, OO design, and pair programming.
  • Agile, Lean, and Continuous Delivery – CI, TDD, BDD.
  • Frontend experience is a plus.
  • Hands-on with Java, Scala, Golang, Rust, Spark, Python, and JS frameworks.
  • Experience with Docker, Kubernetes, and Infrastructure as Code.
  • Excellent understanding of cloud technologies – AWS, GCP, Azure.


Responsibilities

  • Own all aspects of technical development and delivery.
  • Understand project requirements and create architecture documentation.
  • Ensure adherence to development best practices through code reviews.


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Manisha Gouda
Posted by Manisha Gouda
Bengaluru (Bangalore)
3 - 6 yrs
₹5L - ₹18L / yr
DevOps
Red Hat Linux
Apache Kafka
IBM WebSphere MQ
skill iconKubernetes
+1 more

- Experience comparable to DevOps SIRE providing SME-tevel application or platform support with responsibility for designing and automating operational procedures and best practices 


-Experience writing python and shell scripts to perform health checks and automations 


- Experience with Linux System Administration (preferably Red Hat)


- Hands-on experience with multi-tenant hosting environments for middleware applications (for example: centrally managed platform or infrastructure as a service) 


- Experience with implementing observabitity, monitoring and alerting tools 


- Excellent written and oral English communication skills. The candidate must write user-facing documentation, prepare and deliver presentations to an internal audience and effectively interact with upper management, colleagues, and customers 


- Independent problem-solving skills, self-motivated, and a mindset for taking ownership


- A minimum of 5 years of infrastructure production support or DevOps experience 




Additional Technical Skills 


Experience with broker-based messaging infrastructure such as Apache Kafka, IBM MQ (or similar technology like ActiveMQ, Azure Service Bus) including configuration and performance tuning 

Experience with public/private cloud and containerization technologies (e.g. Kubernetes) 

Experience with Agile development methodology, Cl/CD and automated build pipelines 

Experience with DevOps methodology (e.g. Phoenix Project) 

Experience with tools such as Jira, Confluence and ServiceNow 

Experience working with JSON, XML, Google Protocol Buffers, Avro, FIX 

Experience with troubleshooting tools such as TCPdump and Wireshark

Experience with NoSQL databases such as MongoDB and Redis interest and understanding of emerging IT trends 

Experience with system architecture design


Read more
Publicis Sapient

at Publicis Sapient

10 recruiters
Dipika
Posted by Dipika
Bengaluru (Bangalore), Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Hyderabad, Pune
5 - 7 yrs
₹5L - ₹20L / yr
skill iconJava
Microservices
06692
Apache Kafka
Apache ActiveMQ
+3 more

1 Senior Associate Technology L1 – Java Microservices


Company Description

Publicis Sapient is a digital transformation partner helping established organizations get to their future, digitally-enabled state, both in the way they work and the way they serve their customers. We help unlock value through a start-up mindset and modern methods, fusing strategy, consulting and customer experience with agile engineering and problem-solving creativity. United by our core values and our purpose of helping people thrive in the brave pursuit of next, our 20,000+ people in 53 offices around the world combine experience across technology, data sciences, consulting and customer obsession to accelerate our clients’ businesses through designing the products and services their customers truly value.


Job Description

We are looking for a Senior Associate Technology Level 1 - Java Microservices Developer to join our team of bright thinkers and doers. You’ll use your problem-solving creativity to design, architect, and develop high-end technology solutions that solve our clients’ most complex and challenging problems across different industries.

We are on a mission to transform the world, and you will be instrumental in shaping how we do it with your ideas, thoughts, and solutions.


Your Impact:

• Drive the design, planning, and implementation of multifaceted applications, giving you breadth and depth of knowledge across the entire project lifecycle.

• Combine your technical expertise and problem-solving passion to work closely with clients, turning • complex ideas into end-to-end solutions that transform our clients’ business

• Constantly innovate and evaluate emerging technologies and methods to provide scalable and elegant solutions that help clients achieve their business goals.


Qualifications

➢ 5 to 7 Years of software development experience

➢ Strong development skills in Java JDK 1.8 or above

➢ Java fundamentals like Exceptional handling, Serialization/Deserialization and Immutability concepts

➢ Good fundamental knowledge in Enums, Collections, Annotations, Generics, Auto boxing and Data Structure

➢ Database RDBMS/No SQL (SQL, Joins, Indexing)

➢ Multithreading (Re-entrant Lock, Fork & Join, Sync, Executor Framework)

➢ Spring Core & Spring Boot, security, transactions ➢ Hands-on experience with JMS (ActiveMQ, RabbitMQ, Kafka etc)

➢ Memory Mgmt (JVM configuration, Profiling, GC), profiling, Perf tunning, Testing, Jmeter/similar tool)

➢ Devops (CI/CD: Maven/Gradle, Jenkins, Quality plugins, Docker and containersization)

➢ Logical/Analytical skills. Thorough understanding of OOPS concepts, Design principles and implementation of

➢ different type of Design patterns. ➢ Hands-on experience with any of the logging frameworks (SLF4J/LogBack/Log4j) ➢ Experience of writing Junit test cases using Mockito / Powermock frameworks.

➢ Should have practical experience with Maven/Gradle and knowledge of version control systems like Git/SVN etc.

➢ Good communication skills and ability to work with global teams to define and deliver on projects.

➢ Sound understanding/experience in software development process, test-driven development.

➢ Cloud – AWS / AZURE / GCP / PCF or any private cloud would also be fine

➢ Experience in Microservices

Read more
Geektrust
Agency job
via Scaling Theory by Sriraagavi Baskaran
Bengaluru (Bangalore)
2.5 - 5 yrs
₹10L - ₹30L / yr
skill iconJava
Microsoft SharePoint
Apache Kafka
skill iconElastic Search

Role Overview


We are seeking a motivated and technically versatile CMS Engineer (IC2) to support our transition from SharePoint to Contentful, while also contributing to the broader CMS ecosystem. This is an excellent opportunity for an early-career engineer to work on enterprise-grade platforms and microservice-based architectures.


Key Responsibilities

·         3-5 years of experience with SharePoint Online and/or enterprise CMS platforms.

·         Familiarity with Contentful or other headless CMS solutions is a strong plus.

·         Hands-on experience with JavaSpring Boot, and relational databases (e.g., PostgreSQL).

·         Exposure to KafkaElasticsearch, or similar distributed technologies is desirable.

·         Solid problem-solving and communication skills with an eagerness to learn.

 

What would you do here

SharePoint to Contentful Migration | Backend + CMS Integration

 

·         Assist in maintaining and enhancing SharePoint Online content and features during the transition period.

·         Support the migration of pages, documents, and metadata from SharePoint to Contentful.

·         Contribute to the design and development of backend services that integrate with Contentful using Java, Spring Boot, and REST APIs.

·         Write reusable services for content delivery, search indexing (via Elasticsearch), and event processing (via Kafka).

·         Help develop APIs for CMS-based applications that interact with PostgreSQL databases.

·         Troubleshoot CMS-related issues and support testing efforts during platform migration.


Read more
Genspark

at Genspark

2 candid answers
Agency job
via hirezyai by HR Hirezyai
Bengaluru (Bangalore), Chennai, Coimbatore
5 - 9 yrs
₹9L - ₹25L / yr
Apache Kafka
Apache
MLOps
skill iconAmazon Web Services (AWS)

The candidate should have extensive experience in designing and developing scalable data pipelines and real-time data processing solutions. As a key member of the team, the Senior Data Engineer will play a critical role in building end-to-end data workflows, supporting machine learning model deployment, and driving MLOps practices in a fast-paced, agile environment. Strong expertise in Apache Kafka, Apache Flink, AWS SageMaker, and Terraform is essential. Additional experience with infrastructure automation and CI/CD for ML models is a significant advantage.

Key Responsibilities

  1. Design, develop, and maintain high-performance ETL and real-time data pipelines using Apache Kafka and Apache Flink.
  2. Build scalable and automated MLOps pipelines for training, validation, and deployment of models using AWS SageMaker and associated services.
  3. Implement and manage Infrastructure as Code (IaC) using Terraform to provision and manage AWS environments.
  4. Collaborate with data scientists, ML engineers, and DevOps teams to streamline model deployment workflows and ensure reliable production delivery.
  5. Optimize data storage and retrieval strategies for large-scale structured and unstructured datasets.
  6. Develop data transformation logic and integrate data from various internal and external sources into data lakes and warehouses.
  7. Monitor, troubleshoot, and enhance performance of data systems in a cloud-native, fast-evolving production setup.
  8. Ensure adherence to data governance, privacy, and security standards across all data handling activities.
  9. Document data engineering solutions and workflows to facilitate cross-functional understanding and ongoing maintenance.


Read more
IT service Based
Gurugram, Bengaluru (Bangalore)
5 - 8 yrs
₹7L - ₹15L / yr
skill iconJava
Multithreading
skill iconSpring Boot
Microservices
Apache Kafka
+1 more

Looking for Java Developer for Gurugram and Bangalore Location with 5+ years of experience win Java + Microservices , Multithreading , Spring boot , Kafka and any MQ Series

Read more
AI Powered Logistics Company

AI Powered Logistics Company

Agency job
via Recruiting Bond by Pavan Kumar
Bengaluru (Bangalore)
5 - 8 yrs
₹25L - ₹32L / yr
skill iconNodeJS (Node.js)
NOSQL Databases
SQL
skill iconMongoDB
RabbitMQ
+19 more

Job Title: Backend Engineer - NodeJS, NestJS, and Python

Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)


About the role:

We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.


What You'll Do 🛠️ 

  • Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
  • Develop and integrate RESTful APIs and microservices to support scalable systems.
  • Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
  • Collaborate with cross-functional teams including frontend, DevOps, and QA.
  • Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
  • Monitor and troubleshoot production systems, ensuring high availability and performance.
  • Implement security and data protection best practices.


What You'll Bring 💼 

  • 4 to 6 years of professional experience as a backend developer.
  • Strong proficiency in Node.js and NestJS framework.
  • Good hands-on experience with Python (Django/Flask experience is a plus).
  • Solid understanding of relational and non-relational databases.
  • Proficient in writing complex NoSQL queries and SQL queries
  • Experience with microservices architecture and distributed systems.
  • Familiarity with version control systems like Git.
  • Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
  • Excellent problem-solving skills and a collaborative mindset.

 

Bonus Points ➕ 

  • Experience with CI/CD pipelines.
  • Exposure to cloud platforms like AWS, GCP or Azure.
  • Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)


Why this role matters

You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Read more
AI Powered Logistics Company

AI Powered Logistics Company

Agency job
via Recruiting Bond by Pavan Kumar
Bengaluru (Bangalore)
2 - 4 yrs
₹12L - ₹15L / yr
skill iconNodeJS (Node.js)
skill iconFlask
RESTful APIs
skill iconMongoDB
SQL
+13 more

Job Title: Backend Engineer - NodeJS, NestJS, and Python

Location: Hybrid weekly ⅔ days WFO (Bengaluru- India)


About the role:

We are looking for a skilled and passionate Senior Backend Developer to join our dynamic team. The ideal candidate should have strong experience in Node.js and NestJS, along with a solid understanding of database management, query optimization, and microservices architecture. As a backend developer, you will be responsible for developing and maintaining scalable backend systems, building robust APIs, integrating databases, and working closely with frontend and DevOps teams to deliver high-quality software solutions.


What You'll Do 🛠️ 

  • Design, develop, and maintain server-side logic using Node.js, NestJS, and Python.
  • Develop and integrate RESTful APIs and microservices to support scalable systems.
  • Work with NoSQL and SQL databases (e.g., MongoDB, PostgreSQL, MySQL) to create and manage schemas, write complex queries, and optimize performance.
  • Collaborate with cross-functional teams including frontend, DevOps, and QA.
  • Ensure code quality, maintainability, and scalability through code reviews, testing, and documentation.
  • Monitor and troubleshoot production systems, ensuring high availability and performance.
  • Implement security and data protection best practices.


What You'll Bring 💼 

  • 4 to 6 years of professional experience as a backend developer.
  • Strong proficiency in Node.js and NestJS framework.
  • Good hands-on experience with Python (Django/Flask experience is a plus).
  • Solid understanding of relational and non-relational databases.
  • Proficient in writing complex NoSQL queries and SQL queries
  • Experience with microservices architecture and distributed systems.
  • Familiarity with version control systems like Git.
  • Basic understanding of containerization (e.g., Docker) and cloud services is a plus.
  • Excellent problem-solving skills and a collaborative mindset.

 

Bonus Points ➕ 

  • Experience with CI/CD pipelines.
  • Exposure to cloud platforms like AWS, GCP or Azure.
  • Familiarity with event-driven architecture or message brokers (MQTT, Kafka, RabbitMQ)


Why this role matters

You will help build the company from the ground up—shaping our culture and having an impact from Day 1 as part of the foundational team.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Rashmi SR
Posted by Rashmi SR
Bengaluru (Bangalore)
4 - 9 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Apache Kafka
Data Structures
Algorithms
+2 more

Java Developer – Job Description Wissen Technology is now hiring for a Java Developer - Bangalore with hands-on experience in Core Java, algorithms, data structures, multithreading and SQL. We are solving complex technical problems in the industry and need talented software engineers to join our mission and be a part of a global software development team. A brilliant opportunity to become a part of a highly motivated and expert team which has made a mark as a high-end technical consulting. Required Skills: • Exp. - 4 to 7 years. • Experience in Core Java and Spring Boot. • Extensive experience in developing enterprise-scale applications and systems. Should possess good architectural knowledge and be aware of enterprise application design patterns. • Should have the ability to analyze, design, develop and test complex, low-latency client facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-performance server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills. • Strong interpersonal, communication and analytical skills. • Should have the ability to express their design ideas and thoughts. About Wissen Technology: Wissen Technology is a niche global consulting and solutions company that brings unparalleled domain expertise in Banking and Finance, Telecom and Startups. Wissen Technology is a part of Wissen Group and was established in the year 2015. Wissen has offices in the US, India, UK, Australia, Mexico, and Canada, with best-in-class infrastructure and development facilities. Wissen has successfully delivered projects worth $1 Billion for more than 25 of the Fortune 500 companies. The Wissen Group overall includes more than 4000 highly skilled professionals. Wissen Technology provides exceptional value in mission critical projects for its clients, through thought leadership, ownership, and assured on-time deliveries that are always ‘first time right’. Our team consists of 1200+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world. Wissen Technology offers an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation. We have been certified as a Great Place to Work® for two consecutive years (2020-2022) and voted as the Top 20 AI/ML vendor by CIO Insider. 

Read more
Deqode

at Deqode

1 recruiter
Alisha Das
Posted by Alisha Das
Bengaluru (Bangalore), Hyderabad, Delhi, Gurugram, Noida, Ghaziabad, Faridabad
5 - 7 yrs
₹10L - ₹25L / yr
Microsoft Windows Azure
Data engineering
skill iconPython
Apache Kafka

Role Overview:

We are seeking a Senior Software Engineer (SSE) with strong expertise in Kafka, Python, and Azure Databricks to lead and contribute to our healthcare data engineering initiatives. This role is pivotal in building scalable, real-time data pipelines and processing large-scale healthcare datasets in a secure and compliant cloud environment.

The ideal candidate will have a solid background in real-time streaming, big data processing, and cloud platforms, along with strong leadership and stakeholder engagement capabilities.

Key Responsibilities:

  • Design and develop scalable real-time data streaming solutions using Apache Kafka and Python.
  • Architect and implement ETL/ELT pipelines using Azure Databricks for both structured and unstructured healthcare data.
  • Optimize and maintain Kafka applications, Python scripts, and Databricks workflows to ensure performance and reliability.
  • Ensure data integrity, security, and compliance with healthcare standards such as HIPAA and HITRUST.
  • Collaborate with data scientists, analysts, and business stakeholders to gather requirements and translate them into robust data solutions.
  • Mentor junior engineers, perform code reviews, and promote engineering best practices.
  • Stay current with evolving technologies in cloud, big data, and healthcare data standards.
  • Contribute to the development of CI/CD pipelines and containerized environments (Docker, Kubernetes).

Required Skills & Qualifications:

  • 4+ years of hands-on experience in data engineering roles.
  • Strong proficiency in Kafka (including Kafka Streams, Kafka Connect, Schema Registry).
  • Proficient in Python for data processing and automation.
  • Experience with Azure Databricks (or readiness to ramp up quickly).
  • Solid understanding of cloud platforms, with a preference for Azure (AWS/GCP is a plus).
  • Strong knowledge of SQL and NoSQL databases; data modeling for large-scale systems.
  • Familiarity with containerization tools like Docker and orchestration using Kubernetes.
  • Exposure to CI/CD pipelines for data applications.
  • Prior experience with healthcare datasets (EHR, HL7, FHIR, claims data) is highly desirable.
  • Excellent problem-solving abilities and a proactive mindset.
  • Strong communication and interpersonal skills to work in cross-functional teams.


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Robin Silverster
Posted by Robin Silverster
Bengaluru (Bangalore)
5 - 8 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Apache Kafka
Microservices
Multithreading
+2 more

Java Developer – Job Description Wissen Technology is now hiring for a Java Developer - Bangalore with hands-on experience in Core Java, algorithms, data structures, multithreading and SQL. We are solving complex technical problems in the industry and need talented software engineers to join our mission and be a part of a global software development team. A brilliant opportunity to become a part of a highly motivated and expert team which has made a mark as a high-end technical consulting. Required Skills: • Exp. - 4 to 7 years. • Experience in Core Java and Spring Boot. • Extensive experience in developing enterprise-scale applications and systems. Should possess good architectural knowledge and be aware of enterprise application design patterns. • Should have the ability to analyze, design, develop and test complex, low-latency client facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-performance server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills. • Strong interpersonal, communication and analytical skills. • Should have the ability to express their design ideas and thoughts. About Wissen Technology: Wissen Technology is a niche global consulting and solutions company that brings unparalleled domain expertise in Banking and Finance, Telecom and Startups. Wissen Technology is a part of Wissen Group and was established in the year 2015. Wissen has offices in the US, India, UK, Australia, Mexico, and Canada, with best-in-class infrastructure and development facilities. Wissen has successfully delivered projects worth $1 Billion for more than 25 of the Fortune 500 companies. The Wissen Group overall includes more than 4000 highly skilled professionals. Wissen Technology provides exceptional value in mission critical projects for its clients, through thought leadership, ownership, and assured on-time deliveries that are always ‘first time right’. Our team consists of 1200+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world. Wissen Technology offers an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation. We have been certified as a Great Place to Work® for two consecutive years (2020-2022) and voted as the Top 20 AI/ML vendor by CIO Insider. 

Read more
QAgile Services

at QAgile Services

1 recruiter
Radhika Chotai
Posted by Radhika Chotai
Bengaluru (Bangalore)
3 - 8 yrs
₹17L - ₹25L / yr
skill iconMongoDB
skill iconPython
skill iconFlask
skill iconDjango
Windows Azure
+4 more

Employment type- Contract basis


Key Responsibilities

  • Design, develop, and maintain scalable data pipelines using PySpark and distributed computing frameworks.
  • Implement ETL processes and integrate data from structured and unstructured sources into cloud data warehouses.
  • Work across Azure or AWS cloud ecosystems to deploy and manage big data workflows.
  • Optimize performance of SQL queries and develop stored procedures for data transformation and analytics.
  • Collaborate with Data Scientists, Analysts, and Business teams to ensure reliable data availability and quality.
  • Maintain documentation and implement best practices for data architecture, governance, and security.

⚙️ Required Skills

  • Programming: Proficient in PySpark, Python, and SQL, MongoDB
  • Cloud Platforms: Hands-on experience with Azure Data Factory, Databricks, or AWS Glue/Redshift.
  • Data Engineering Tools: Familiarity with Apache Spark, Kafka, Airflow, or similar tools.
  • Data Warehousing: Strong knowledge of designing and working with data warehouses like Snowflake, BigQuery, Synapse, or Redshift.
  • Data Modeling: Experience in dimensional modeling, star/snowflake schema, and data lake architecture.
  • CI/CD & Version Control: Exposure to Git, Terraform, or other DevOps tools is a plus.

🧰 Preferred Qualifications

  • Bachelor's or Master's in Computer Science, Engineering, or related field.
  • Certifications in Azure/AWS are highly desirable.
  • Knowledge of business intelligence tools (Power BI, Tableau) is a bonus.
Read more
Deqode

at Deqode

1 recruiter
Apoorva Jain
Posted by Apoorva Jain
Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Pune, Bengaluru (Bangalore), Mumbai, Nagpur, Ahmedabad, Kochi (Cochin), Chennai
6 - 11 yrs
₹4L - ₹15L / yr
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka
Spring

Job Description:

We are looking for a Senior Java Developer with strong expertise in Apache Kafka and backend systems. The ideal candidate will have hands-on experience in Java (8/11+), Spring Boot, and building scalable, real-time data pipelines using Kafka.

Key Responsibilities:

  • Develop and maintain backend services using Java and Spring Boot
  • Design and implement Kafka-based messaging and streaming solutions
  • Optimize Kafka performance (topics, partitions, consumers)
  • Collaborate with cross-functional teams to deliver scalable microservices
  • Ensure code quality and maintain best practices in a distributed environment

Required Skills:

  • 6+ years in Java development
  • 3+ years of hands-on Kafka experience (producers, consumers, streams)
  • Strong knowledge of Spring Boot, REST APIs, and microservices
  • Familiarity with Kafka Connect, Schema Registry, and stream processing
  • Experience with containerization (Docker), CI/CD, and cloud platforms (AWS/GCP/Azure)


Read more
KJBN labs

at KJBN labs

2 candid answers
sakthi ganesh
Posted by sakthi ganesh
Bengaluru (Bangalore)
4 - 7 yrs
₹10L - ₹30L / yr
Hadoop
Apache Kafka
Spark
skill iconPython
skill iconJava
+8 more

Senior Data Engineer Job Description

Overview

The Senior Data Engineer will design, develop, and maintain scalable data pipelines and

infrastructure to support data-driven decision-making and advanced analytics. This role requires deep

expertise in data engineering, strong problem-solving skills, and the ability to collaborate with

cross-functional teams to deliver robust data solutions.

Key Responsibilities


Data Pipeline Development: Design, build, and optimize scalable, secure, and reliable data

pipelines to ingest, process, and transform large volumes of structured and unstructured data.

Data Architecture: Architect and maintain data storage solutions, including data lakes, data

warehouses, and databases, ensuring performance, scalability, and cost-efficiency.

Data Integration: Integrate data from diverse sources, including APIs, third-party systems,

and streaming platforms, ensuring data quality and consistency.

Performance Optimization: Monitor and optimize data systems for performance, scalability,

and cost, implementing best practices for partitioning, indexing, and caching.

Collaboration: Work closely with data scientists, analysts, and software engineers to

understand data needs and deliver solutions that enable advanced analytics, machine

learning, and reporting.

Data Governance: Implement data governance policies, ensuring compliance with data

security, privacy regulations (e.g., GDPR, CCPA), and internal standards.

Automation: Develop automated processes for data ingestion, transformation, and validation

to improve efficiency and reduce manual intervention.

Mentorship: Guide and mentor junior data engineers, fostering a culture of technical

excellence and continuous learning.

Troubleshooting: Diagnose and resolve complex data-related issues, ensuring high

availability and reliability of data systems.

Required Qualifications

Education: Bachelor’s or Master’s degree in Computer Science, Engineering, Data Science,

or a related field.

Experience: 5+ years of experience in data engineering or a related role, with a proven track

record of building scalable data pipelines and infrastructure.

Technical Skills:

Proficiency in programming languages such as Python, Java, or Scala.

Expertise in SQL and experience with NoSQL databases (e.g., MongoDB, Cassandra).

Strong experience with cloud platforms (e.g., AWS, Azure, GCP) and their data services

(e.g., Redshift, BigQuery, Snowflake).

Hands-on experience with ETL/ELT tools (e.g., Apache Airflow, Talend, Informatica) and

data integration frameworks.

Familiarity with big data technologies (e.g., Hadoop, Spark, Kafka) and distributed

systems.

Knowledge of containerization and orchestration tools (e.g., Docker, Kubernetes) is a

plus.

Soft Skills:

Excellent problem-solving and analytical skills.

Strong communication and collaboration abilities.

Ability to work in a fast-paced, dynamic environment and manage multiple priorities.

Certifications (optional but preferred): Cloud certifications (e.g., AWS Certified Data Analytics,

Google Professional Data Engineer) or relevant data engineering certifications.

Preferred Qualifica

Experience with real-time data processing and streaming architectures.

Familiarity with machine learning pipelines and MLOps practices.

Knowledge of data visualization tools (e.g., Tableau, Power BI) and their integration with data

pipelines.

Experience in industries with high data complexity, such as finance, healthcare, or

e-commerce.

Work Environment

Location: Hybrid/Remote/On-site (depending on company policy).

Team: Collaborative, cross-functional team environment with data scientists, analysts, and

business stakeholders.

Hours: Full-time, with occasional on-call responsibilities for critical data systems.

Read more
NeoGenCode Technologies Pvt Ltd
Akshay Patil
Posted by Akshay Patil
Hyderabad, Bengaluru (Bangalore), Mumbai, Pune, Gurugram, Chennai
4 - 10 yrs
₹5L - ₹20L / yr
Red Hat PAM Developer
Red Hat Process Automation Manager (PAM)
jBPM
JBPM (Java Business Process Management)
BPMN 2.0 (Business Process Model and Notation)
+6 more

Job Title : Red Hat PAM Developer

Experience Required :

  • Relevant Experience : 4+ Years
  • Total Experience : Up to 8 Years

No. of Positions : 4

Work Locations : Hyderabad / Bangalore / Mumbai / Pune / Gurgaon / Chennai

Work Mode : Hybrid

Work Timings : 1:00 PM to 10:00 PM IST

Interview Mode : Virtual

Interview Rounds : 2


Mandatory Skills :

  • Excellent communication skills – must be comfortable in client-facing roles
  • Red Hat Process Automation Manager (PAM)
  • JBPM (Java Business Process Management)
  • BPMN 2.0 (Business Process Model and Notation) – low-code platform
  • DMN (Decision Model and Notation) – business processes and rules
  • Spring Boot
  • JavaScript

Good-to-Have Skills :

  • Red Hat Fuse
  • Apache Kafka
  • Apache Camel
Read more
Wissen Technology

at Wissen Technology

4 recruiters
Robin Silverster
Posted by Robin Silverster
Bengaluru (Bangalore)
11 - 18 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
Multithreading
Data Structures
+3 more

Java Developer – Job Description


Wissen Technology is now hiring for a Java Developer - Bangalore with hands-on experience in Core Java, algorithms, data structures, multithreading and SQL. We are solving complex technical problems in the industry and need talented software engineers to join our mission and be a part of a global software development team. A brilliant opportunity to become a part of a highly motivated and expert team which has made a mark as a high-end technical consulting. Required Skills: • Exp. - 5 to 12 years. • Experience in Core Java and Spring Boot. • Extensive experience in developing enterprise-scale applications and systems. Should possess good architectural knowledge and be aware of enterprise application design patterns. • Should have the ability to analyze, design, develop and test complex, low-latency client facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-performance server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills. • Strong interpersonal, communication and analytical skills. • Should have the ability to express their design ideas and thoughts. About Wissen Technology: Wissen Technology is a niche global consulting and solutions company that brings unparalleled domain expertise in Banking and Finance, Telecom and Startups. Wissen Technology is a part of Wissen Group and was established in the year 2015. Wissen has offices in the US, India, UK, Australia, Mexico, and Canada, with best-in-class infrastructure and development facilities. Wissen has successfully delivered projects worth $1 Billion for more than 25 of the Fortune 500 companies. The Wissen Group overall includes more than 4000 highly skilled professionals. Wissen Technology provides exceptional value in mission critical projects for its clients, through thought leadership, ownership, and assured on-time deliveries that are always ‘first time right’. Our team consists of 1200+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world. Wissen Technology offers an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation. We have been certified as a Great Place to Work® for two consecutive years (2020-2022) and voted as the Top 20 AI/ML vendor by CIO Insider. 

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Seema Srivastava
Posted by Seema Srivastava
Bengaluru (Bangalore), Mumbai
5 - 10 yrs
Best in industry
skill iconJava
06692
Microservices
skill iconAmazon Web Services (AWS)
Apache Kafka
+1 more

Job Description: We are looking for a talented and motivated Software Engineer with

expertise in both Windows and Linux operating systems and solid experience in Java

technologies. The ideal candidate should be proficient in data structures and algorithms, as

well as frameworks like Spring MVC, Spring Boot, and Hibernate. Hands-on experience

working with MySQL databases is also essential for this role.


Responsibilities:

● Design, develop, test, and maintain software applications using Java technologies.

● Implement robust solutions using Spring MVC, Spring Boot, and Hibernate frameworks.

● Develop and optimize database operations with MySQL.

● Analyze and solve complex problems by applying knowledge of data structures and

algorithms.

● Work with both Windows and Linux environments to develop and deploy solutions.

● Collaborate with cross-functional teams to deliver high-quality products on time.

● Ensure application security, performance, and scalability.

● Maintain thorough documentation of technical solutions and processes.

● Debug, troubleshoot, and upgrade legacy systems when required.

Requirements:

● Operating Systems: Expertise in Windows and Linux environments.

● Programming Languages & Technologies: Strong knowledge of Java (Core Java, Java 8+).

● Frameworks: Proficiency in Spring MVC, Spring Boot, and Hibernate.

● Algorithms and Data Structures: Good understanding and practical application of DSA

concepts.

● Databases: Experience with MySQL – writing queries, stored procedures, and performance

tuning.

● Version Control Systems: Experience with tools like Git.

● Deployment: Knowledge of CI/CD pipelines and tools such as Jenkins, Docker (optional)

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Deepa Shankar
Posted by Deepa Shankar
Bengaluru (Bangalore)
4 - 9 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka
Multitasking
+2 more

Required Skillset

• Experience in Core Java 1.8 and above, Data Structures, OOPS, Multithreading, Algorithms, Collections, System Design, Unix/Linux. • Possess good architectural knowledge and be aware of enterprise application design patterns. • Should be able to analyze, design, develop and test complex, low-latency client-facing applications. • Good development experience with RDBMS. • Good knowledge of multi-threading and high-volume server-side development. • Basic working knowledge of Unix/Linux. • Excellent problem solving and coding skills in Java. • Strong interpersonal, communication and analytical skills. • Should be able to express their design ideas and thoughts.


Job Brief-

• Understand product requirements and come up with solution approaches. • Build and enhance large scale domain centric applications. • Deploy high quality deliverables into production adhering to the security, compliance and SDLC guidelines.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
VenkataRamanan S
Posted by VenkataRamanan S
Bengaluru (Bangalore)
3 - 8 yrs
Best in industry
06692
Microservices
Apache Kafka

Key Responsibilities:

  • Design, develop, and maintain robust and scalable backend applications using Core Java and Spring Boot.
  • Build and manage microservices-based architectures and ensure smooth inter-service communication.
  • Integrate and manage real-time data streaming using Apache Kafka.
  • Write clean, maintainable, and efficient code following best practices.
  • Collaborate with cross-functional teams including QA, DevOps, and product management.
  • Participate in code reviews and provide constructive feedback.
  • Troubleshoot, debug, and optimize applications for performance and scalability.

Required Skills:

  • Strong knowledge of Core Java (Java 8 or above).
  • Hands-on experience with Spring Boot and the Spring ecosystem (Spring MVC, Spring Data, Spring Security).
  • Experience in designing and developing RESTful APIs.
  • Solid understanding of Microservices architecture and related patterns.
  • Practical experience with Apache Kafka for real-time data processing.
  • Familiarity with SQL/NoSQL databases such as MySQL, PostgreSQL, or MongoDB.
  • Good understanding of CI/CD tools and practices.
  • Knowledge of containerization tools like Docker is a plus.
  • Strong problem-solving skills and attention to detail.


Read more
Cloudesign Technology Solutions
Anshul Saxena
Posted by Anshul Saxena
Noida, Hyderabad, Bengaluru (Bangalore)
5 - 8 yrs
₹25L - ₹30L / yr
Apache Kafka
Kafka Cluster
Schema Registry
Streaming
Kafka
+1 more

Job Title: Tech Lead and SSE – Kafka, Python, and Azure Databricks (Healthcare Data Project)

Experience: 4 to 12 years


Role Overview:

We are looking for a highly skilled Tech Lead with expertise in Kafka, Python, and Azure Databricks (preferred) to drive our healthcare data engineering projects. The ideal candidate will have deep experience in real-time data streaming, cloud-based data platforms, and large-scale data processing. This role requires strong technical leadership, problem-solving abilities, and the ability to collaborate with cross-functional teams.


Key Responsibilities:

  • Lead the design, development, and implementation of real-time data pipelines using Kafka, Python, and Azure Databricks.
  • Architect scalable data streaming and processing solutions to support healthcare data workflows.
  • Develop, optimize, and maintain ETL/ELT pipelines for structured and unstructured healthcare data.
  • Ensure data integrity, security, and compliance with healthcare regulations (HIPAA, HITRUST, etc.).
  • Collaborate with data engineers, analysts, and business stakeholders to understand requirements and translate them into technical solutions.
  • Troubleshoot and optimize Kafka streaming applications, Python scripts, and Databricks workflows.
  • Mentor junior engineers, conduct code reviews, and ensure best practices in data engineering.
  • Stay updated with the latest cloud technologies, big data frameworks, and industry trends.


Required Skills & Qualifications:

  • 4+ years of experience in data engineering, with strong proficiency in Kafka and Python.
  • Expertise in Kafka Streams, Kafka Connect, and Schema Registry for real-time data processing.
  • Experience with Azure Databricks (or willingness to learn and adopt it quickly).
  • Hands-on experience with cloud platforms (Azure preferred, AWS or GCP is a plus).
  • Proficiency in SQL, NoSQL databases, and data modeling for big data processing.
  • Knowledge of containerization (Docker, Kubernetes) and CI/CD pipelines for data applications.
  • Experience working with healthcare data (EHR, claims, HL7, FHIR, etc.) is a plus.
  • Strong analytical skills, problem-solving mindset, and ability to lead complex data projects.
  • Excellent communication and stakeholder management skills.


Read more
Talent Pro
Bengaluru (Bangalore)
4 - 8 yrs
₹26L - ₹35L / yr
skill iconJava
skill iconSpring Boot
Google Cloud Platform (GCP)
Distributed Systems
Microservices
+3 more

Role & Responsibilities

Responsible for ensuring that the architecture and design of the platform remains top-notch with respect to scalability, availability, reliability and maintainability

Act as a key technical contributor as well as a hands-on contributing member of the team.

Own end-to-end availability and performance of features, driving rapid product innovation while ensuring a reliable service.

Working closely with the various stakeholders like Program Managers, Product Managers, Reliability and Continuity Engineering(RCE) team, QE team to estimate and execute features/tasks independently.

Maintain and drive tech backlog execution for non-functional requirements of the platform required to keep the platform resilient

Assist in release planning and prioritization based on technical feasibility and engineering constraints

A zeal to continually find new ways to improve architecture, design and ensure timely delivery and high quality.

Read more
Trika Tech
bhagya a
Posted by bhagya a
Bengaluru (Bangalore), Coimbatore
7 - 8 yrs
₹12L - ₹25L / yr
skill iconNodeJS (Node.js)
AWS Lambda
Apache Kafka
skill iconKubernetes

WHO WE ARE


We are a team of digital practitioners with roots stretching back to the earliest days of online commerce, who dedicate themselves to serving our client companies.

We’ve seen the advancements first-hand over the last 25 years and believe our experiences allow us to innovate. Utilizing cutting-edge technology and providing bespoke, innovative services, we believe we can help you stay ahead of the curve.

We take a holistic view of digital strategy. Our approach to transformation is based on conscious Intent to delight customers through continuous Insight and creative Innovation with an enduring culture of problem-solving.

We bring every element together to create innovative, high-performing commerce experiences for enterprise B2C, B2B, D2C and Marketplace brands across industries. From mapping out business and functional requirements, to developing the infrastructure to optimize traditionally fragmented processes, we help you create integrated, future-proofed commerce solutions.

 

WHAT YOU’LL BE DOING

As part of our team, you'll play a key role in building and evolving our Integration Platform as a Service (iPaaS) solution. This platform empowers our clients to seamlessly connect systems, automate workflows, and scale integrations with modern cloud-native tools.

Here’s what your day-to-day will look like:

  • Designing and Building Integrations
  • Collaborate with clients to understand integration needs and build scalable, event-driven solutions using Apache Kafka, AWS Lambda, API Gateway, and EventBridge.

  • Cloud-Native Development
  • Develop and deploy microservices and serverless functions using TypeScript (Node.js), hosted on Kubernetes (EKS) and fully integrated with core AWS services like S3, SQS, and SNS.

  • Managing Data Pipelines
  • Build robust data flows and streaming pipelines using Kafka and NoSQL databases like MongoDB, ensuring high availability and fault tolerance.

  • Client Collaboration
  • Work directly with customers to gather requirements, design integration patterns, and provide guidance on best practices for cloud-native architectures.

  • Driving Platform Evolution
  • Contribute to the ongoing improvement of our iPaaS platform—enhancing observability, scaling capabilities, and CI/CD processes using modern DevOps practices.


WHAT WE NEED IN YOU


  • Solid Experience in Apache Kafka for data streaming and event-driven systems
  • Production experience with Kubernetes (EKS) and containerized deployments
  • Deep knowledge of AWS, including S3, EC2, SQS, SNS, EventBridge, Lambda
  • Proficient in TypeScript (Node.js environment)
  • Experience with MongoDB or other NoSQL databases
  • Familiarity with microservices architecture, async messaging, and DevOps practices
  • AWS Certification (e.g., Solutions Architect or Developer Associate) is a plus


Qualification

  • Graduate - BE / Btech or equivalent.
  • 5 to 8 years of experience
  • Self motivated and quick learner with excellent problem solving skills. 
  • A good team player with nice communication skills.
  • Energy and real passion to work in a startup environment.

Visit our website - https://www.trikatechnologies.com



Read more
NeoGenCode Technologies Pvt Ltd
Akshay Patil
Posted by Akshay Patil
Bengaluru (Bangalore), Pune, Chennai
10 - 20 yrs
₹30L - ₹60L / yr
skill iconJava
skill iconSpring Boot
Microservices
Apache Kafka
skill iconAmazon Web Services (AWS)
+8 more

📍 Position : Java Architect

📅 Experience : 10 to 15 Years

🧑‍💼 Open Positions : 3+

📍 Work Location : Bangalore, Pune, Chennai

💼 Work Mode : Hybrid

📅 Notice Period : Immediate joiners preferred; up to 1 month maximum

🔧 Core Responsibilities :

  • Lead architecture design and development for scalable enterprise-level applications.
  • Own and manage all aspects of technical development and delivery.
  • Define and enforce best coding practices, architectural guidelines, and development standards.
  • Plan and estimate the end-to-end technical scope of projects.
  • Conduct code reviews, ensure CI/CD, and implement TDD/BDD methodologies.
  • Mentor and lead individual contributors and small development teams.
  • Collaborate with cross-functional teams, including DevOps, Product, and QA.
  • Engage in high-level and low-level design (HLD/LLD), solutioning, and cloud-native transformations.

🛠️ Required Technical Skills :

  • Strong hands-on expertise in Java, Spring Boot, Microservices architecture
  • Experience with Kafka or similar messaging/event streaming platforms
  • Proficiency in cloud platformsAWS and Azure (must-have)
  • Exposure to frontend technologies (nice-to-have)
  • Solid understanding of HLD, system architecture, and design patterns
  • Good grasp of DevOps concepts, Docker, Kubernetes, and Infrastructure as Code (IaC)
  • Agile/Lean development, Pair Programming, and Continuous Integration practices
  • Polyglot mindset is a plus (Scala, Golang, Python, etc.)

🚀 Ideal Candidate Profile :

  • Currently working in a product-based environment
  • Already functioning as an Architect or Principal Engineer
  • Proven track record as an Individual Contributor (IC)
  • Strong engineering fundamentals with a passion for scalable software systems
  • No compromise on code quality, craftsmanship, and best practices

🧪 Interview Process :

  1. Round 1: Technical pairing round
  2. Rounds 2 & 3: Technical rounds with panel (code pairing + architecture)
  3. Final Round: HR and offer discussion
Read more
Talent Pro
Mayank choudhary
Posted by Mayank choudhary
Bengaluru (Bangalore)
3 - 5 yrs
₹20L - ₹25L / yr
ETL
SQL
Apache Spark
Apache Kafka

Role & Responsibilities

About the Role:


We are seeking a highly skilled Senior Data Engineer with 5-7 years of experience to join our dynamic team. The ideal candidate will have a strong background in data engineering, with expertise in data warehouse architecture, data modeling, ETL processes, and building both batch and streaming pipelines. The candidate should also possess advanced proficiency in Spark, Databricks, Kafka, Python, SQL, and Change Data Capture (CDC) methodologies.

Key responsibilities:


Design, develop, and maintain robust data warehouse solutions to support the organization's analytical and reporting needs.

Implement efficient data modeling techniques to optimize performance and scalability of data systems.

Build and manage data lakehouse infrastructure, ensuring reliability, availability, and security of data assets.

Develop and maintain ETL pipelines to ingest, transform, and load data from various sources into the data warehouse and data lakehouse.

Utilize Spark and Databricks to process large-scale datasets efficiently and in real-time.

Implement Kafka for building real-time streaming pipelines and ensure data consistency and reliability.

Design and develop batch pipelines for scheduled data processing tasks.

Collaborate with cross-functional teams to gather requirements, understand data needs, and deliver effective data solutions.

Perform data analysis and troubleshooting to identify and resolve data quality issues and performance bottlenecks.

Stay updated with the latest technologies and industry trends in data engineering and contribute to continuous improvement initiatives.

Read more
Bengaluru (Bangalore)
6 - 9 yrs
₹30L - ₹60L / yr
skill iconPython
skill iconDjango
skill iconFlask
skill iconPostgreSQL
Apache Kafka
+2 more

Role & Responsibilities

Lead the design, development, and deployment of complex, scalable, reliable, and highly available features for world-class SaaS products and services.

Guide the engineering team in adopting best practices for software development, code quality, and architecture.

Make strategic architectural and technical decisions, ensuring the scalability, security, and performance of software applications.

Proactively identify, prioritize, and address technical debt to improve system performance, maintainability, and long-term scalability, ensuring a solid foundation for future development.

Collaborate with cross-functional teams (product managers, designers, and stakeholders) to define project scope, requirements, and timelines.

Mentor and coach team members, providing technical guidance and fostering professional development.

Oversee code reviews, ensuring adherence to best practices and maintaining high code quality standards.

Drive continuous improvement in development processes, tools, and technologies to increase team productivity and product quality.

Stay updated with the latest industry trends and emerging technologies to drive innovation and keep the team at the cutting edge.

Ensure project timelines and goals are met, managing risks and resolving any technical challenges that arise during development.

Foster a collaborative and inclusive team culture, promoting open communication and problem-solving.

Imbibe and maintain a strong customer delight attitude while designing and building products.

Read more
Talent Pro
Mayank choudhary
Posted by Mayank choudhary
Bengaluru (Bangalore)
5 - 8 yrs
₹30L - ₹45L / yr
skill iconSpring Boot
Spring
Microservices
skill iconJava
skill iconAmazon Web Services (AWS)
+7 more

What we Require


We are recruiting technical experts with the following core skills and hands-on experience on


Mandatory skills : Core java, Microservices, AWS/Azure/GCP, Spring, Spring Boot

Hands on experience on : Kafka , Redis ,SQL, Docker, Kubernetes

Expert proficiency in designing both producer and consumer types of Rest services.

Expert proficiency in Unit testing and Code Quality tools.

Expert proficiency in ensuring code coverage.

Expert proficiency in understanding High-Level Design and translating that to Low-Level design

Hands-on experience working with no-SQL databases.

Experience working in an Agile development process - Scrum.

Experience working closely with engineers and software cultures.

Ability to think at a high level about product strategy and customer journeys.

Ability to produce low level design considering the paradigm that journeys will be extensible in the future and translate that into components that can be easily extended and reused.

Excellent communication skills to clearly articulate design decisions.

Read more
Jio Tesseract
TARUN MISHRA
Posted by TARUN MISHRA
Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Bengaluru (Bangalore), Pune, Hyderabad, Mumbai, Navi Mumbai
5 - 40 yrs
₹8.5L - ₹75L / yr
Microservices
Architecture
API
NOSQL Databases
skill iconMongoDB
+33 more

JioTesseract, a digital arm of Reliance Industries, is India's leading and largest AR/VR organization with the mission to democratize mixed reality for India and the world. We make products at the cross of hardware, software, content and services with focus on making India the leader in spatial computing. We specialize in creating solutions in AR, VR and AI, with some of our notable products such as JioGlass, JioDive, 360 Streaming, Metaverse, AR/VR headsets for consumers and enterprise space.


Mon-Fri, In office role with excellent perks and benefits!


Key Responsibilities:

1. Design, develop, and maintain backend services and APIs using Node.js or Python, or Java.

2. Build and implement scalable and robust microservices and integrate API gateways.

3. Develop and optimize NoSQL database structures and queries (e.g., MongoDB, DynamoDB).

4. Implement real-time data pipelines using Kafka.

5. Collaborate with front-end developers to ensure seamless integration of backend services.

6. Write clean, reusable, and efficient code following best practices, including design patterns.

7. Troubleshoot, debug, and enhance existing systems for improved performance.


Mandatory Skills:

1. Proficiency in at least one backend technology: Node.js or Python, or Java.


2. Strong experience in:

i. Microservices architecture,

ii. API gateways,

iii. NoSQL databases (e.g., MongoDB, DynamoDB),

iv. Kafka

v. Data structures (e.g., arrays, linked lists, trees).


3. Frameworks:

i. If Java : Spring framework for backend development.

ii. If Python: FastAPI/Django frameworks for AI applications.

iii. If Node: Express.js for Node.js development.


Good to Have Skills:

1. Experience with Kubernetes for container orchestration.

2. Familiarity with in-memory databases like Redis or Memcached.

3. Frontend skills: Basic knowledge of HTML, CSS, JavaScript, or frameworks like React.js.

Read more
Wekan Enterprise Solutions

at Wekan Enterprise Solutions

2 candid answers
Deepak  N
Posted by Deepak N
Bengaluru (Bangalore), Chennai
12 - 22 yrs
Best in industry
skill iconNodeJS (Node.js)
skill iconMongoDB
Microservices
skill iconJavascript
TypeScript
+3 more

Architect


Experience - 12+ yrs


About Wekan Enterprise Solutions


Wekan Enterprise Solutions is a leading Technology Consulting company and a strategic investment partner of MongoDB. We help companies drive innovation in the cloud by adopting modern technology solutions that help them achieve their performance and availability requirements. With strong capabilities around Mobile, IOT and Cloud environments, we have an extensive track record helping Fortune 500 companies modernize their most critical legacy and on-premise applications, migrating them to the cloud and leveraging the most cutting-edge technologies.

 

Job Description

We are looking for passionate architects eager to be a part of our growth journey. The right candidate needs to be interested in working in high-paced and challenging environments leading technical teams, designing system architecture and reviewing peer code. Interested in constantly upskilling, learning new technologies and expanding their domain knowledge to new industries. This candidate needs to be a team player and should be looking to help build a culture of excellence. Do you have what it takes?

You will be working on complex data migrations, modernizing legacy applications and building new applications on the cloud for large enterprise and/or growth stage startups. You will have the opportunity to contribute directly into mission critical projects directly interacting with business stakeholders, customer’s technical teams and MongoDB solutions Architects.

Location - Chennai or Bangalore


●     Relevant experience of 12+ years building high-performance applications with at least 3+ years as an architect.

●     Good problem solving skills

●     Strong mentoring capabilities

●     Good understanding of software development life cycle

●     Strong experience in system design and architecture

●     Strong focus on quality of work delivered

●     Excellent verbal and written communication skills

 

Required Technical Skills

 

● Extensive hands-on experience building high-performance applications using Node.Js (Javascript/Typescript) and .NET/ Golang / Java / Python.

● Strong experience with appropriate framework(s).

● Wellversed in monolithic and microservices architecture.

● Hands-on experience with data modeling on MongoDB and any other Relational or NoSQL databases

● Experience working with 3rd party integrations ranging from authentication, cloud services, etc.

● Hands-on experience with Kafka or RabbitMQ.

● Handsonexperience with CI/CD pipelines and atleast 1 cloud provider- AWS / GCP / Azure

● Strong experience writing and maintaining clear documentation

  

Good to have skills:

 

●     Experience working with frontend technologies - React.Js or Vue.Js or Angular.

●     Extensive experience consulting with customers directly for defining architecture or system design.

●     Technical certifications in AWS / Azure / GCP / MongoDB or other relevant technologies

Read more
Mernplus Technologies
Bengaluru (Bangalore)
4 - 6 yrs
₹10L - ₹17L / yr
skill iconJava
camunda
Apache Camel
Apache Kafka
karaf

We are seeking a skilled Java Developer with 5+ years of experience in Java, Camunda, Apache Camel, Kafka, and Apache Karaf. The ideal candidate should have expertise in workflow automation, message-driven architectures, and enterprise integration patterns. Strong problem-solving skills and hands-on experience in microservices and event-driven systems are required.

Read more
Gipfel & Schnell Consultings Pvt Ltd
Bengaluru (Bangalore)
5 - 12 yrs
Best in industry
DevOps
azure
Terraform
Powershell
Apache Kafka
+1 more

Mandatory Skills:


  • AZ-104 (Azure Administrator) experience
  • CI/CD migration expertise
  • Proficiency in Windows deployment and support
  • Infrastructure as Code (IaC) in Terraform
  • Automation using PowerShell
  • Understanding of SDLC for C# applications (build/ship/run strategy)
  • Apache Kafka experience
  • Azure web app


Good to Have Skills:


  • AZ-400 (Azure DevOps Engineer Expert)
  • AZ-700 Designing and Implementing Microsoft Azure Networking Solutions
  • Apache Pulsar
  • Windows containers
  • Active Directory and DNS
  • SAST and DAST tool understanding
  • MSSQL database
  • Postgres database
  • Azure security
Read more
Rigel Networks Pvt Ltd
Minakshi Soni
Posted by Minakshi Soni
Bengaluru (Bangalore), Pune, Mumbai, Chennai
8 - 12 yrs
₹8L - ₹10L / yr
skill iconAmazon Web Services (AWS)
Terraform
Amazon Redshift
Redshift
Snowflake
+16 more

Dear Candidate,


We are urgently Hiring AWS Cloud Engineer for Bangalore Location.

Position: AWS Cloud Engineer

Location: Bangalore

Experience: 8-11 yrs

Skills: Aws Cloud

Salary: Best in Industry (20-25% Hike on the current ctc)

Note:

only Immediate to 15 days Joiners will be preferred.

Candidates from Tier 1 companies will only be shortlisted and selected

Candidates' NP more than 30 days will get rejected while screening.

Offer shoppers will be rejected.


Job description:

 

Description:

 

Title: AWS Cloud Engineer

Prefer BLR / HYD – else any location is fine

Work Mode: Hybrid – based on HR rule (currently 1 day per month)


Shift Timings 24 x 7 (Work in shifts on rotational basis)

Total Experience in Years- 8+ yrs, 5 yrs of relevant exp is required.

Must have- AWS platform, Terraform, Redshift / Snowflake, Python / Shell Scripting



Experience and Skills Requirements:


Experience:

8 years of experience in a technical role working with AWS


Mandatory

Technical troubleshooting and problem solving

AWS management of large-scale IaaS PaaS solutions

Cloud networking and security fundamentals

Experience using containerization in AWS

Working Data warehouse knowledge Redshift and Snowflake preferred

Working with IaC – Terraform and Cloud Formation

Working understanding of scripting languages including Python and Shell

Collaboration and communication skills

Highly adaptable to changes in a technical environment

 

Optional

Experience using monitoring and observer ability toolsets inc. Splunk, Datadog

Experience using Github Actions

Experience using AWS RDS/SQL based solutions

Experience working with streaming technologies inc. Kafka, Apache Flink

Experience working with a ETL environments

Experience working with a confluent cloud platform


Certifications:


Minimum

AWS Certified SysOps Administrator – Associate

AWS Certified DevOps Engineer - Professional



Preferred


AWS Certified Solutions Architect – Associate


Responsibilities:


Responsible for technical delivery of managed services across NTT Data customer account base. Working as part of a team providing a Shared Managed Service.


The following is a list of expected responsibilities:


To manage and support a customer’s AWS platform

To be technical hands on

Provide Incident and Problem management on the AWS IaaS and PaaS Platform

Involvement in the resolution or high priority Incidents and problems in an efficient and timely manner

Actively monitor an AWS platform for technical issues

To be involved in the resolution of technical incidents tickets

Assist in the root cause analysis of incidents

Assist with improving efficiency and processes within the team

Examining traces and logs

Working with third party suppliers and AWS to jointly resolve incidents


Good to have:


Confluent Cloud

Snowflake




Best Regards,

Minakshi Soni

Executive - Talent Acquisition (L2)

Rigel Networks

Worldwide Locations: USA | HK | IN 

Read more
Posspole
Vibha Shashidhar
Posted by Vibha Shashidhar
Bengaluru (Bangalore)
3 - 6 yrs
₹8L - ₹14L / yr
Apache Kafka
IBM DB2
  • Strong knowledge in Kafka development and architecture.
  • Hands-on experience on KSQL Database.
  • Very good communication, analytical & problem-solving. skills.
  • Proven hands-on Development experience Kafka platforms, lenses, confluent.
  • Strong knowledge of the framework (Kafka Connect).
  • Very comfortable with Shell scripting & Linux commands.
  • Experience in DB2 database


Read more
Aadvi tech
Sravan Kumar
Posted by Sravan Kumar
Bengaluru (Bangalore), Mumbai, Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Pune, Hyderabad
4 - 10 yrs
₹15L - ₹30L / yr
Kafka
Cucumber
skill iconJava
Test Automation (QA)
Selenium
+1 more


Job description:

  • Hands on skills in Java programming languages
  • Experience of testing of Cloud Native applications with exposure of Kafka.
  • Understanding the concepts of K8, Caching, Rest / GRPC and Observability
  • Experience with good programming or scripting practices and tools: code review, ADO/Jenkin etc
  • Apply expertise in Java, API Testing, Cucumber or other test frameworks to design, develop and maintain automation test suites.
  • Intimate familiarity with QA concepts: white-/black-/grey-box testing, acceptance/regression test, system integration test, performance/stress test, and security tests
Read more
Cargill Business Services
Paramjit Kaur
Posted by Paramjit Kaur
Bengaluru (Bangalore)
2 - 6 yrs
Best in industry
Apache Kafka
Kerberos
Zookeeper
Terraform
Linux administration

As a Kafka Administrator at Cargill you will work across the full set of data platform technologies spanning on-prem and SAS solutions empowering highly performant modern data centric solutions. Your work will play a critical role in enabling analytical insights and process efficiencies for Cargill’s diverse and complex business environments. You will work in a small team who shares your passion for building, configuring, and supporting platforms while sharing, learning and growing together.  


  • Develop and recommend improvements to standard and moderately complex application support processes and procedures. 
  • Review, analyze and prioritize incoming incident tickets and user requests. 
  • Perform programming, configuration, testing and deployment of fixes or updates for application version releases. 
  • Implement security processes to protect data integrity and ensure regulatory compliance. 
  • Keep an open channel of communication with users and respond to standard and moderately complex application support requests and needs. 


MINIMUM QUALIFICATIONS

  • 2-4 year of minimum experience
  • Knowledge of Kafka cluster management, alerting/monitoring, and performance tuning
  • Full ecosystem Kafka administration (kafka, zookeeper, kafka-rest, connect)
  • Experience implementing Kerberos security
  • Preferred:
  • Experience in Linux system administration
  • Authentication plugin experience such as basic, SSL, and Kerberos
  • Production incident support including root cause analysis
  • AWS EC2
  • Terraform
Read more
Radisys India

at Radisys India

1 recruiter
Sagar bh
Posted by Sagar bh
Bengaluru (Bangalore)
5 - 10 yrs
₹20L - ₹35L / yr
skill iconJava
J2EE
skill iconSpring Boot
Microservices
Apache Kafka
+3 more

Radisys Corporation is looking for JAVA Backend developers with 6-10 years of experience for their Bangalore location.


The ideal candidate will be able to design and develop code for tasks after brainstorming sessions and applying best practices and coding conventions.


This position requires experience in Java, Spring, Spring Boot, microservices, message broker, and DB knowledge. Candidates should be skilled in developing enterprise applications that consist of FE, BE, and DB integration.


If you have experience with Docker and Kubernetes, that's an added advantage.

Read more
Radisys India

at Radisys India

1 recruiter
Sai Kiran
Posted by Sai Kiran
Bengaluru (Bangalore)
5 - 10 yrs
₹5L - ₹25L / yr
skill iconJava
J2EE
skill iconSpring Boot
Hibernate (Java)
skill iconMongoDB
+4 more

Radisys Corporation, a global leader in open telecom solutions, enables service providers to drive disruption with new open architecture business models. Our innovative technology solutions leverage open reference architectures and standards, combined with open software and hardware, to power business transformation for the telecom industry. Our services organization delivers systems integration expertise necessary to solve complex deployment challenges for communications and content providers.


Job Overview :


We are looking for a Lead Engineer - Java with a strong background in Java development and hands-on experience with J2EE, Springboot, Kubernetes, Microservices, NoSQL, and SQL. As a Lead Engineer, you will be responsible for designing and developing high-quality software solutions and ensuring the successful delivery of projects. role with 7 to 10 years of experience, based in Bangalore, Karnataka, India. This position is a full-time role with excellent growth opportunities.


Qualifications and Skills :


- Bachelor's or master's degree in Computer Science or a related field


- Strong knowledge of Core Java, J2EE, and Springboot frameworks


- Hands-on experience with Kubernetes and microservices architecture


- Experience with NoSQL and SQL databases


- Proficient in troubleshooting and debugging complex system issues


- Experience in Enterprise Applications


- Excellent communication and leadership skills


- Ability to work in a fast-paced and collaborative environment


- Strong problem-solving and analytical skills


Roles and Responsibilities :


- Work closely with product management and cross-functional teams to define requirements and deliverables


- Design scalable and high-performance applications using Java, J2EE, and Springboot


- Develop and maintain microservices using Kubernetes and containerization


- Design and implement data models using NoSQL and SQL databases


- Ensure the quality and performance of software through code reviews and testing


- Collaborate with stakeholders to identify and resolve technical issues


- Stay up-to-date with the latest industry trends and technologies


Read more
Egen Solutions
Anshul Saxena
Posted by Anshul Saxena
Remote, Hyderabad, Ahmedabad, Noida, Delhi, Gurugram, Ghaziabad, Faridabad, Kolkata, Indore, Bhopal, Kochi (Cochin), Chennai, Bengaluru (Bangalore), Pune
3 - 5 yrs
Best in industry
skill iconJava
J2EE
skill iconSpring Boot
Hibernate (Java)
skill iconKotlin
+3 more

Egen is a data engineering and cloud modernization firm helping industry-leading companies achieve digital breakthroughs and deliver for the future, today. We are catalysts for change who create digital breakthroughs at warp speed. Our team of cloud and data engineering experts are trusted by top clients in pursuit of the extraordinary. An Inc. 5000 Fastest Growing Company 7 times, and recently recognized on the Crain’s Chicago Business Fast 50 list, Egen has also been recognized as a great place to work 3 times.


You will join a team of insatiably curious data engineers, software architects, and product experts who never settle for "good enough". Our Java Platform team's tech stack is based on Java8 (Spring Boot) and RESTful web services. We typically build and deploy applications as cloud-native Kubernetes microservices and integrate with scalable technologies such as Kafka in Docker container environments. Our developers work in an agile process to efficiently deliver high value data driven applications and product packages.


Required Experience:

  • Minimum of Bachelor’s Degree or its equivalent in Computer Science, Computer Information Systems, Information Technology and Management, Electrical Engineering or a related field.
  • Have experience working and strong understanding of object-oriented programing and cloud technologies
  • End to end experience delivering production ready code with Java8, Spring Boot, Spring Data, and API libraries
  • Strong experience with unit and integration testing of the Spring Boot APIs.
  • Strong understanding and production experience of RESTful API's and microservice architecture.
  • Strong understanding of SQL databases and NoSQL databases and experience with writing abstraction layers to communicate with the databases.

Nice to have's (but not required):

  • Exposure to Kotlin or other JVM programming languages
  • Strong understanding and production experience working with Docker container environments
  • Strong understanding and production experience working with Kafka
  • Cloud Environments: AWS, GCP or Azure


Read more
Arroz Technology

at Arroz Technology

2 candid answers
Amogh Saxena
Posted by Amogh Saxena
Bengaluru (Bangalore)
1 - 3 yrs
₹1L - ₹5L / yr
skill iconAngularJS (1.x)
skill iconAngular (2+)
skill iconReact.js
skill iconNodeJS (Node.js)
skill iconMongoDB
+5 more

Job Description: Full Stack Developer Company: Arroz Technology Private Limited CTC: 5 LPA

Location : Bangalore (Onsite)

Responsibilities:

- Design and develop scalable and high-performance web applications using the

MERN (MongoDB, Express.js, React.js, Node.js) stack.

- Collaborate with cross-functional teams to gather requirements and translate them into high-level designs.

- Write clean, reusable, and well-structured code following industry best practices and coding standards.

- Mentor and guide junior developers, providing technical expertise and promoting Professional growth.

- Conduct code reviews and provide constructive feedback to ensure code quality and adherence to standards.

- Collaborate with frontend and backend developers to integrate components and ensure smooth data flow.

- Work with UI/UX designers to implement responsive and user-friendly interfaces.

- Stay updated with the latest trends and advancements in full-stack development technologies.

- Work in a 10 AM to 6 PM, six-day office role, maintaining regular attendance and punctuality.

Required Skills and Qualifications:

-Strong proficiency in MERN (MongoDB, Express.js, React.js, Node.js) stack development.

-Experience with Redux or similar state management libraries.

-Solid understanding of front-end technologies such as HTML, CSS, and JavaScript.

-Proficiency in RESTful API development and integration.

-Familiarity with version control systems like Git and agile development methodologies.

-Good problem-solving and debugging skills.

-Excellent communication and teamwork abilities.

-Bachelor's degree in Computer Science or a related field (preferred).

Join Arroz Technology Private Limited as a Full Stack Developer and contribute to the development of cutting-edge web applications. This role offers competitive compensation and growth opportunities within a dynamic work environment. 

Read more
Prolifics Corporation Ltd.,

at Prolifics Corporation Ltd.,

1 video
1 recruiter
Sandhya Patel
Posted by Sandhya Patel
Bengaluru (Bangalore)
5 - 9 yrs
Best in industry
skill iconJava
06692
Apache Kafka
Azure

Job Description:


Organization - Prolifics Corporation


Skill - Java developer


Job type - Full time/Permanent


Location - Bangalore/Mumbai


Experience - 5 to 10 Years


Notice Period – Immediate to 30 Days



Required Skillset:


Spring framework concepts, Spring boot(Mandatory)

Spring batch and dashboard

Apache Kafka(Mandatory)

Azure (Mandatory)

GIT / Maven / Griddle / CI/CD

MS SQL database

Cloud and Data Exposure


Docker, Orchestration using Kubernetes

Genesys pure cloud or any cloud-based contact center platform that can be used to manage customer interactions.

Technical Experience:


The candidate should have 5+ years of experience, preferably at technology or financial firm.

Must have at least 2- 3 years of experience in spring batch / java / Kafka / SQL

Must have hands on experience in database tools and technologies.

Must have exposure to CI / CD and Cloud.

Work scope


Build the spring batch framework to pull the required data from Genesys

Cloud to MS reporting data storage – on prem / Cloud.

Build MS WM Contact Center Data Hub (on Prem / Cloud)

Build dashboard to monitor and manage the data injection, fusion jobs.

Event bridge implementation for real time data ingestion and monitoring

MS Private Cloud

Read more
Molecular Connections

at Molecular Connections

4 recruiters
Molecular Connections
Posted by Molecular Connections
Bengaluru (Bangalore)
8 - 10 yrs
₹15L - ₹20L / yr
Spark
Hadoop
Big Data
Data engineering
PySpark
+4 more
  1. Big data developer with 8+ years of professional IT experience with expertise in Hadoop ecosystem components in ingestion, Data modeling, querying, processing, storage, analysis, Data Integration and Implementing enterprise level systems spanning Big Data.
  2. A skilled developer with strong problem solving, debugging and analytical capabilities, who actively engages in understanding customer requirements.
  3. Expertise in Apache Hadoop ecosystem components like Spark, Hadoop Distributed File Systems(HDFS), HiveMapReduce, Hive, Sqoop, HBase, Zookeeper, YARN, Flume, Pig, Nifi, Scala and Oozie.
  4. Hands on experience in creating real - time data streaming solutions using Apache Spark core, Spark SQL & DataFrames, Kafka, Spark streaming and Apache Storm.
  5. Excellent knowledge of Hadoop architecture and daemons of Hadoop clusters, which include Name node,Data node, Resource manager, Node Manager and Job history server.
  6. Worked on both Cloudera and Horton works in Hadoop Distributions. Experience in managing Hadoop clustersusing Cloudera Manager tool.
  7. Well versed in installation, Configuration, Managing of Big Data and underlying infrastructure of Hadoop Cluster.
  8. Hands on experience in coding MapReduce/Yarn Programs using Java, Scala and Python for analyzing Big Data.
  9. Exposure to Cloudera development environment and management using Cloudera Manager.
  10. Extensively worked on Spark using Scala on cluster for computational (analytics), installed it on top of Hadoop performed advanced analytical application by making use of Spark with Hive and SQL/Oracle .
  11. Implemented Spark using PYTHON and utilizing Data frames and Spark SQL API for faster processing of data and handled importing data from different data sources into HDFS using Sqoop and performing transformations using Hive, MapReduce and then loading data into HDFS.
  12. Used Spark Data Frames API over Cloudera platform to perform analytics on Hive data.
  13. Hands on experience in MLlib from Spark which are used for predictive intelligence, customer segmentation and for smooth maintenance in Spark streaming.
  14. Experience in using Flume to load log files into HDFS and Oozie for workflow design and scheduling.
  15. Experience in optimizing MapReduce jobs to use HDFS efficiently by using various compression mechanisms.
  16. Working on creating data pipeline for different events of ingestion, aggregation, and load consumer response data into Hive external tables in HDFS location to serve as feed for tableau dashboards.
  17. Hands on experience in using Sqoop to import data into HDFS from RDBMS and vice-versa.
  18. In-depth Understanding of Oozie to schedule all Hive/Sqoop/HBase jobs.
  19. Hands on expertise in real time analytics with Apache Spark.
  20. Experience in converting Hive/SQL queries into RDD transformations using Apache Spark, Scala and Python.
  21. Extensive experience in working with different ETL tool environments like SSIS, Informatica and reporting tool environments like SQL Server Reporting Services (SSRS).
  22. Experience in Microsoft cloud and setting cluster in Amazon EC2 & S3 including the automation of setting & extending the clusters in AWS Amazon cloud.
  23. Extensively worked on Spark using Python on cluster for computational (analytics), installed it on top of Hadoop performed advanced analytical application by making use of Spark with Hive and SQL.
  24. Strong experience and knowledge of real time data analytics using Spark Streaming, Kafka and Flume.
  25. Knowledge in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH3, CDH4) distributions and on Amazon web services (AWS).
  26. Experienced in writing Ad Hoc queries using Cloudera Impala, also used Impala analytical functions.
  27. Experience in creating Data frames using PySpark and performing operation on the Data frames using Python.
  28. In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS and MapReduce Programming Paradigm, High Availability and YARN architecture.
  29. Establishing multiple connections to different Redshift clusters (Bank Prod, Card Prod, SBBDA Cluster) and provide the access for pulling the information we need for analysis. 
  30. Generated various kinds of knowledge reports using Power BI based on Business specification. 
  31. Developed interactive Tableau dashboards to provide a clear understanding of industry specific KPIs using quick filters and parameters to handle them more efficiently.
  32. Well Experience in projects using JIRA, Testing, Maven and Jenkins build tools.
  33. Experienced in designing, built, and deploying and utilizing almost all the AWS stack (Including EC2, S3,), focusing on high-availability, fault tolerance, and auto-scaling.
  34. Good experience with use-case development, with Software methodologies like Agile and Waterfall.
  35. Working knowledge of Amazon's Elastic Cloud Compute( EC2 ) infrastructure for computational tasks and Simple Storage Service ( S3 ) as Storage mechanism.
  36. Good working experience in importing data using Sqoop, SFTP from various sources like RDMS, Teradata, Mainframes, Oracle, Netezza to HDFS and performed transformations on it using Hive, Pig and Spark .
  37. Extensive experience in Text Analytics, developing different Statistical Machine Learning solutions to various business problems and generating data visualizations using Python and R.
  38. Proficient in NoSQL databases including HBase, Cassandra, MongoDB and its integration with Hadoop cluster.
  39. Hands on experience in Hadoop Big data technology working on MapReduce, Pig, Hive as Analysis tool, Sqoop and Flume data import/export tools.
Read more
Kapture CX

at Kapture CX

1 video
Arunashree JS
Posted by Arunashree JS
Bengaluru (Bangalore)
3 - 4 yrs
₹8L - ₹15L / yr
skill iconJava
J2EE
skill iconSpring Boot
Hibernate (Java)
Apache Kafka
+3 more

Roles and Responsibilities:


  • Proven experience in Java 8, Spring Boot, Microservices and API
  • Strong experience with Kafka, Kubernetes
  • strong experience in using RDBMS (Mysql) and NoSQL.
  • Experience working in Eclipse or Maven environments
  • Hands-on experience in Unix and Shell scripting
  • hands-on experience in fine-tuning application response and performance testing.
  • experience in Web Services.
  • Strong analysis and problem-solving skills
  • Strong communication skills, both verbal and written
  • Ability to work independently with limited supervision
  • Proven ability to use own initiative to resolve issues
  • Full ownership of projects and tasks
  • Ability and willingness to work under pressure, on multiple concurrent tasks, and to deliver to agreed deadlines
  • Eagerness to learn
  • Strong team-working skills
Read more
Zycus

at Zycus

10 recruiters
Nafis Kurne
Posted by Nafis Kurne
Pune, Mumbai, Bangalore
14 - 26 yrs
₹25L - ₹55L / yr
skill iconVue.js
skill iconAngularJS (1.x)
skill iconAngular (2+)
skill iconReact.js
skill iconJavascript
+18 more

EXPERTISE AND QUALIFICATIONS

  • 14+ years of experience in Software Engineering with at least 6+ years as a Lead Enterprise Architect preferably in a software product company
  • High technical credibility - ability to lead technical brainstorming, take decisions and push for the best solution to a problem
  • Experience in architecting Microservices based E2E Enterprise Applications
  • Experience in UI technologies such as Angular, Node.js or Fullstack technology is desirable
  • Experience with NoSQL technologies (MongoDB, Neo4j etc.)
  • Elastic Search, Kibana, ELK, Logstash.
  • Good understanding of Kafka, Redis, ActiveMQ, RabbitMQ, Solr etc.
  • Exposure in SaaS cloud-based platform.
  • Experience on Docker, Kubernetes etc.
  • Experience in planning, designing, developing and delivering Enterprise Software using Agile Methodology
  • Key Programming Skills: Java, J2EE with cutting edge technologies
  • Hands-on technical leadership with proven ability to recruit and mentor high performance talents including Architects, Technical Leads, Developers
  • Excellent team building, mentoring and coaching skills are a must-have
  • A proven track record of consistently setting and achieving high standards

Five Reasons Why You Should Join Zycus

1. Cloud Product Company: We are a Cloud SaaS Company, and our products are created by using the latest technologies like ML and AI. Our UI is in Angular JS and we are developing our mobile apps using React.

2. A Market Leader: Zycus is recognized by Gartner (world’s leading market research analyst) as a Leader in Procurement Software Suites.

3. Move between Roles: We believe that change leads to growth and therefore we allow our employees to shift careers and move to different roles and functions within the organization

4. Get a Global Exposure: You get to work and deal with our global customers.

5. Create an Impact: Zycus gives you the environment to create an impact on the product and transform your ideas into reality. Even our junior engineers get the opportunity to work on different product features.


About Us

Zycus is a pioneer in Cognitive Procurement software and has been a trusted partner of choice for large global enterprises for two decades. Zycus has been consistently recognized by Gartner, Forrester, and other analysts for its Source to Pay integrated suite. Zycus powers its S2P software with the revolutionary Merlin AI Suite. Merlin AI takes over the tactical tasks and empowers procurement and AP officers to focus on strategic projects; offers data-driven actionable insights for quicker and smarter decisions, and its conversational AI offers a B2C type user-experience to the end-users.


Zycus helps enterprises drive real savings, reduce risks, and boost compliance, and its seamless, intuitive, and easy-

to-use user interface ensures high adoption and value across the organization.


Start your #CognitiveProcurement journey with us, as you are #MeantforMore

Read more
Mobile Programming LLC

at Mobile Programming LLC

1 video
34 recruiters
Sukhdeep Singh
Posted by Sukhdeep Singh
Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹15L / yr
skill iconJava
J2EE
skill iconSpring Boot
Hibernate (Java)
Windows Azure
+3 more
  1. Role: IoT Application Development (Java) Skill Set:
  • Proficiency in Java 11.
  • Strong knowledge of Spring Boot framework.
  • Experience with Kubernetes.
  • Familiarity with Kafka.
  • Understanding of Azure Cloud services.

1 Experience: 3 to 5 years Location: Bangalore ; Notice period : Immediate Joiners

  1. Job Description: We are seeking an experienced IoT Application Developer with expertise in Java to join our team in Bangalore. As a Java Developer, you will be responsible for designing, developing, and deploying IoT applications. You should have a solid understanding of Java 11 and the Spring Boot framework. Experience with Kubernetes and Kafka is also required. Familiarity with Azure Cloud services is essential. Your role will involve collaborating with the development team to build scalable and efficient IoT solutions using Java and related technologies.


Read more
Blue Yonder

at Blue Yonder

5 recruiters
GnanaMalleshwar Karri
Posted by GnanaMalleshwar Karri
Hyderabad, Bengaluru (Bangalore)
10 - 12 yrs
₹10L - ₹30L / yr
skill iconNodeJS (Node.js)
skill iconReact.js
skill iconAngular (2+)
skill iconAngularJS (1.x)
skill iconMongoDB
+8 more

About Merchandise Operation (Merch Ops): Merchandise Operations (Merch Ops) is a merchandise management system, it is positioned as a host system in the retail solutions, it has ability to maintain the Master/Foundation data, create and manage Purchase Orders, create, and manage Prices & Promotions, perform Replenishment, effective inventory control and financial management. Merc Ops provides Business users with consistent, accurate, and timely data across an enterprise by allowing them to get the:


Right Goods in the...

• Right Silhouettes, Sizes and Colors; at the...

• Right Price; at the...

• Right Location; for the...

• Right Consumer; at the...

• Right Time; at the...

• Right Quantity.


About Team:

• Proven, passionate bunch of disruptors providing solutions that solve real-time supply chain problems.

• Well mixed experienced team with young members and experienced in product, domain, and Industry knowledge.

• Gained Expertise in designing and deploying massively scalable cloud native SaaS products

• The team currently comprises of associates across the globe and is expected to grow rapidly.


Our current technical environment:

• Software: React JS, Node JS, Oracle PL/SQL, GIT, Rest API. Java script.

• Application Architecture: Scalable three tier web application.

• Cloud Architecture: Private cloud, MS Azure (ARM templates, AKS, HD insight, Application gateway, Virtue Networks, Event Hub, Azure AD) • Frameworks/Others: Tomcat Apache, RDBMS, Jenkins, Nginx, Oracle Type ORM, Express.


What you'll be doing:

• As a Staff Engineer you will be responsible for the design of the features in the product roadmap

• Creating and encouraging good software development practices engineering-wide, driving strategic technical improvements, and mentoring other engineers.

• You will write code as we expect our technical leadership to be in the trenches alongside junior engineers, understanding root causes and leading by example

• You will mentor engineers

• You will own relationships with other engineering teams and collaborate with other functions within Blue Yonder

• Drive architecture and designs to become simpler, more robust, and more efficient.


• Lead designs discussion and come up with robust and more efficient designs to achieve features in product roadmap

• Take complete responsibility of the features developed right from coding till deployments

• Introduce new technology and tools for the betterment of the product

• Guides fellow engineers to look beyond the surface and fix the root causes rather than symptoms.


What we are looking for:

• Bachelor’s degree (B.E/B.Tech/M.Tech Computer science or related specialization) and minimum 7 to 10 years of experience in Software development, has been an Architect, within the last 1-2 years minimum. • Strong programming experience and background in Node JS and React JS.

• Hands-on development skills along with architecture/design experience.

• Hands-on experience on designing, building deploying and maintenance of enterprise cloud solutions.

• Demonstrable experience, thorough knowledge, and interests in Cloud native architecture, Distributed micro-services, Multi-tenant SaaS solution and Cloud Scalability, performance, and High availability

• Experience with API management platforms & providing / consuming RESTful APIs

• Experience with varied tools such as REST, Hibernate, RDBMS, Docker, Kubernetes, Kafka, React.

• Hands-on development experience on Oracle PL/SQL.

• Experience with DevOps and infrastructure automation.

Read more
Recro

at Recro

1 video
32 recruiters
Amrita Singh
Posted by Amrita Singh
Bengaluru (Bangalore)
2 - 6 yrs
₹5L - ₹20L / yr
skill iconJava
J2EE
skill iconSpring Boot
Hibernate (Java)
Microservices
+5 more
  • 2.5+ year of experience in Development in JAVA technology.
  • Strong Java Basics
  • SpringBoot or Spring MVC
  • Hands on experience on Relationl Databases (SQL query or Hibernate) + Mongo (JSON parsing)
  • Proficient in REST API development
  • Messaging Queue (RabitMQ or Kafka)
  • Microservices
  • Any Caching Mechanism
  • Good at problem solving


Good to Have Skills:


  • 3+ years of experience in using Java/J2EE tech stacks
  • Good understanding of data structures and algorithms.
  • Excellent analytical and problem solving skills.
  • Ability to work in a fast paced internet start-up environment.
  • Experience in technical mentorship/coaching is highly desirable.
  • Understanding of AI/ML algorithms is a plus.
Read more
BlueYonder
Bengaluru (Bangalore), Hyderabad
10 - 14 yrs
Best in industry
skill iconJava
J2EE
skill iconSpring Boot
Hibernate (Java)
Gradle
+13 more

·      Core responsibilities to include analyze business requirements and designs for accuracy and completeness. Develops and maintains relevant product.

·      BlueYonder is seeking a Senior/Principal Architect in the Data Services department (under Luminate Platform ) to act as one of key technology leaders to build and manage BlueYonder’ s technology assets in the Data Platform and Services.

·      This individual will act as a trusted technical advisor and strategic thought leader to the Data Services department. The successful candidate will have the opportunity to lead, participate, guide, and mentor other people in the team on architecture and design in a hands-on manner. You are responsible for technical direction of Data Platform. This position reports to the Global Head, Data Services and will be based in Bangalore, India.

·      Core responsibilities to include Architecting and designing (along with counterparts and distinguished Architects) a ground up cloud native (we use Azure) SaaS product in Order management and micro-fulfillment

·      The team currently comprises of 60+ global associates across US, India (COE) and UK and is expected to grow rapidly. The incumbent will need to have leadership qualities to also mentor junior and mid-level software associates in our team. This person will lead the Data platform architecture – Streaming, Bulk with Snowflake/Elastic Search/other tools

Our current technical environment:

·      Software: Java, Springboot, Gradle, GIT, Hibernate, Rest API, OAuth , Snowflake

·      • Application Architecture: Scalable, Resilient, event driven, secure multi-tenant Microservices architecture

·      • Cloud Architecture: MS Azure (ARM templates, AKS, HD insight, Application gateway, Virtue Networks, Event Hub, Azure AD)

·      Frameworks/Others: Kubernetes, Kafka, Elasticsearch, Spark, NOSQL, RDBMS, Springboot, Gradle GIT, Ignite

Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort