50+ Kubernetes Jobs in India
Apply to 50+ Kubernetes Jobs on CutShort.io. Find your next job, effortlessly. Browse Kubernetes Jobs and apply today!


Position Overview
We are seeking a skilled Developer to join our engineering team. The ideal candidate will have strong expertise in Java and Python ecosystems, with hands-on experience in modern web technologies, messaging systems, and cloud-native development using Kubernetes.
Key Responsibilities
- Design, develop, and maintain scalable applications using Java and Spring Boot framework
- Build robust web services and APIs using Python and Flask framework
- Implement event-driven architectures using NATS messaging server
- Deploy, manage, and optimize applications in Kubernetes environments
- Develop microservices following best practices and design patterns
- Collaborate with cross-functional teams to deliver high-quality software solutions
- Write clean, maintainable code with comprehensive documentation
- Participate in code reviews and contribute to technical architecture decisions
- Troubleshoot and optimize application performance in containerized environments
- Implement CI/CD pipelines and follow DevOps best practices
- Required Qualifications
- Bachelor's degree in Computer Science, Information Technology, or related field
- 4+ years of experience in software development
- Strong proficiency in Java with deep understanding of web technology stack
- Hands-on experience developing applications with Spring Boot framework
- Solid understanding of Python programming language with practical Flask framework experience
- Working knowledge of NATS server for messaging and streaming data
- Experience deploying and managing applications in Kubernetes
- Understanding of microservices architecture and RESTful API design
- Familiarity with containerization technologies (Docker)
- Experience with version control systems (Git)
- Skills & Competencies
- Skills Java (Spring Boot, Spring Cloud, Spring Security)
- Python (Flask, SQL Alchemy, REST APIs)
- NATS messaging patterns (pub/sub, request/reply, queue groups)
- Kubernetes (deployments, services, ingress, ConfigMaps, Secrets)
- Web technologies (HTTP, REST, WebSocket, gRPC)
- Container orchestration and management
- Soft Skills Problem-solving and analytical thinking
- Strong communication and collaboration
- Self-motivated with ability to work independently
- Attention to detail and code quality
- Continuous learning mindset
- Team player with mentoring capabilities

We are looking for a Senior Software Engineer to join our team and contribute to key business functions. The ideal candidate will bring relevant experience, strong problem-solving skills, and a collaborative
mindset.
Responsibilities:
- Design, build, and maintain high-performance systems using modern C++
- Architect and implement containerized services using Docker, with orchestration via Kubernetes or ECS
- Build, monitor, and maintain data ingestion, transformation, and enrichment pipelines
- Deep understanding of cloud platforms (preferably AWS) and hands-on experience in deploying and
- managing applications in the cloud.
- Implement and maintain modern CI/CD pipelines, ensuring seamless integration, testing, and delivery
- Participate in system design, peer code reviews, and performance tuning
Qualifications:
- 5+ years of software development experience, with strong command over modern C++
- Deep understanding of cloud platforms (preferably AWS) and hands-on experience in deploying and managing applications in the cloud.
- Apache Airflow for orchestrating complex data workflows.
- EKS (Elastic Kubernetes Service) for managing containerized workloads.
- Proven expertise in designing and managing robust data pipelines & Microservices.
- Proficient in building and scaling data processing workflows and working with structured/unstructured data
- Strong hands-on experience with Docker, container orchestration, and microservices architecture
- Working knowledge of CI/CD practices, Git, and build/release tools
- Strong problem-solving, debugging, and cross-functional collaboration skills
This position description is intended to describe the duties most frequently performed by an individual in this position. It is not intended to be a complete list of assigned duties but to describe a position level.
Job Title: Backend Developer
Location: In-Office, Bangalore, Karnataka, India
Job Summary:
We are seeking a highly skilled and experienced Backend Developer with a minimum of 1 year of experience in product building to join our dynamic and innovative team. In this role, you will be responsible for designing, developing, and maintaining robust backend systems that drive our applications. You will collaborate with cross-functional teams to ensure seamless integration between frontend and backend components, and your expertise will be critical in architecting scalable, secure, and high-performance backend solutions.
Annual Compensation: 6-10 LPA
Responsibilities:
- Design, develop, and maintain scalable and efficient backend systems and APIs using NodeJS.
- Architect and implement complex backend solutions, ensuring high availability and performance.
- Collaborate with product managers, frontend developers, and other stakeholders to deliver comprehensive end-to-end solutions.
- Design and optimize data storage solutions using relational databases (e.g., MySQL) and NoSQL databases (e.g., MongoDB, Redis).
- Promoting a culture of collaboration, knowledge sharing, and continuous improvement.
- Implement and enforce best practices for code quality, security, and performance optimization.
- Develop and maintain CI/CD pipelines to automate build, test, and deployment processes.
- Ensure comprehensive test coverage, including unit testing, and implement various testing methodologies and tools to validate application functionality.
- Utilize cloud services (e.g., AWS, Azure, GCP) for infrastructure deployment, management, and optimization.
- Conduct system design reviews and contribute to architectural discussions.
- Stay updated with industry trends and emerging technologies to drive innovation within the team.
- Implement secure authentication and authorization mechanisms and ensure data encryption for sensitive information.
- Design and develop event-driven applications utilizing serverless computing principles to enhance scalability and efficiency.
Requirements:
- Minimum of 1 year of proven experience as a Backend Developer, with a strong portfolio of product-building projects.
- Extensive experience with JavaScript backend frameworks (e.g., Express, Socket) and a deep understanding of their ecosystems.
- Strong expertise in SQL and NoSQL databases (MySQL and MongoDB) with a focus on data modeling and scalability.
- Practical experience with Redis and caching mechanisms to enhance application performance.
- Proficient in RESTful API design and development, with a strong understanding of API security best practices.
- In-depth knowledge of asynchronous programming and event-driven architecture.
- Familiarity with the entire web stack, including protocols, web server optimization techniques, and performance tuning.
- Experience with containerization and orchestration technologies (e.g., Docker, Kubernetes) is highly desirable.
- Proven experience working with cloud technologies (AWS/GCP/Azure) and understanding of cloud architecture principles.
- Strong understanding of fundamental design principles behind scalable applications and microservices architecture.
- Excellent problem-solving, analytical, and communication skills.
- Ability to work collaboratively in a fast-paced, agile environment and lead projects to successful completion.
How to Apply
Visit: https://www.thealteroffice.com/about


Build a dynamic solution empowering companies to optimize promotional activities for maximum impact. It collects and validates data, analyzes promotion effectiveness, plans calendars, and integrates seamlessly with existing systems. The tool enhances vendor collaboration, negotiates better deals, and employs machine learning to optimize promotional plans, enabling companies to make informed decisions and maximize return on investment.
Technology Stack: Scala, Go-Lang, Docker, Kubernetes, Databricks, Python optional.
Working Time Zone: EU
Specialty: Data Science
Level of the Candidate: more then 5 years of experience
Language Proficiency: English Upper-Intermediate
Required Soft Skills:
- Prefer problem solving style over experience
- Ability to clarify requirements with the customer
- Willingness to pair with other engineers when solving complex issues
- Good communication skills
Hard Skills / Need to Have:
- Experience in Scala and/or Go, designing and building scalable high-performing applications
- Experience in containerization and microservices orchestration using Docker and Kubernetes
- Experience in building data pipelines and ETL solutions using Databricks
- Experience in data storage and retrieval with PostgreSQL and Elasticsearch
- Experience in deploying and maintaining solutions in the Azure cloud environment
- Experience in Python is nice to have
Responsibilities and Tasks:
- Develop and maintain distributed systems using Scala and/or Go
- Work with Docker and Kubernetes for containerization and microservices orchestration
- Build data pipelines and ETL solutions using Databricks
- Work with PostgreSQL and Elasticsearch for data storage and retrieval
- Deploy and maintain solutions in the Azure cloud environment

About Astra:
Astra is a cyber security SaaS company that makes otherwise chaotic pentests a breeze with its one of a kind Pentest Platform. Astra's continuous vulnerability scanner emulates hacker behavior to scan applications for 9300+ security tests. CTOs & CISOs love Astra because it helps them fix vulnerabilities in record time and move from DevOps to DevSecOps with Astra's CI/CD integrations.
Astra is loved by 650+ companies across the globe. In 2023 Astra uncovered 2 million+ vulnerabilities for its customers, saving customers $69M+ in potential losses due to security vulnerabilities.
We've been awarded by the President of France Mr. François Hollande at the La French Tech program and Prime Minister of India Shri Narendra Modi at the Global Conference on Cyber Security. Loom, MamaEarth, Muthoot Finance, Canara Robeco, ScripBox etc. are a few of Astra’s customers.
Role Overview:
As an SDE 2 back-end engineer you will be responsible for leading the development and maintenance of the platform/dashboard backend. This would involve architecting, developing & maintaining RESTful services for vulnerability management, scan orchestration, inventory management and platform features such as on-boarding, trust centers, certificates, payment integrations, vulnerability ingestion, authentication etc.
You should have a strong background in backend programming (Symfony preferred, or Laravel) and have experience with an event driven, async & distributed architecture.
During the first 6 months of your role, you will be involved in the 0->1 journey of two of our upcoming products in our platform along with the respective product owners.
Love solving hard problems? Want to build high impact products rooted in first principles? Is coding your poetry? Join us in shaping the future of cyber security.
Roles & Responsibilities:
- Lead the design, development, and maintenance of backend services and APIs using Symfony PHP framework. What you create is also what you own.
- Collaborate with front-end developers to integrate user-facing elements with server-side logic.
- Collaborate with scanner teams to orchestrate scans, ingest vulnerabilities, configure scanners etc.
- Write clean, well-documented, and efficient code.
- Optimize and refactor existing code to improve performance and reliability.
- Implement security and data protection measures.
- Triage, troubleshoot and upgrade existing systems.
- Ship code to production multiple times a day/week.
- Ensure timely delivery of the features.
- Test your own features, and write test cases for continuous automated testing.
- Troubleshoot complex issues, debug, and upgrade existing systems.
- Mentor junior engineers and lead code reviews, promoting best practices and high standards.
- Drive architectural decisions and technical strategy for backend systems.
What we are looking for:
- 4-6 years experience in a similar role or similar working experience.
- Advanced proficiency in PHP backend programming. Symfony preferred, or Laravel.
- Strong understanding of system design, software architecture principles and performance optimization.
- Experience with an event driven (subscribers/listeners), async (MessageHandlers & job queues) & distributed (load-balanced) architecture.
- Experience with PostgreSQL database and data modeling.
- Understanding of concepts such as Dependency Injection, ORM, data validation, error handling etc.
- Experience with Docker, Kubernetes, GitHub Actions (good to have).
- Experience with Agile methodologies.
- Excellent problem-solving skills and the ability to think strategically about technical solutions.
- Strong communication and interpersonal skills, with the ability to collaborate effectively with cross-functional remote teams.
- Demonstrated track record of delivering high-quality software products on schedule.
- Knowledge of industry best practices in software development, security, and compliance.
- Knowledge of application & infrastructure security - helps you stand out.
What We offer:
- Adrenalin rush of being a part of a fast-growing company, and working on hard problems that matter.
- Fully remote, agile working environment.
- Good engineering culture with full ownership in design, development, release lifecycle.
- A wholesome opportunity where you get to build things from scratch, improve and ship code to production in hours, not weeks.
- Holistic understanding of SaaS and enterprise security business.
- Experience with security side of things.
- Annual trips to beaches or mountains (next one is to Wayanad).
- Open and supportive culture.
- Health insurance & other benefits for you and your spouse. Maternity benefits included.
Job Title: Lead Java Developer
Location: Bangalore
Experience: 8-12 years (only)
Job Overview:
We are looking for a highly experienced Java / Backend Developer with a strong background in investment banking/trading/brokerages to join our team. The ideal candidate will have extensive experience in developing front office and middle office systems which require high availability, high reliability and low latency/high throughput performance profiles.
Skills & Qualifications:
- 8 to 12 years of experience in Java development, with at least 4 years in the investment banking/financial services domain.
- Strong knowledge of Java, Spring Framework, RESTful web services.
- Hands-on experience with AWS Cloud and related services. Knowledge of Kubernetes is a plus.
- Solid experience with Kafka and/or messaging protocols.
- Familiarity with SQL databases (e.g., Postgres/Oracle).
- Strong problem-solving skills and ability to work in a team.
- Ability to understand and work with distributed systems / microservices architecture.
- Solid written and verbal communication skills to interface with other technology and business stakeholders.
1.4+ years of software development experience
2. Strong experience with Kubernetes, Docker, and CI/CD pipelines in cloud-native environments.
3. Hands-on with NATS for event-driven architecture and streaming.
4. Skilled in microservices, RESTful APIs, and containerized app performance optimization.
5. Strong in problem-solving, team collaboration, clean code practices, and continuous learning.
6. Proficient in Java (Spring Boot) and Python (Flask) for building scalable applications and APIs.
7. Focus: Java, Python, Kubernetes, Cloud-native development
About Us:
At Vahan, we are building the first AI powered recruitment marketplace for India’s 300 million strong Blue Collar workforce, opening doors to economic opportunities and brighter futures.
Already India’s largest recruitment platform, Vahan is supported by marquee investors like Khosla Ventures, Y Combinator, Airtel, Vijay Shekhar Sharma (CEO, Paytm), and leading executives from Google and Facebook.
Our customers include names like Swiggy, Zomato, Rapido, Zepto, and many more. We leverage cutting-edge technology and AI to recruit for the workforces of some of the most recognized companies in the country.
Our vision is ambitious: to become the go-to platform for blue-collar professionals worldwide, empowering them with not just earning opportunities but also the tools, benefits, and support they need to thrive. We aim to impact over a billion lives worldwide, creating a future where everyone has access to economic prosperity.
If our vision excites you, Vahan might just be your next adventure. We’re on the hunt for driven individuals who love tackling big challenges. If this sounds like your kind of journey, dive into the details and see where you can make your mark.
What You Will Be Doing:
- Build & Automate Cloud Infrastructure – Design, deploy, and optimize cloud environments, ensuring scalability, reliability, and cost efficiency.
- Set Up CI/CD & Deployment Pipelines – Develop automated workflows to streamline code integration, testing, and deployment for faster releases.
- Monitor & Improve System Performance – Implement robust monitoring, logging, and alerting mechanisms to proactively identify and resolve issues.
- Manage Containers & Scalability – Deploy and maintain containerized applications, ensuring efficient resource utilization and high availability.
- Ensure Security & Reliability – Enforce access controls, backup strategies, and disaster recovery plans to safeguard infrastructure and data.
- Adapt & Scale with the Startup – Take on dynamic responsibilities, quickly learn new technologies, and evolve processes to meet growing business needs.
You Will Thrive in This Role If You:
Must Haves:
- Experience: 3+ years in DevOps or related roles, focusing on cloud environments, automation, CI/CD, and Linux system administration. Strong expertise in debugging and infrastructure performance improvements.
- Cloud Expertise: In-depth experience with one or more cloud platforms (AWS, GCP), including services like EC2, RDS, S3, VPC, etc.
- IaC Tools: Proficiency in Terraform, Ansible, CloudFormation, or similar tools.
- Scripting Skills: Strong scripting abilities in Python, Bash, or PowerShell.
- Containerization: Experience with Docker, including managing containers in production.
- Monitoring Tools: Hands-on experience with tools like ELK, Prometheus, Grafana, CloudWatch, New Relic, and Data dog.
- Version Control: Proficiency with Git and code repository management.
- Soft Skills: Excellent problem-solving skills, attention to detail, and effective communication with both technical and non-technical team members.
- Database Management: Experience with managing and tuning databases like MySQL and PostgreSQL.
- Deployment Pipelines: Experience with Jenkins and similar CI/CD tools.
- Message Queues: Experience with rabbitMQ/SQS/Kafka.
Nice to Have:
- Certifications: AWS Certified DevOps Engineer, Certified Kubernetes Administrator (CKA), or similar.
- SRE Practices: Familiarity with Site Reliability Engineering (SRE) principles, including error budgeting and service level objectives (SLOs).
- Serverless Computing: Knowledge of AWS Lambda, Azure Functions, or similar architectures.
- Containerization: Experience with Docker and Kubernetes, including managing production clusters.
- Security: Awareness of security best practices and implementations.
- Cloud Cost Optimization: Experience with cost-saving initiatives in cloud environments.
- Data Pipelines & ETL: Experience in setting up and managing data pipelines and ETL workflows.
- Familiarity with Modern Tech Stacks: Exposure to Python, Node.js, React.js, and Kotlin for app deployment CI/CD pipelines.
- MLOps Pipelines: Understanding of ML model deployment and operationalization.
- Data Retrieval & Snapshots: Experience with PITI, EC2, and RDS snapshots.
- System Resiliency & Recovery: Strategies for ensuring system reliability and recovery in case of downtime.
At Vahan, you’ll have the opportunity to make a real impact in a sector that touches millions of lives. We’re committed to not only advancing the livelihoods of our workforce but also, in taking care of the people who make this mission possible. Here’s what we offer:
- Unlimited PTO: Trust and flexibility to manage your time in the way that works best for you.
- Comprehensive Medical Insurance: We’ve got you covered with plans designed to support you and your loved ones.
- Monthly Wellness Leaves: Regular time off to recharge and focus on what matters most.
- Competitive Pay: Your contributions are recognized and rewarded with a compensation package that reflects your impact.
Join us, and be part of something bigger—where your work drives real,
positive change in the world.


Job Overview:
We are looking for a highly skilled Full-Stack Developer with expertise in .NET Core, to develop and maintain scalable web applications and microservices. The ideal candidate will have strong problem-solving skills, experience in modern software development, and a passion for creating robust, high-performance applications.
Key Responsibilities:
Backend Development:
- Design, develop, and maintain microservices and APIs using.NET Core. Should have a good understanding of .NET Framework.
- Implement RESTful APIs, ensuring high performance and security.
- Optimize database queries and design schemas for SQL Server / Snowflake / MongoDB.
Software Architecture & DevOps:
- Design and implement scalable microservices architecture.
- Work with Docker, Kubernetes, and CI/CD pipelines for deployment and automation.
- Ensure best practices in security, scalability, and performance.
Collaboration & Agile Development:
- Work closely with UI/UX designers, backend engineers, and product managers.
- Participate in Agile/Scrum ceremonies, code reviews, and knowledge-sharing sessions.
- Write clean, maintainable, and well-documented code.
Required Skills & Qualifications:
- 7+ years of experience as a Full-Stack Developer.
- Strong experience in .NET Core, C#.
- Proficiency in React.js, JavaScript (ES6+), TypeScript.
- Experience with RESTful APIs, Microservices architecture.
- Knowledge of SQL / NoSQL databases (SQL Server, Snowflake, MongoDB).
- Experience with Git, CI/CD pipelines, Docker, and Kubernetes.
- Familiarity with Cloud services (Azure, AWS, or GCP) is a plus.
- Strong debugging and troubleshooting skills.
Nice-to-Have:
- Experience with GraphQL, gRPC, WebSockets.
- Exposure to serverless architecture and cloud-based solutions.
- Knowledge of authentication/authorization frameworks (OAuth, JWT, Identity Server).
- Experience with unit testing and integration testing.

Role Overview:
As a Backend Developer at LearnTube.ai, you will ship the backbone that powers 2.3 million learners in 64 countries—owning APIs that crunch 1 billion learning events & the AI that supports it with <200 ms latency.
Skip the wait and get noticed faster by completing our AI-powered screening. Click this link to start your quick interview. It only takes a few minutes and could be your shortcut to landing the job! - https://bit.ly/LT_Python
What You'll Do:
At LearnTube, we’re pushing the boundaries of Generative AI to revolutionise how the world learns. As a Backend Engineer, you will be building the backend for an AI system and working directly on AI. Your roles and responsibilities will include:
- Ship Micro-services – Build FastAPI services that handle ≈ 800 req/s today and will triple within a year (sub-200 ms p95).
- Power Real-Time Learning – Drive the quiz-scoring & AI-tutor engines that crunch millions of events daily.
- Design for Scale & Safety – Model data (Postgres, Mongo, Redis, SQS) and craft modular, secure back-end components from scratch.
- Deploy Globally – Roll out Dockerised services behind NGINX on AWS (EC2, S3, SQS) and GCP (GKE) via Kubernetes.
- Automate Releases – GitLab CI/CD + blue-green / canary = multiple safe prod deploys each week.
- Own Reliability – Instrument with Prometheus / Grafana, chase 99.9 % uptime, trim infra spend.
- Expose Gen-AI at Scale – Publish LLM inference & vector-search endpoints in partnership with the AI team.
- Ship Fast, Learn Fast – Work with founders, PMs, and designers in weekly ship rooms; take a feature from Figma to prod in < 2 weeks.
What makes you a great fit?
Must-Haves:
- 2+ yrs Python back-end experience (FastAPI)
- Strong with Docker & container orchestration
- Hands-on with GitLab CI/CD, AWS (EC2, S3, SQS) or GCP (GKE / Compute) in production
- SQL/NoSQL (Postgres, MongoDB) + You’ve built systems from scratch & have solid system-design fundamentals
Nice-to-Haves
- k8s at scale, Terraform,
- Experience with AI/ML inference services (LLMs, vector DBs)
- Go / Rust for high-perf services
- Observability: Prometheus, Grafana, OpenTelemetry
About Us:
At LearnTube, we’re on a mission to make learning accessible, affordable, and engaging for millions of learners globally. Using Generative AI, we transform scattered internet content into dynamic, goal-driven courses with:
- AI-powered tutors that teach live, solve doubts in real time, and provide instant feedback.
- Seamless delivery through WhatsApp, mobile apps, and the web, with over 1.4 million learners across 64 countries.
Meet the Founders:
LearnTube was founded by Shronit Ladhani and Gargi Ruparelia, who bring deep expertise in product development and ed-tech innovation. Shronit, a TEDx speaker, is an advocate for disrupting traditional learning, while Gargi’s focus on scalable AI solutions drives our mission to build an AI-first company that empowers learners to achieve career outcomes. We’re proud to be recognised by Google as a Top 20 AI Startup and are part of their 2024 Startups Accelerator: AI First Program, giving us access to cutting-edge technology, credits, and mentorship from industry leaders.
Why Work With Us?
At LearnTube, we believe in creating a work environment that’s as transformative as the products we build. Here’s why this role is an incredible opportunity:
- Cutting-Edge Technology: You’ll work on state-of-the-art generative AI applications, leveraging the latest advancements in LLMs, multimodal AI, and real-time systems.
- Autonomy and Ownership: Experience unparalleled flexibility and independence in a role where you’ll own high-impact projects from ideation to deployment.
- Rapid Growth: Accelerate your career by working on impactful projects that pack three years of learning and growth into one.
- Founder and Advisor Access: Collaborate directly with founders and industry experts, including the CTO of Inflection AI, to build transformative solutions.
- Team Culture: Join a close-knit team of high-performing engineers and innovators, where every voice matters, and Monday morning meetings are something to look forward to.
- Mission-Driven Impact: Be part of a company that’s redefining education for millions of learners and making AI accessible to everyone.
Job Title: QA Engineer
Location- Bangalore / Hybrid
Desired skills- Java / Node.js, Selenium, Rest Assured, Cucumber, Docker, Kubernetes, Nomad, Grafana, Kibana
Experience range- 5-8 years
Job Description:
- Develop automated tests and test frameworks to enhance software quality.
- Conduct functional and non-functional testing, including performance, security, resiliency, and chaos testing.
- Optimize and improve CI/CD pipelines for faster and more efficient deployments.
- Identify and implement quality processes to enhance engineering efficiency.
- Automate various aspects of the software development lifecycle.
- Review code and designs, providing constructive feedback for improvement.
- Continuously upskill and mentor team members to drive growth and excellence.
- Strong automation testing expertise for REST APIs, with coding proficiency in Node.js Or Java.
- Prior software development experience before transitioning into testing and automation is a plus.
- Hands-on experience with testing frameworks and CI/CD tools like Jenkins or GitHub Actions.
- Experience working with Docker, Kubernetes, and Nomad for containerized environments.
- Familiarity with cloud platforms, particularly AWS,
- Experience with observability and monitoring tools like Grafana and Kibana.
- Ensure comprehensive test coverage through automation and manual testing where necessary.
- Research automation tools and infrastructure for improvement
- Review plans and consult junior QAs
- Excellent problem-solving skills and attention to detail.
- Strong communication and interpersonal skills.
- Define E2E testing requirements (scenarios, conditions, testing types, metrics formonitoring)
- Execute E2E manual tests
- Automate E2E regression


About Us
DAITA is a German AI startup revolutionizing the global textile supply chain by digitizing factory-to-brand workflows. We are building cutting-edge AI-powered SaaS and Agentic Systems that automate order management, production tracking, and compliance — making the supply chain smarter, faster, and more transparent.
Fresh off a $500K pre-seed raise, our passionate team is on the ground in India, collaborating directly with factories and brands to build our MVP and create real-world impact. If you’re excited by the intersection of AI, SaaS, and supply chain innovation, join us to help reshape how textiles move from factory floors to global brands.
Role Overview
We’re seeking a versatile Full-Stack Engineer to join our growing engineering team. You’ll be instrumental in designing and building scalable, secure, and high-performance applications that power our AI-driven platform. Working closely with Founders, ML Engineers, and Pilot Customers, you’ll transform complex AI workflows into intuitive, production-ready features.
What You’ll Do
• Design, develop, and deploy backend services, APIs, and microservices powering our platform.
• Build responsive, user-friendly frontend applications tailored for factory and brand users.
• Integrate AI/ML models and agentic workflows into seamless production environments.
• Develop features supporting order parsing, supply chain tracking, compliance, and reporting.
• Collaborate cross-functionally to iterate rapidly, test with users, and deliver impactful releases.
• Optimize applications for performance, scalability, and cost-efficiency on cloud platforms.
• Establish and improve CI/CD pipelines, deployment processes, and engineering best practices.
• Write clear documentation and maintain clean, maintainable code.
Required Skills
• 3–5 years of professional Full-Stack development experience
• Strong backend skills with frameworks like Node.js, Python (FastAPI, Django), Go, or similar
• Frontend experience with React, Vue.js, Next.js, or similar modern frameworks
• Solid knowledge and experience with relational databases (PostgreSQL, MySQL) and NoSQL databases (MongoDB, Redis, Neon)
• Strong API design skills (REST mandatory; GraphQL a plus)
• Containerization expertise with Docker
• Container orchestration and management with Kubernetes (including experience with Helm charts, operators, or custom resource definitions)
• Cloud deployment and infrastructure experience on AWS, GCP or Azure
• Hands-on experience deploying AI/ML models in cloud-native environments (AWS, GCP or Azure) with scalable infrastructure and monitoring.
• Experience with managed AI/ML services like AWS SageMaker, GCP Vertex AI, Azure ML, Together.ai, or similar
• Experience with CI/CD pipelines and DevOps tools such as Jenkins, GitHub Actions, Terraform, Ansible, or ArgoCD
• Familiarity with monitoring, logging, and observability tools like Prometheus, Grafana, ELK stack (Elasticsearch, Logstash, Kibana), or Helicone
Nice-to-have
• Experience with TypeScript for full-stack AI SaaS development
• Use of modern UI frameworks and tooling like Tailwind CSS
• Familiarity with modern AI-first SaaS concepts viz. vector databases for fast ML data retrieval, prompt engineering for LLM integration, integrating with OpenRouter or similar LLM orchestration frameworks etc.
• Knowledge of MLOps tools like Kubeflow, MLflow, or Seldon for model lifecycle management.
• Background in building data pipelines, real-time analytics, and predictive modeling.
• Knowledge of AI-driven security tools and best practices for SaaS compliance.
• Proficiency in cloud automation, cost optimization, and DevOps for AI workflows.
• Ability to design and implement hyper-personalized, adaptive user experiences.
What We Value
• Ownership: You take full responsibility for your work and ship high-quality solutions quickly.
• Bias for Action: You’re pragmatic, proactive, and focused on delivering results.
• Clear Communication: You articulate ideas, challenges, and solutions effectively across teams.
• Collaborative Spirit: You thrive in a cross-functional, distributed team environment.
• Customer Focus: You build with empathy for end users and real-world usability.
• Curiosity & Adaptability: You embrace learning, experimentation, and pivoting when needed.
• Quality Mindset: You write clean, maintainable, and well-tested code.
Why Join DAITA?
• Be part of a mission-driven startup transforming a $1+ Trillion global industry.
• Work closely with founders and AI experts on cutting-edge technology.
• Directly impact real-world supply chains and sustainability.
• Grow your skills in AI, SaaS, and supply chain tech in a fast-paced environment.
DevOps Engineer
AiSensy
Gurugram, Haryana, India (On-site)
About AiSensy
AiSensy is a WhatsApp based Marketing & Engagement platform helping businesses like Adani, Delhi Transport Corporation, Yakult, Godrej, Aditya Birla Hindalco., Wipro, Asian Paints, India Today Group Skullcandy, Vivo, Physicswallah, Cosco grow their revenues via WhatsApp.
- Enabling 100,000+ Businesses with WhatsApp Engagement & Marketing
- 400Crores + WhatsApp Messages done between Businesses and Users via AiSensy per year
- Working with top brands like Delhi Transport Corporation, Vivo, Physicswallah & more
- High Impact as Businesses drive 25-80% Revenues using AiSensy Platform
- Mission-Driven and Growth Stage Startup backed by Marsshot.vc, Bluelotus.vc & 50+ Angel Investors
Now, we’re looking for a DevOps Engineer to help scale our infrastructure and optimize performance for millions of users. 🚀
What You’ll Do (Key Responsibilities)
🔹 CI/CD & Automation:
- Implement, manage, and optimize CI/CD pipelines using AWS CodePipeline, GitHub Actions, or Jenkins.
- Automate deployment processes to improve efficiency and reduce downtime.
🔹 Infrastructure Management:
- Use Terraform, Ansible, Chef, Puppet, or Pulumi to manage infrastructure as code.
- Deploy and maintain Dockerized applications on Kubernetes clusters for scalability.
🔹 Cloud & Security:
- Work extensively with AWS (Preferred) or other cloud platforms to build and maintain cloud infrastructure.
- Optimize cloud costs and ensure security best practices are in place.
🔹 Monitoring & Troubleshooting:
- Set up and manage monitoring tools like CloudWatch, Prometheus, Datadog, New Relic, or Grafana to track system performance and uptime.
- Proactively identify and resolve infrastructure-related issues.
🔹 Scripting & Automation:
- Use Python or Bash scripting to automate repetitive DevOps tasks.
- Build internal tools for system health monitoring, logging, and debugging.
What We’re Looking For (Must-Have Skills)
✅ Version Control: Proficiency in Git (GitLab / GitHub / Bitbucket)
✅ CI/CD Tools: Hands-on experience with AWS CodePipeline, GitHub Actions, or Jenkins
✅ Infrastructure as Code: Strong knowledge of Terraform, Ansible, Chef, or Pulumi
✅ Containerization & Orchestration: Experience with Docker & Kubernetes
✅ Cloud Expertise: Hands-on experience with AWS (Preferred) or other cloud providers
✅ Monitoring & Alerting: Familiarity with CloudWatch, Prometheus, Datadog, or Grafana
✅ Scripting Knowledge: Python or Bash for automation
Bonus Skills (Good to Have, Not Mandatory)
➕ AWS Certifications: Solutions Architect, DevOps Engineer, Security, Networking
➕ Experience with Microsoft/Linux/F5 Technologies
➕ Hands-on knowledge of Database servers

Position : Tech Lead – Fullstack Developer
Experience : 7 to 15 Years
Location : MG Road, Bengaluru (Hybrid – 3 Days in Office)
Notice Period : Immediate / Serving / 15 Days or Less
About the Opportunity :
We are hiring a Tech Lead – Fullstack Developer for a well-funded product startup building an enterprise-grade SaaS platform in the Cybersecurity domain.
The role involves designing and delivering scalable microservices and cloud-native applications in a high-performing, Agile engineering environment.
You'll work alongside industry veterans from billion-dollar digital firms, contributing to technical design, product architecture, and engineering best practices.
Mandatory Skills : Java, Spring Boot, ReactJS (or any modern JavaScript framework), RESTful APIs, PostgreSQL, Docker, Kubernetes, CI/CD, Hibernate/JPA, Multithreading, and Microservices architecture.
Role Highlights :
- Lead fullstack product development using Java (Spring Boot) and ReactJS (or similar frameworks).
- Design, develop, test, and deploy scalable microservices and RESTful APIs.
- Collaborate with Product, DevOps, and QA teams in a fast-paced Agile environment.
- Write modular, secure, and efficient code optimized for performance and maintainability.
- Mentor junior developers and influence architecture decisions across the team.
- Participate in all stages of the product lifecycle, from design to deployment.
- Create technical documentation, UML diagrams, and contribute to knowledge-sharing through blogs or whitepapers.
Key Skills Required :
- Strong expertise in Java (mandatory) and Spring Boot.
- Proficient in frontend development using ReactJS or similar frameworks.
- Hands-on experience building and consuming RESTful APIs.
- Solid knowledge of PostgreSQL, Hibernate/JPA, and transaction management.
- Familiarity with Docker, Kubernetes, and cloud platforms (Azure/GCP).
- Understanding of API Gateway, ACID properties, multithreading, and performance tuning.
- Experience with CI/CD tools (Jenkins, GitHub Actions, GitLab CI) and Agile methodologies.
- Strong debugging and profiling skills for performance bottlenecks.
Nice to Have :
- Experience with data integration tools (e.g., Pentaho, Apache NiFi).
- Exposure to the Healthcare or Cybersecurity domain.
- Familiarity with OpenAI APIs or real-time analytics tools.
- Willingness to contribute to internal documentation, blog posts, or whitepapers.
Perks & Benefits :
- Opportunity to build a product from scratch.
- Flat hierarchy and direct access to leadership.
- Strong focus on learning, mentorship, and technical innovation.
- Collaborative startup culture with long-term growth opportunities.
Interview Process :
- Technical Round – Technical Assessment
- Technical Interview – Core Development
- Advanced Technical Interview – Design & Problem Solving
- Final Round – CTO Discussion
About the Company
We are hiring for a fast-growing, well-funded product startup backed by a leadership team with a proven track record of building billion-dollar digital businesses. The company is focused on delivering enterprise-grade SaaS products in the Cybersecurity domain for B2B markets. You’ll be part of a passionate and dynamic engineering team building innovative solutions using modern tech stacks.
Key Responsibilities
- Design and develop scalable microservices using Java and Spring Boot
- Build and manage robust RESTful APIs
- Collaborate with cross-functional teams in an Agile setup
- Lead and mentor junior engineers, driving technical excellence
- Contribute to architecture discussions and code reviews
- Work with PostgreSQL, implement data integrity and consistency
- Deploy and manage services on cloud platforms like GCP or Azure
- Utilize Docker/Kubernetes for containerization and orchestration
Must-Have Skills
- Strong backend experience with Java, Spring Boot, REST APIs
- Proficiency in frontend development with React.js
- Experience with PostgreSQL and database optimization
- Hands-on with cloud platforms (GCP or Azure)
- Familiarity with Docker and Kubernetes
- Strong understanding of:
- API Gateways
- Hibernate & JPA
- Transaction management & ACID properties
- Multi-threading and context switching
Good to Have
- Experience in Cybersecurity or Healthcare domain
Exposure to CI/CD pipelines and DevOps practices
About the Role
We are looking for a highly motivated DevOps Engineer with a strong background in cloud technologies, big data ecosystems, and software development lifecycles to lead cross-functional teams in delivering high-impact projects. The ideal candidate will combine excellent project management skills with technical acumen in GCP, DevOps, and Python-based applications.
Key Responsibilities
- Lead end-to-end project planning, execution, and delivery, ensuring alignment
- Create and maintain project documentation including detailed timelines, sprint boards, risk logs, and weekly status reports.
- Facilitate Agile ceremonies: daily stand-ups, sprint planning, retrospectives, and backlog grooming.
- Actively manage risks, scope changes, resource allocation, and project dependencies to ensure delivery without disruptions.
- Ensure compliance with QA processes and security/compliance standards throughout the SDLC.
- Collaborate with stakeholders and senior leadership to communicate progress, blockers, and key milestones.
- Provide mentorship and support to cross-functional team members to drive continuous improvement and team performance.
- Coordinate with clients and act as a key point of contact for requirement gathering, updates, and escalations.
Required Skills & Experience
Cloud & DevOps
- Proficient in Google Cloud Platform (GCP) services: Compute, Storage, Networking, IAM.
- Hands-on experience with cloud deployments and infrastructure as code.
- Strong working knowledge of CI/CD pipelines, Docker, Kubernetes, and Terraform (or similar tools).
Big Data & Data Engineering
- Experience with large-scale data processing using tools like PySpark, Hadoop, Hive, HDFS, and Spark Streaming (preferred).
- Proven experience in managing and optimizing big data pipelines and ensuring high performance.
Programming & Frameworks
- Strong proficiency in Python with experience in Django (REST APIs, ORM, deployment workflows).
- Familiarity with Git and version control best practices.
- Basic knowledge of Linux administration and shell scripting.
Nice to Have
- Knowledge or prior experience in the Media & Advertising domain.
- Experience in client-facing roles and handling stakeholder communications.
- Proven ability to manage technical teams (5–6 members).
Why Join Us?
- Work on cutting-edge cloud and data engineering projects
- Collaborate with a talented, fast-paced team
- Flexible work setup and culture of ownership

Role overview:
As a founding senior software engineer, you will play a key role in shaping our AI-powered visual search engine for fashion and e-commerce. Responsibilities include solving complex deep-tech challenges to build scalable AI/ML solutions, leading backend development for performance and scalability, and architecting and integrating software aligned with product strategy and innovation goals. You will collaborate with cross-functional teams to address real consumer problems and build robust AI/ML pipelines to drive product innovation.
What we’re looking for:
- 3–5 years of Python experience (Golang is a plus), with expertise in concurrency, FastAPI, restful APIs, and microservices.
- proficiency in PostgreSQL/MongoDB, cloud platforms (AWS/GCP/Azure), and containerization tools like Docker/Kubernetes.
- Strong experience in asynchronous programming, CI/CD pipelines, and version control (git).
- Excellent problem-solving and communication skills are essential.
What we offer:
- Competitive salary and ESOPs, along with hackerhouse living: live and work with a Gen-Z team in a 7bhk house on MG Road, Gurgaon.
- Hands-on experience in shipping world-class products, professional development opportunities, flexible hours, and a collaborative, supportive culture.

As a senior software engineer (AI/ML), you will play a key role in shaping our AI-powered visual search engine for fashion and e-commerce. Responsibilities include solving complex deep-tech challenges to build scalable AI/ML solutions, leading backend development for performance and scalability, and architecting and integrating software aligned with product strategy and innovation goals. You will collaborate with cross-functional teams to address real consumer problems and build robust AI/ML pipelines to drive product innovation.
What we’re looking for:
- Design and deploy advanced machine learning models in computer vision, including object detection & similarity matching
- Implement scalable data pipelines, optimize models for performance and accuracy, and ensure they are production-ready with MLOps
- 3–5 years of Python experience (Golang is a plus), with expertise in concurrency, FastAPI, restful APIs, and microservices.
- proficiency in PostgreSQL/MongoDB, cloud platforms (AWS/GCP/Azure), and containerization tools like Docker/Kubernetes.
- Take part in code reviews, share knowledge, and lead by example to maintain high-quality engineering practices
- Strong experience in asynchronous programming, CI/CD pipelines, and version control (git). Excellent problem-solving and communication skills are essential.
What we offer:
- Competitive salary and ESOPs, along with HackerHouse living: live and work with a Gen Z team in a 7BHK house on MG Road, Gurgaon.
- hands-on experience in shipping world-class products, professional development opportunities, flexible hours, and a collaborative, supportive culture.
What does a successful Senior DevOps Engineer do at Fiserv?
This role’s focus will be on contributing and enhancing our DevOps environment within Issuer Solution group, where our cross functional Scrum teams are delivering solutions built on cutting-edge mobile technology and products. You will be expected to support across the wider business unit, leading DevOps practices and initiatives.
What will you do:
• Build, manage, and deploy CI/CD pipelines.
• DevOps Engineer - Helm Chart, Rundesk, Openshift
• Strive for continuous improvement and build continuous integration, continuous development, and constant deployment pipeline.
• Implementing various development, testing, automation tools, and IT infrastructure
• Optimize and automate release/development cycles and processes.
• Be part of and help promote our DevOps culture.
• Identify and implement continuous improvements to the development practice
What you must have:
• 3+ years of experience in devops with hands-on experience in the following:
- Writing automation scripts for deployments and housekeeping using shell scripts (bash) and ansible playbooks
- Building docker images and running/managing docker instances
- Building Jenkins pipelines using groovy scripts
- Working knowledge on kubernetes including application deployments, managing application configurations and persistence volumes
• Has good understanding on infrastructure as code
• Ability to write and update documentation
• Demonstrate a logical, process orientated approach to problems and troubleshooting
• Ability to collaborate with multi development teams
What you are preferred to have:
• 8+ years of development experience
• Jenkins administration experience
• Hands-on experience in building and deploying helm charts
Process Skills:
• Should have worked in Agile Project

Job Title: Java Full Stack Developer
Experience: 6+ Years
Locations: Bangalore, Mumbai, Pune, Gurgaon
Work Mode: Hybrid
Notice Period: Immediate Joiners Preferred / Candidates Who Have Completed Their Notice Period
About the Role
We are looking for a highly skilled and experienced Java Full Stack Developer with a strong command over backend technologies and modern frontend frameworks. The ideal candidate will have deep experience with Java, ReactJS, and DevOps tools like Jenkins, Docker, and basic Kubernetes knowledge. You’ll be contributing to complex software solutions across industries, collaborating with cross-functional teams, and deploying production-grade systems in a cloud-native, CI/CD-driven environment.
Key Responsibilities
- Design and develop scalable web applications using Java (Spring Boot) and ReactJS
- Collaborate with UX/UI designers and backend developers to implement robust, efficient front-end interfaces
- Develop and maintain CI/CD pipelines using Jenkins, ensuring high-quality software delivery
- Containerize applications using Docker and ensure smooth deployment and orchestration using Kubernetes (basic level)
- Write clean, modular, and testable code and participate in code reviews
- Troubleshoot and resolve performance, reliability, and functional issues in production
- Work in Agile teams and participate in daily stand-ups, sprint planning, and retrospective meetings
- Ensure all security, compliance, and performance standards are met in the development lifecycle
Mandatory Skills
- Backend: Java, Spring Boot
- Frontend: ReactJS
- DevOps Tools: Jenkins, Docker
- Containers & Orchestration: Basic knowledge of Kubernetes
- Strong understanding of RESTful services and APIs
- Familiarity with Git and version control workflows
- Good understanding of SDLC, Agile/Scrum methodologies
Profile: Senior Java Developer
🔷 Experience: 5+ Years
🔷 Location: Remote
🔷 Shift: Day Shift
(Only immediate joiners & candidates who have completed notice period)
✨ What we want:
✅ AWS cloud services (MANDATORY)
✅ Docker containerization (MANDATORY)
✅ Spring/Spring Boot framework
✅ RESTful API development
✅ Microservices architecture
✅ Database experience (SQL/NoSQL)
✅ Git version control & CI/CD
✅ Kubernetes orchestration
About Sun King
Sun King is the world’s leading off-grid solar energy company, delivering energy access to 1.8 billion people without reliable grid connections through innovative product design, fintech solutions, and field operations.
Key highlights:
- Connected over 20 million homes to solar power across Africa and Asia, adding 200,000 homes monthly.
- Affordable ‘pay-as-you-go’ financing model; after 1-2 years, customers own their solar equipment.
- Saved customers over $4 billion to date.
- Collect 650,000 daily payments via 28,000 field agents using mobile money systems.
- Products range from home lighting to high-energy appliances, with expansion into clean cooking, electric mobility, and entertainment.
With 2,800 staff across 12 countries, our team includes experts in various fields, all passionate about serving off-grid communities.
Diversity Commitment:
44% of our workforce are women, reflecting our commitment to gender diversity.
About the role:
Sun King is looking for a self-driven Infrastructure engineer, who is comfortable working in a fast-paced startup environment and balancing the needs of multiple development teams and systems. You will work on improving our current IAC, observability stack, and incident response processes. You will work with the data science, analytics, and engineering teams to build optimized CI/CD pipelines, scalable AWS infrastructure, and Kubernetes deployments.
What you would be expected to do:
- Work with engineering, automation, and data teams to work on various infrastructure requirements.
- Designing modular and efficient GitOps CI/CD pipelines, agnostic to the underlying platform.
- Managing AWS services for multiple teams.
- Managing custom data store deployments like sharded MongoDB clusters, Elasticsearch clusters, and upcoming services.
- Deployment and management of Kubernetes resources.
- Deployment and management of custom metrics exporters, trace data, custom application metrics, and designing dashboards, querying metrics from multiple resources, as an end-to-end observability stack solution.
- Set up incident response services and design effective processes.
- Deployment and management of critical platform services like OPA and Keycloak for IAM.
- Advocate best practices for high availability and scalability when designing AWS infrastructure, observability dashboards, implementing IAC, deploying to Kubernetes, and designing GitOps CI/CD pipelines.
You might be a strong candidate if you have/are:
- Hands-on experience with Docker or any other container runtime environment and Linux with the ability to perform basic administrative tasks.
- Experience working with web servers (nginx, apache) and cloud providers (preferably AWS).
- Hands-on scripting and automation experience (Python, Bash), experience debugging and troubleshooting Linux environments and cloud-native deployments.
- Experience building CI/CD pipelines, with familiarity with monitoring & alerting systems (Grafana, Prometheus, and exporters).
- Knowledge of web architecture, distributed systems, and single points of failure.
- Familiarity with cloud-native deployments and concepts like high availability, scalability, and bottleneck.
- Good networking fundamentals — SSH, DNS, TCP/IP, HTTP, SSL, load balancing, reverse proxies, and firewalls.
Good to have:
- Experience with backend development and setting up databases and performance tuning using parameter groups.
- Working experience in Kubernetes cluster administration and Kubernetes deployments.
- Experience working alongside sec ops engineers.
- Basic knowledge of Envoy, service mesh (Istio), and SRE concepts like distributed tracing.
- Setup and usage of open telemetry, central logging, and monitoring systems.
What Sun King offers:
- Professional growth in a dynamic, rapidly expanding, high-social-impact industry.
- An open-minded, collaborative culture made up of enthusiastic colleagues who are driven by the challenge of innovation towards profound impact on people and the planet.
- A truly multicultural experience: you will have the chance to work with and learn from people from different geographies, nationalities, and backgrounds.
- Structured, tailored learning and development programs that help you become a better leader, manager, and professional through the Sun King Center for Leadership.

About the Role
We are looking for a highly motivated Project Manager with a strong background in cloud technologies, big data ecosystems, and software development lifecycles to lead cross-functional teams in delivering high-impact projects. The ideal candidate will combine excellent project management skills with technical acumen in GCP, DevOps, and Python-based applications.
Key Responsibilities
- Lead end-to-end project planning, execution, and delivery, ensuring alignment with business goals and timelines.
- Create and maintain project documentation including detailed timelines, sprint boards, risk logs, and weekly status reports.
- Facilitate Agile ceremonies: daily stand-ups, sprint planning, retrospectives, and backlog grooming.
- Actively manage risks, scope changes, resource allocation, and project dependencies to ensure delivery without disruptions.
- Ensure compliance with QA processes and security/compliance standards throughout the SDLC.
- Collaborate with stakeholders and senior leadership to communicate progress, blockers, and key milestones.
- Provide mentorship and support to cross-functional team members to drive continuous improvement and team performance.
- Coordinate with clients and act as a key point of contact for requirement gathering, updates, and escalations.
Required Skills & Experience
Cloud & DevOps
- Proficient in Google Cloud Platform (GCP) services: Compute, Storage, Networking, IAM.
- Hands-on experience with cloud deployments and infrastructure as code.
- Strong working knowledge of CI/CD pipelines, Docker, Kubernetes, and Terraform (or similar tools).
Big Data & Data Engineering
- Experience with large-scale data processing using tools like PySpark, Hadoop, Hive, HDFS, and Spark Streaming (preferred).
- Proven experience in managing and optimizing big data pipelines and ensuring high performance.
Programming & Frameworks
- Strong proficiency in Python with experience in Django (REST APIs, ORM, deployment workflows).
- Familiarity with Git and version control best practices.
- Basic knowledge of Linux administration and shell scripting.
Nice to Have
- Knowledge or prior experience in the Media & Advertising domain.
- Experience in client-facing roles and handling stakeholder communications.
- Proven ability to manage technical teams (5–6 members).
Why Join Us?
- Work on cutting-edge cloud and data engineering projects
- Collaborate with a talented, fast-paced team
- Flexible work setup and culture of ownership
- Continuous learning and upskilling environment
- Inclusive health benefits included
About Company:
Auditoria is an AI-driven SaaS automation provider for corporate finance that automates back-office business processes involving tasks, analytics, and responses in Vendor Management, Accounts Payable, Accounts Receivable, and Planning. By leveraging natural language processing, artificial intelligence, and machine learning, Auditoria removes friction and repetition from mundane tasks
while automating complex functions and providing real-time visibility into cash performance. Corporate finance and accounting teams use Auditoria to accelerate business value while minimizing heavy IT
involvement, improving business resilience, lowering attrition, and accelerating business insights.
Founded in 2019 and backed by Venrock, Workday Ventures, Neotribe Ventures, Engineering Capital, and Firebolt Ventures, we build intelligent automation by combining fine-grained analytical
orchestration of a company's typical financial and audit workflows with conversational AI, delivering rapid value to the finance/audit back office.
In 2021, Auditoria earned industry recognition by being named to the Intelligent Apps Top 40 List, SSON's Shared Services & Outsourcing Impact Awards, the Constellation Research ShortList for AI-Driven Cognitive Applications, HFS Research Hot Vendors, 2021 CRN Emerging Vendors List, TiE50 Award, and the winner of the inaugural Pitch Event by Constellation Research.
The opportunity for you:
We are building an AI/ML-enabled SAAS solution to help manage the cash performance of enterprises. You would be working on solving complex problems in the FinTech space.
Responsibilities:
- Own the design and development of the core areas of Auditoria’s product, leveraging the latest tech stack hosted on AWS cloud.
- Collaborating across multiple teams/timezones to help deliver quality solutions as per roadmap.
- Partner with business teams to deliver incremental value to clients
- Champion Auditoria’s values and tech culture.
Requirements:
- Worked with many of the following: multi-tenant SaaS, CI/CD environments, monitoring tools, Kubernetes and containers, Istio, workflow engines, data stores (RDBMS, Cassandra, Neo4j), AWS services, integrations with enterprise systems (SSO, email, etc.).
- Experience with system design & data modeling; familiarity with RDBMS and Big Data
Job Description:
- Hands-on experience building applications leveraging AWS services like Step Functions, RDS, Cassandra, Kinesis, and ELK is a big plus.
- Fluent coding skills in Node.js.
- 10+ years of professional, hands-on experience developing and shipping complex software systems.
- Embraces startup setup, can work through unknowns, resource constraints & multiple priorities with creativity & resourcefulness.
- Experience with the Agile development process and zeal for engineering best practices.
- BS or MS in Computer Science or related Engineering degree. Preferably from IIT/NIT/BITS

Senior Software Engineer - Backend
A Senior Software Backend Engineer is responsible for designing, building, and maintaining the server-side
logic and infrastructure of web applications or software systems. They typically work closely with frontend
engineers, DevOps teams, and other stakeholders to ensure that the back-end services perform optimally and
meet business requirements. Below is an outline of a typical Senior Backend Developer job profile:
Key Responsibilities:
1. System Architecture & Design:
- Design scalable, high-performance backend services and APIs.
- Participate in the planning, design, and development of new features.
- Ensure that systems are designed with fault tolerance, security, and scalability in mind.
2. Development & Implementation:
- Write clean, maintainable, and efficient code.
- Implement server-side logic, databases, and data storage solutions.
- Work with technologies like REST, GraphQL, and other backend communication methods.
- Design and optimize database schemas, queries, and indexes.
3. Performance Optimization:
- Diagnose and fix performance bottlenecks.
- Optimize backend processes and database queries for speed and efficiency.
- Implement caching strategies and load balancing.
4. Security:
- Ensure the security of the backend systems by implementing secure coding practices.
- Protect against common security threats such as SQL injection, cross-site scripting (XSS), and others.
5. Collaboration & Leadership:
- Collaborate with frontend teams, product managers, and DevOps engineers.
- Mentor junior developers and guide them in best practices.
- Participate in code reviews and ensure that the development team follows consistent coding standards.
6. Testing & Debugging:
- Develop and run unit, integration, and performance tests to ensure code quality.
- Troubleshoot, debug, and upgrade existing systems.
7. Monitoring & Maintenance:
- Monitor system performance and take preventive measures to ensure uptime and reliability.
- Maintain technical documentation for reference and reporting.
- Stay updated on emerging technologies and incorporate them into the backend tech stack.
Required Skills:
1. Programming Languages:
- Expertise in one or more backend programming languages in the list Python, Java, Go, Rust.
2. Database Management:
- Strong understanding of both relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g.,
MongoDB, Redis).
- Knowledge of data modeling, query optimization, and database scaling strategies.
3. API Design & Development:
- Proficiency in designing and implementing RESTful and GraphQL APIs.
- Experience with microservices architecture.
- Good understanding of containers
4. Cloud & DevOps:
- Familiarity with cloud platforms like AWS, Azure, or Google Cloud.
- Understanding of DevOps principles, CI/CD pipelines, containerization (Docker), and orchestration
(Kubernetes).
5. Version Control:
- Proficiency with Git and branching strategies.
6. Testing & Debugging Tools:
- Familiarity with testing frameworks, debugging tools, and performance profiling.
7. Soft Skills:
- Strong problem-solving skills.
- Excellent communication and teamwork abilities.
- Leadership and mentorship qualities.
Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Software Engineering, or related field.
- 5+ years of experience in backend development or software engineering.
- Proven experience with system design, architecture, and high-scale application development.
Preferred Qualifications:
- Experience with distributed systems, event-driven architectures, and asynchronous processing.
- Familiarity with message queues (e.g., RabbitMQ, Kafka) and caching layers (e.g., Redis, Memcached).
- Knowledge of infrastructure as code (IaC) tools like Terraform or Ansible.
Tools & Technologies:
- Languages: Python, Java, Golang, Rust.
- Databases: PostgreSQL, MySQL, MongoDB, Redis, Cassandra.
- Frameworks: Django, Flask, Spring Boot, Go Micro.
- Cloud Providers: AWS, Azure, Google Cloud.
- Containerization: Docker, Kubernetes.
- CI/CD: Jenkins, GitLab CI, CircleCI.
This job profile will vary depending on the company and industry, but the core principles of designing,
developing, and maintaining back-end systems remain the same.
Minimum 5 years of experience in a customer-facing role such as pre-sales, solutions engineering or technical architecture.
- Exceptional communication and presentation skills.
- Proven ability in technical integrations and conducting POCs.
- Proficiency in coding with high-level programming languages (Java, Go, Python).
- Solid understanding of Monitoring, Observability, Log Management, SIEM.
- Background in Engineering/DevOps will be considered an advantage.
- Previous experience in Technical Sales of Log Analytics, Monitoring, APM, RUM, SIEM is desirable.
Technical Expertise :
- In-depth knowledge of Kubernetes, AWS, Azure, GCP, Docker, Prometheus, OpenTelemetry.
- Candidates should have hands-on experience and the ability to integrate these technologies into customer environments, providing tailored solutions that meet diverse operational requirements.
Requirements
- Bachelors/Masters in Computer Science or a related field
- 5-8 years of relevant experience
- Proven track record of Team Leading/Mentoring a team successfully.
- Experience with web technologies and microservices architecture both frontend and backend.
- Java, Spring framework, hibernate
- MySQL, Mongo, Solr, Redis,
- Kubernetes, Docker
- Strong understanding of Object-Oriented Programming, Data Structures, and Algorithms.
- Excellent teamwork skills, flexibility, and ability to handle multiple tasks.
- Experience with API Design, ability to architect and implement an intuitive customer and third-party integration story
- Ability to think and analyze both breadth-wise (client, server, DB, control flow) and depth-wise (threads, sessions, space-time complexity) while designing and implementing services
- Exceptional design and architectural skills
- Experience of cloud providers/platforms like GCP and AWS
Roles & Responsibilities
- Develop new user-facing features.
- Work alongside the product to understand our requirements, and design, develop and iterate, think through the complex architecture.
- Writing clean, reusable, high-quality, high-performance, maintainable code.
- Encourage innovation and efficiency improvements to ensure processes are productive.
- Ensure the training and mentoring of the team members.
- Ensure the technical feasibility of UI/UX designs and optimize applications for maximum speed.
- Research and apply new technologies, techniques, and best practices.
- Team mentorship and leadership.

Product company for financial operations automation platform

Mandatory Criteria :
- Candidate must have Strong hands-on experience with Kubernetes of atleast 2 years in production environments.
- Candidate should have Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI).
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Candidate should have strong Backend experience.
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
About the Role
We are looking for a highly skilled and motivated Cloud Backend Engineer with 4–7 years of experience, who has worked extensively on at least one major cloud platform (GCP, AWS, Azure, or OCI). Experience with multiple cloud providers is a strong plus. As a Senior Development Engineer, you will play a key role in designing, building, and scaling backend services and infrastructure on cloud-native platforms.
# Experience with Kubernetes is mandatory.
Key Responsibilities
- Design and develop scalable, reliable backend services and cloud-native applications.
- Build and manage RESTful APIs, microservices, and asynchronous data processing systems.
- Deploy and operate workloads on Kubernetes with best practices in availability, monitoring, and cost-efficiency.
- Implement and manage CI/CD pipelines and infrastructure automation.
- Collaborate with frontend, DevOps, and product teams in an agile environment.
- Ensure high code quality through testing, reviews, and documentation.
Required Skills
- Strong hands-on experience with Kubernetes of atleast 2 years in production environments (mandatory).
- Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI].
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Solid understanding of distributed systems, microservices, and cloud-native architecture.
- Experience with containerization using Docker and Kubernetes-native deployment workflows.
- Working knowledge of SQL and relational databases.
Preferred Qualifications
- Experience working across multiple cloud platforms.
- Familiarity with infrastructure-as-code tools like Terraform or CloudFormation.
- Exposure to monitoring, logging, and observability stacks (e.g., Prometheus, Grafana, Cloud Monitoring).
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
Nice to Have
- Knowledge of NoSQL databases or event-driven/message-based architectures.
- Experience with serverless services, managed data pipelines, or data lake platforms.
We are looking for an experienced GCP Cloud Engineer to design, implement, and manage cloud-based solutions on Google Cloud Platform (GCP). The ideal candidate should have expertise in GKE (Google Kubernetes Engine), Cloud Run, Cloud Loadbalancer, Cloud function, Azure DevOps, and Terraform, with a strong focus on automation, security, and scalability.
You will work closely with development, operations, and security teams to ensure robust cloud infrastructure and CI/CD pipelines while optimizing performance and cost.
Key Responsibilities:
1. Cloud Infrastructure Design & Management
· Architect, deploy, and maintain GCP cloud resources via terraform/other automation.
· Implement Google Cloud Storage, Cloud SQL, file store, for data storage and processing needs.
· Manage and configure Cloud Load Balancers (HTTP(S), TCP/UDP, and SSL Proxy) for high availability and scalability.
· Optimize resource allocation, monitoring, and cost efficiency across GCP environments.
2. Kubernetes & Container Orchestration
· Deploy, manage, and optimize workloads on Google Kubernetes Engine (GKE).
· Work with Helm charts, Istio, and service meshes for microservices deployments.
· Automate scaling, rolling updates, and zero-downtime deployments.
3. Serverless & Compute Services
· Deploy and manage applications on Cloud Run and Cloud Functions for scalable, serverless workloads.
· Optimize containerized applications running on Cloud Run for cost efficiency and performance.
4. CI/CD & DevOps Automation
· Design, implement, and manage CI/CD pipelines using Azure DevOps.
· Automate infrastructure deployment using Terraform, Bash and Power shell scripting
· Integrate security and compliance checks into the DevOps workflow (DevSecOps).
Required Skills & Qualifications:
✔ Experience: 8+ years in Cloud Engineering, with a focus on GCP.
✔ Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions).
✔ Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm.
✔ DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation.
✔ Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources.
✔ Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation.
✔ Security & Compliance: Knowledge of cloud security principles, IAM, and compliance standards.

Role Overview:
As a Backend Developer at LearnTube.ai, you will ship the backbone that powers 2.3 million learners in 64 countries—owning APIs that crunch 1 billion learning events & the AI that supports it with <200 ms latency.
Skip the wait and get noticed faster by completing our AI-powered screening. Click this link to start your quick interview. It only takes a few minutes and could be your shortcut to landing the job! -https://bit.ly/LT_Python
What You'll Do:
At LearnTube, we’re pushing the boundaries of Generative AI to revolutionize how the world learns. As a Backend Engineer, your roles and responsibilities will include:
- Ship Micro-services – Build FastAPI services that handle ≈ 800 req/s today and will triple within a year (sub-200 ms p95).
- Power Real-Time Learning – Drive the quiz-scoring & AI-tutor engines that crunch millions of events daily.
- Design for Scale & Safety – Model data (Postgres, Mongo, Redis, SQS) and craft modular, secure back-end components from scratch.
- Deploy Globally – Roll out Dockerised services behind NGINX on AWS (EC2, S3, SQS) and GCP (GKE) via Kubernetes.
- Automate Releases – GitLab CI/CD + blue-green / canary = multiple safe prod deploys each week.
- Own Reliability – Instrument with Prometheus / Grafana, chase 99.9 % uptime, trim infra spend.
- Expose Gen-AI at Scale – Publish LLM inference & vector-search endpoints in partnership with the AI team.
- Ship Fast, Learn Fast – Work with founders, PMs, and designers in weekly ship rooms; take a feature from Figma to prod in < 2 weeks.
What makes you a great fit?
Must-Haves:
- 2+ yrs Python back-end experience (FastAPI)
- Strong with Docker & container orchestration
- Hands-on with GitLab CI/CD, AWS (EC2, S3, SQS) or GCP (GKE / Compute) in production
- SQL/NoSQL (Postgres, MongoDB) + You’ve built systems from scratch & have solid system-design fundamentals
Nice-to-Haves
- k8s at scale, Terraform,
- Experience with AI/ML inference services (LLMs, vector DBs)
- Go / Rust for high-perf services
- Observability: Prometheus, Grafana, OpenTelemetry
About Us:
At LearnTube, we’re on a mission to make learning accessible, affordable, and engaging for millions of learners globally. Using Generative AI, we transform scattered internet content into dynamic, goal-driven courses with:
- AI-powered tutors that teach live, solve doubts in real time, and provide instant feedback.
- Seamless delivery through WhatsApp, mobile apps, and the web, with over 1.4 million learners across 64 countries.
Meet the Founders:
LearnTube was founded by Shronit Ladhani and Gargi Ruparelia, who bring deep expertise in product development and ed-tech innovation. Shronit, a TEDx speaker, is an advocate for disrupting traditional learning, while Gargi’s focus on scalable AI solutions drives our mission to build an AI-first company that empowers learners to achieve career outcomes. We’re proud to be recognised by Google as a Top 20 AI Startup and are part of their 2024 Startups Accelerator: AI First Program, giving us access to cutting-edge technology, credits, and mentorship from industry leaders.
Why Work With Us?
At LearnTube, we believe in creating a work environment that’s as transformative as the products we build. Here’s why this role is an incredible opportunity:
- Cutting-Edge Technology: You’ll work on state-of-the-art generative AI applications, leveraging the latest advancements in LLMs, multimodal AI, and real-time systems.
- Autonomy and Ownership: Experience unparalleled flexibility and independence in a role where you’ll own high-impact projects from ideation to deployment.
- Rapid Growth: Accelerate your career by working on impactful projects that pack three years of learning and growth into one.
- Founder and Advisor Access: Collaborate directly with founders and industry experts, including the CTO of Inflection AI, to build transformative solutions.
- Team Culture: Join a close-knit team of high-performing engineers and innovators, where every voice matters, and Monday morning meetings are something to look forward to.
- Mission-Driven Impact: Be part of a company that’s redefining education for millions of learners and making AI accessible to everyone.
Who we are
At CoinCROWD, were building the next-gen wallet for real-world crypto utility. Our flagship product, CROWD Wallet, is secure, intuitive, gasless, and designed to bring digital currencies into everyday spending from a coffee shop to cross-border payments.
Were redefining the wallet experience for everyday users, combining the best of Web3 + AI to create a secure, scalable, and delightful platform
Were more than just a blockchain company, were an AI-native, crypto-forward startup. We ship fast, think long, and believe in building agentic, self-healing infrastructure that can scale across geographies and blockchains. If that excites you, lets talk.
What Youll Be Doing :
As the DevOps Lead at CoinCROWD, youll own our infrastructure from end to end, designing, deploying, and scaling secure systems to support blockchain transactions, AI agents, and token operations across global users.
You will :
- Lead the CI/CD, infra automation, observability, and multi-region deployments of CoinCROWD products.
- Manage cloud and container infrastructure using GCP, Docker, Kubernetes, Terraform.
- Deploy and maintain scalable, secure blockchain infrastructure using QuickNode, Alchemy, Web3Auth, and other Web3 APIs.
- Implement infrastructure-level AI agents or scripts for auto-scaling, failure prediction, anomaly detection, and alert management (using LangChain, LLMs, or tools like n8n).
- Ensure 99.99% uptime for wallet systems, APIs, and smart contract layers.
- Build and optimize observability across on-chain/off-chain systems using tools like Prometheus,
Grafana, Sentry, Loki, ELK Stack.
- Create auto-healing, self-monitoring pipelines that reduce human ops time via Agentic AI workflows.
- Collaborate with engineering and security teams on smart contract deployment pipelines, token rollouts, and app store release automation.
Agentic Ops : What it means
- Use GPT-based agents to auto-document infra changes or failure logs.
- Run LangChain agents that triage alerts, perform log analysis, or suggest infra optimizations.
- Build CI/CD workflows that self-update or auto-tune based on system usage.
- Integrate AI to detect abnormal wallet behaviors, fraud attempts, or suspicious traffic spikes
What Were Looking For :
- 5 to 10 years of DevOps/SRE experience, with at least 2 to 3 years in Web3, fintech, or high-scale infra.
- Deep expertise with Docker, Kubernetes, Helm, and cloud providers (GCP preferred).
- Hands-on with Terraform, Ansible, GitHub Actions, Jenkins, or similar IAC and pipeline tools.
- Experience maintaining or scaling blockchain infra (EVM nodes, RPC endpoints, APIs).
- Understanding of smart contract CI/CD, token lifecycle (ICO, vesting, etc.), and wallet integrations.
- Familiarity with AI DevOps tools, or interest in building LLM-enhanced internal tooling.
- Strong grip on security best practices, key management, and secrets infrastructure (Vault, SOPS, AWS KMS).
Bonus Points :
- You've built or run infra for a token launch, DEX, or high-TPS crypto wallet.
- You've deployed or automated a blockchain node network at scale.
- You've used AI/LLMs to write ops scripts, manage logs, or analyze incidents.
- You've worked with systems handling real-money movement with tight uptime and security requirements.
Why Join CoinCROWD :
- Equity-first model: Build real value as we scale.
- Be the architect of infrastructure that supports millions of real-world crypto transactions.
- Build AI-powered ops that scale without a 24/7 pager culture
- Work remotely with passionate people who ship fast and iterate faster.
- Be part of one of the most ambitious crossovers of AI + Web3 in 2025.
Role: Sr. Java Developer
Experience: 6+ Years
Location: Bangalore (Whitefield)
Work Mode: Hybrid (2-3 days WFO)
Shift Timing: Regular Morning Shift
About the Role:
We are looking for a seasoned Java Developer with 6+ years of experience to join our growing engineering team. The ideal candidate should have strong expertise in Java, Spring Boot, Microservices, and cloud-based deployment using AWS or DevOps tools. This is a hybrid role based out of our Whitefield, Bangalore location.
Key Responsibilities:
- Participate in agile development processes and scrum ceremonies.
- Translate business requirements into scalable and maintainable technical solutions.
- Design and develop applications using Java, Spring Boot, and Microservices architecture.
- Ensure robust and reliable code through full-scale unit testing and TDD/BDD practices.
- Contribute to CI/CD pipeline setup and cloud deployments.
- Work independently and as an individual contributor on complex features.
- Troubleshoot production issues and optimize application performance.
Mandatory Skills:
- Strong Core Java and Spring Boot expertise.
- Proficiency in AWS or DevOps (Docker & Kubernetes).
- Experience with relational and/or non-relational databases (SQL, NoSQL).
- Sound understanding of Microservices architecture and RESTful APIs.
- Containerization experience using Docker and orchestration via Kubernetes.
- Familiarity with Linux-based development environments.
- Exposure to modern SDLC tools – Maven, Git, Jenkins, etc.
- Good understanding of CI/CD pipelines and cloud-based deployment.
Soft Skills:
- Self-driven, proactive, and an individual contributor.
- Strong problem-solving and analytical skills.
- Excellent communication and interpersonal abilities.
- Able to plan, prioritize, and manage tasks independently.
Nice-to-Have Skills:
- Exposure to frontend technologies like Angular, JavaScript, HTML5, and CSS.
Looking for Fresher developers
Responsibilities:
- Implement integrations requested by customers
- Deploy updates and fixes
- Provide Level 2 technical support
- Build tools to reduce occurrences of errors and improve customer experience
- Develop software to integrate with internal back-end systems
- Perform root cause analysis for production errors
- Investigate and resolve technical issues
- Develop scripts to automate visualization
- Design procedures for system troubleshooting and maintenance
Requirements and skill:
Knowledge in DevOps Engineer or similar software engineering role
Good knowledge of Terraform, Kubernetes
Working knowledge on AWS, Google Cloud
You can directly contact me on nine three one six one two zero one three two

Job Title : Senior Consultant (Java / NodeJS + Temporal)
Experience : 5 to 12 Years
Location : Bengaluru, Chennai, Hyderabad, Pune, Mumbai, Gurugram, Coimbatore
Work Mode : Remote (Must be open to travel for occasional team meetups)
Notice Period : Immediate Joiners or Serving Notice
Interview Process :
- R1 : Tech Interview (60 mins)
- R2 : Technical Interview
- R3 : (Optional) Interview with Client
Job Summary :
We are seeking a Senior Backend Consultant with strong hands-on expertise in Temporal (BPM/Workflow Engine) and either Node.js or Java.
The ideal candidate will have experience in designing and developing microservices and process-driven applications, as well as orchestrating complex workflows using Temporal.io.
You will work on high-scale systems, collaborating closely with cross-functional teams.
Mandatory Skills :
Temporal.io, Node.js (or Java), React.js, Keycloak IAM, PostgreSQL, Terraform, Kubernetes, Azure, Jest, OpenAPI
Key Responsibilities :
- Design and implement scalable backend services using Node.js or Java.
- Build and manage complex workflow orchestrations using Temporal.io.
- Integrate with IAM solutions like Keycloak for role-based access control.
- Work with React (v17+), TypeScript, and component-driven frontend design.
- Use PostgreSQL for structured data persistence and optimized queries.
- Manage infrastructure using Terraform and orchestrate via Kubernetes.
- Leverage Azure Services like Blob Storage, API Gateway, and AKS.
- Write and maintain API documentation using Swagger/Postman/Insomnia.
- Conduct unit and integration testing using Jest.
- Participate in code reviews and contribute to architectural decisions.
Must-Have Skills :
- Temporal.io – BPMN modeling, external task workers, Operate, Tasklist
- Node.js + TypeScript (preferred) or strong Java experience
- React.js (v17+) and component-driven UI development
- Keycloak IAM, PostgreSQL, and modern API design
- Infrastructure automation with Terraform, Kubernetes
- Experience in using GitFlow, OpenAPI, Jest for testing
Nice-to-Have Skills :
- Blockchain integration experience for secure KYC/identity flows
- Custom Camunda Connectors or exporter plugin development
- CI/CD experience using Azure DevOps or GitHub Actions
- Identity-based task completion authorization enforcement
About the Role
We are looking for a skilled Backend Engineer with strong experience in building scalable microservices, integrating with distributed data systems, and deploying web APIs that serve UI applications in the cloud. You’ll work on high-performance systems involving Kafka, DynamoDB, Redis, and other modern backend technologies.
Responsibilities
- Design, develop, and deploy backend microservices and APIs that power UI applications.
- Implement event-driven architectures using Apache Kafka or similar messaging platforms.
- Build scalable and highly available systems using NoSQL databases (e.g., DynamoDB, MongoDB).
- Optimize backend systems using caching layers like Redis to enhance performance.
- Ensure seamless deployment and operation of services in cloud environments (AWS, GCP, or Azure).
- Write clean, maintainable, and well-tested code; contribute to code reviews and architecture discussions.
- Collaborate closely with frontend, DevOps, and product teams to deliver integrated solutions.
- Monitor and troubleshoot production issues and participate in on-call rotations as needed.
Required Qualifications
- 3–7 years of professional experience in backend development.
- Strong programming skills in one or more languages: Java, Python, Go, Node.js.
- Hands-on experience with microservices architecture and API design (REST/gRPC).
- Practical experience with Kafka, RabbitMQ, or other event streaming/message queue systems.
- Solid knowledge of NoSQL databases, especially DynamoDB or equivalents.
- Experience using Redis or Memcached for caching or pub/sub mechanisms.
- Proficiency with cloud platforms (preferably AWS – e.g., Lambda, ECS, EKS, API Gateway).
- Familiarity with Docker, Kubernetes, and CI/CD pipelines.
Job Description
What does a successful Senior DevOps Engineer do at Fiserv?
This role’s focus will be on contributing and enhancing our DevOps environment within Issuer Solution group, where our cross functional Scrum teams are delivering solutions built on cutting-edge mobile technology and products. You will be expected to support across the wider business unit, leading DevOps practices and initiatives.
What will you do:
• Build, manage, and deploy CI/CD pipelines.
• DevOps Engineer - Helm Chart, Rundesk, Openshift
• Strive for continuous improvement and build continuous integration, continuous development, and constant deployment pipeline.
• Implementing various development, testing, automation tools, and IT infrastructure
• Optimize and automate release/development cycles and processes.
• Be part of and help promote our DevOps culture.
• Identify and implement continuous improvements to the development practice
What you must have:
• 3+ years of experience in devops with hands-on experience in the following:
- Writing automation scripts for deployments and housekeeping using shell scripts (bash) and ansible playbooks
- Building docker images and running/managing docker instances
- Building Jenkins pipelines using groovy scripts
- Working knowledge on kubernetes including application deployments, managing application configurations and persistence volumes
• Has good understanding on infrastructure as code
• Ability to write and update documentation
• Demonstrate a logical, process orientated approach to problems and troubleshooting
• Ability to collaborate with multi development teams
What you are preferred to have:
• 8+ years of development experience
• Jenkins administration experience
• Hands-on experience in building and deploying helm charts
Process Skills:
• Should have worked in Agile Project
Behavioral Skills :
• Good Communication skills
Skills
PRIMARY COMPETENCY : Cloud Infra PRIMARY SKILL : DevOps PRIMARY SKILL PERCENTAGE : 100
About the Role
We are looking for a DevOps Engineer to build and maintain scalable, secure, and high-
performance infrastructure for our next-generation healthcare platform. You will be
responsible for automation, CI/CD pipelines, cloud infrastructure, and system reliability,
ensuring seamless deployment and operations.
Responsibilities
1. Infrastructure & Cloud Management
• Design, deploy, and manage cloud-based infrastructure (AWS, Azure, GCP)
• Implement containerization (Docker, Kubernetes) and microservices orchestration
• Optimize infrastructure cost, scalability, and performance
2. CI/CD & Automation
• Build and maintain CI/CD pipelines for automated deployments
• Automate infrastructure provisioning using Terraform, Ansible, or CloudFormation
• Implement GitOps practices for streamlined deployments
3. Security & Compliance
• Ensure adherence to ABDM, HIPAA, GDPR, and healthcare security standards
• Implement role-based access controls, encryption, and network security best
practices
• Conduct Vulnerability Assessment & Penetration Testing (VAPT) and compliance
audits
4. Monitoring & Incident Management
• Set up monitoring, logging, and alerting systems (Prometheus, Grafana, ELK,
Datadog, etc.)
• Optimize system reliability and automate incident response mechanisms
• Improve MTTR (Mean Time to Recovery) and system uptime KPIs
5. Collaboration & Process Improvement
• Work closely with development and QA teams to streamline deployments
• Improve DevSecOps practices and cloud security policies
• Participate in architecture discussions and performance tuning
Required Skills & Qualifications
• 2+ years of experience in DevOps, cloud infrastructure, and automation
• Hands-on experience with AWS and Kubernetes
• Proficiency in Docker and CI/CD tools (Jenkins, GitHub Actions, ArgoCD, etc.)
• Experience with Terraform, Ansible, or CloudFormation
• Strong knowledge of Linux, shell scripting, and networking
• Experience with cloud security, monitoring, and logging solutions
Nice to Have
• Experience in healthcare or other regulated industries
• Familiarity with serverless architectures and AI-driven infrastructure automation
• Knowledge of big data pipelines and analytics workflows
What You'll Gain
• Opportunity to build and scale a mission-critical healthcare infrastructure
• Work in a fast-paced startup environment with cutting-edge technologies
• Growth potential into Lead DevOps Engineer or Cloud Architect roles
Job Role : Azure DevSecOps Engineer (Security-Focused)
Experience : 12 to 18 Years
Location : Preferably Delhi NCR (Hybrid); Remote possible with 1–2 office visits per quarter (Gurgaon)
Joining Timeline : Max 45 days (Buyout option available)
Work Mode : Full-time | 5 Days Working
About the Role :
We are looking for a highly experienced Azure DevSecOps Engineer with a strong focus on cloud security practices.
This role is 60–70% security-driven, involving threat modeling, secure cloud architecture, and infrastructure security on Azure using Terraform.
Key Responsibilities :
- Architect and maintain secure, scalable Azure cloud infrastructure using Terraform.
- Implement security best practices : IAM, threat modeling, network security, data protection, and compliance (e.g., GDPR).
- Build CI/CD pipelines and automate deployments using Azure DevOps, Jenkins, Prometheus.
- Monitor, analyze, and proactively improve security posture.
- Collaborate with global teams to ensure secure design, development, and operations.
- Stay updated on cloud security trends and lead mitigation efforts.
Mandatory Skills :
Azure, Terraform, DevSecOps, Cloud Security, Threat Modelling, IAM, CI/CD (Azure DevOps), Docker, Kubernetes, Prometheus, Infrastructure as Code (IaC), Compliance Frameworks (GDPR)
Preferred Certifications :
Certified DevSecOps Professional (CDP), Microsoft Azure Certifications

Job Role : DevOps Engineer (Python + DevOps)
Experience : 4 to 10 Years
Location : Hyderabad
Work Mode : Hybrid
Mandatory Skills : Python, Ansible, Docker, Kubernetes, CI/CD, Cloud (AWS/Azure/GCP)
Job Description :
We are looking for a skilled DevOps Engineer with expertise in Python, Ansible, Docker, and Kubernetes.
The ideal candidate will have hands-on experience automating deployments, managing containerized applications, and ensuring infrastructure reliability.
Key Responsibilities :
- Design and manage containerization and orchestration using Docker & Kubernetes.
- Automate deployments and infrastructure tasks using Ansible & Python.
- Build and maintain CI/CD pipelines for streamlined software delivery.
- Collaborate with development teams to integrate DevOps best practices.
- Monitor, troubleshoot, and optimize system performance.
- Enforce security best practices in containerized environments.
- Provide operational support and contribute to continuous improvements.
Required Qualifications :
- Bachelor’s in Computer Science/IT or related field.
- 4+ years of DevOps experience.
- Proficiency in Python and Ansible.
- Expertise in Docker and Kubernetes.
- Hands-on experience with CI/CD tools and pipelines.
- Experience with at least one cloud provider (AWS, Azure, or GCP).
- Strong analytical, communication, and collaboration skills.
Preferred Qualifications :
- Experience with Infrastructure-as-Code tools like Terraform.
- Familiarity with monitoring/logging tools like Prometheus, Grafana, or ELK.
- Understanding of Agile/Scrum practices.

You will:
- Collaborate with the I-Stem Voice AI team and CEO to design, build and ship new agent capabilities
- Develop, test and refine end-to-end voice agent models (ASR, NLU, dialog management, TTS)
- Stress-test agents in noisy, real-world scenarios and iterate for improved robustness and low latency
- Research and prototype cutting-edge techniques (e.g. robust speech recognition, adaptive language understanding)
- Partner with backend and frontend engineers to seamlessly integrate AI components into live voice products
- Monitor agent performance in production, analyze failure cases, and drive continuous improvement
- Occasionally demo our Voice AI solutions at industry events and user forums
You are:
- An AI/Software Engineer with hands-on experience in speech-centric ML (ASR, NLU or TTS)
- Skilled in building and tuning transformer-based speech models and handling real-time audio pipelines
- Obsessed with reliability: you design experiments to push agents to their limits and root-cause every error
- A clear thinker who deconstructs complex voice interactions from first principles
- Passionate about making voice technology inclusive and accessible for diverse users
- Comfortable moving fast in a small team, yet dogged about code quality, testing and reproducibility
We are seeking an experienced and passionate Cloud and DevOps Trainer to join our training and development team. The trainer will be responsible for delivering high-quality, hands-on training in Cloud technologies (such as AWS, Azure, or GCP) and DevOps tools and practices to students or working professionals.


Skill Sets:
- Expertise in ML/DL, model lifecycle management, and MLOps (MLflow, Kubeflow)
- Proficiency in Python, TensorFlow, PyTorch, Scikit-learn, and Hugging Face models
- Strong experience in NLP, fine-tuning transformer models, and dataset preparation
- Hands-on with cloud platforms (AWS, GCP, Azure) and scalable ML deployment (Sagemaker, Vertex AI)
- Experience in containerization (Docker, Kubernetes) and CI/CD pipelines
- Knowledge of distributed computing (Spark, Ray), vector databases (FAISS, Milvus), and model optimization (quantization, pruning)
- Familiarity with model evaluation, hyperparameter tuning, and model monitoring for drift detection
Roles and Responsibilities:
- Design and implement end-to-end ML pipelines from data ingestion to production
- Develop, fine-tune, and optimize ML models, ensuring high performance and scalability
- Compare and evaluate models using key metrics (F1-score, AUC-ROC, BLEU etc)
- Automate model retraining, monitoring, and drift detection
- Collaborate with engineering teams for seamless ML integration
- Mentor junior team members and enforce best practices
What you’ll do
- Tame data → pull, clean, and shape structured & unstructured data.
- Orchestrate pipelines → Airflow / Step Functions / ADF… your call.
- Ship models → build, tune, and push to prod on SageMaker, Azure ML, or Vertex AI.
- Scale → Spark / Databricks for the heavy lifting.
- Automate everything → Docker, Kubernetes, CI/CD, MLFlow, Seldon, Kubeflow.
- Pair up → work with engineers, architects, and business folks to solve real problems, fast.
What you bring
- 3+ yrs hands-on MLOps (4-5 yrs total software experience).
- Proven chops on one hyperscaler (AWS, Azure, or GCP).
- Confidence with Databricks / Spark, Python, SQL, TensorFlow / PyTorch / Scikit-learn.
- You debug Kubernetes in your sleep and treat Dockerfiles like breathing.
- You prototype with open-source first, choose the right tool, then make it scale.
- Sharp mind, low ego, bias for action.
Nice-to-haves
- Sagemaker, Azure ML, or Vertex AI in production.
- Love for clean code, clear docs, and crisp PRs.

Job Title : Python Data Engineer
Experience : 4+ Years
Location : Bangalore / Hyderabad (On-site)
Job Summary :
We are seeking a skilled Python Data Engineer to work on cloud-native data platforms and backend services.
The role involves building scalable APIs, working with diverse data systems, and deploying containerized services using modern cloud infrastructure.
Mandatory Skills : Python, AWS, RESTful APIs, Microservices, SQL/PostgreSQL/NoSQL, Docker, Kubernetes, CI/CD (Jenkins/GitLab CI/AWS CodePipeline)
Key Responsibilities :
- Design, develop, and maintain backend systems using Python.
- Build and manage RESTful APIs and microservices architectures.
- Work extensively with AWS cloud services for deployment and data storage.
- Implement and manage SQL, PostgreSQL, and NoSQL databases.
- Containerize applications using Docker and orchestrate with Kubernetes.
- Set up and maintain CI/CD pipelines using Jenkins, GitLab CI, or AWS CodePipeline.
- Collaborate with teams to ensure scalable and reliable software delivery.
- Troubleshoot and optimize application performance.
Must-Have Skills :
- 4+ years of hands-on experience in Python backend development.
- Strong experience with AWS cloud infrastructure.
- Proficiency in building microservices and APIs.
- Good knowledge of relational and NoSQL databases.
- Experience with Docker and Kubernetes.
- Familiarity with CI/CD tools and DevOps processes.
- Strong problem-solving and collaboration skills.

A leader in telecom, fintech, AI-led marketing automation.

We are looking for a talented MERN Developer with expertise in MongoDB/MySQL, Kubernetes, Python, ETL, Hadoop, and Spark. The ideal candidate will design, develop, and optimize scalable applications while ensuring efficient source code management and implementing Non-Functional Requirements (NFRs).
Key Responsibilities:
- Develop and maintain robust applications using MERN Stack (MongoDB, Express.js, React.js, Node.js).
- Design efficient database architectures (MongoDB/MySQL) for scalable data handling.
- Implement and manage Kubernetes-based deployment strategies for containerized applications.
- Ensure compliance with Non-Functional Requirements (NFRs), including source code management, development tools, and security best practices.
- Develop and integrate Python-based functionalities for data processing and automation.
- Work with ETL pipelines for smooth data transformations.
- Leverage Hadoop and Spark for processing and optimizing large-scale data operations.
- Collaborate with solution architects, DevOps teams, and data engineers to enhance system performance.
- Conduct code reviews, troubleshooting, and performance optimization to ensure seamless application functionality.
Required Skills & Qualifications:
- Proficiency in MERN Stack (MongoDB, Express.js, React.js, Node.js).
- Strong understanding of database technologies (MongoDB/MySQL).
- Experience working with Kubernetes for container orchestration.
- Hands-on knowledge of Non-Functional Requirements (NFRs) in application development.
- Expertise in Python, ETL pipelines, and big data technologies (Hadoop, Spark).
- Strong problem-solving and debugging skills.
- Knowledge of microservices architecture and cloud computing frameworks.
Preferred Qualifications:
- Certifications in cloud computing, Kubernetes, or database management.
- Experience in DevOps, CI/CD automation, and infrastructure management.
- Understanding of security best practices in application development.
Minimum requirements
5+ years of industry software engineering experience (does not include internships nor includes co-ops)
Strong coding skills in any programming language (we understand new languages can be learned on the job so our interview process is language agnostic)
Strong collaboration skills, can work across workstreams within your team and contribute to your peers’ success
Have the ability to thrive on a high level of autonomy, responsibility, and think of yourself as entrepreneurial
Interest in working as a generalist across varying technologies and stacks to solve problems and delight both internal and external users
Preferred Qualifications
Experience with large-scale financial tracking systems
Good understanding and practical knowledge in cloud based services (e.g. gRPC, GraphQL, Docker/Kubernetes, cloud services such as AWS, etc.)
Job Summary:
We are seeking a skilled DevOps Engineer to design, implement, and manage CI/CD pipelines, containerized environments, and infrastructure automation. The ideal candidate should have hands-on experience with ArgoCD, Kubernetes, and Docker, along with a deep understanding of cloud platforms and deployment strategies.
Key Responsibilities:
- CI/CD Implementation: Develop, maintain, and optimize CI/CD pipelines using ArgoCD, GitOps, and other automation tools.
- Container Orchestration: Deploy, manage, and troubleshoot containerized applications using Kubernetes and Docker.
- Infrastructure as Code (IaC): Automate infrastructure provisioning with Terraform, Helm, or Ansible.
- Monitoring & Logging: Implement and maintain observability tools like Prometheus, Grafana, ELK, or Loki.
- Security & Compliance: Ensure best security practices in containerized and cloud-native environments.
- Cloud & Automation: Manage cloud infrastructure on AWS, Azure, or GCP with automated deployments.
- Collaboration: Work closely with development teams to optimize deployments and performance.
Required Skills & Qualifications:
- Experience: 5+ years in DevOps, Site Reliability Engineering (SRE), or Infrastructure Engineering.
- Tools & Tech: Strong knowledge of ArgoCD, Kubernetes, Docker, Helm, Terraform, and CI/CD pipelines.
- Cloud Platforms: Experience with AWS, GCP, or Azure.
- Programming & Scripting: Proficiency in Python, Bash, or Go.
- Version Control: Hands-on with Git and GitOps workflows.
- Networking & Security: Knowledge of ingress controllers, service mesh (Istio/Linkerd), and container security best practices.
Nice to Have:
- Experience with Kubernetes Operators, Kustomize, or FluxCD.
- Exposure to serverless architectures and multi-cloud deployments.
- Certifications in CKA, AWS DevOps, or similar.
Job Title : Java Backend Developer
Experience : 3 to 6 Years
Locations : Bangalore / Gurgaon (Hybrid – 3 Days Work From Office)
Shift Timings : 11:00 AM – 8:00 PM IST
Notice Period : Immediate to 15 Days Only
Job Description :
We are looking for experienced Java Backend Developers with strong expertise in building scalable microservices-based architectures. The ideal candidate should have hands-on experience with Spring Boot, containerized deployments, and DevOps tools.
✅ Must-Have Skills :
- Java – Strong programming skills in core Java.
- Spring Boot (2.x / 3.x) – Deep understanding of microservices architecture and patterns.
- Microservices – Design and implementation experience.
- Kubernetes – Experience deploying and managing microservices
- Jenkins & Maven – Build and CI/CD pipeline experience
- PostgreSQL – Experience with relational database management
✨ Good-to-Have Skills :
- Git – Source control management
- CI/CD Pipeline Tools – End-to-end pipeline automation
- Cloud & DevOps Knowledge – Experience with cloud-based deployments
Job description
● Design effective, scalable architectures on top of cloud technologies such as AWS and Kubernetes
● Mentor other software engineers, including actively participating in peer code and architecture review
● Participate in all parts of the development lifecycle from design to coding to deployment to maintenance and operations
● Kickstart new ideas, build proof of concepts and jumpstart newly funded projects
● Demonstrate ability to work independently with minimal supervision
● Embed with other engineering teams on challenging initiatives and time sensitive projects
● Collaborate with other engineering teams on challenging initiatives and time sensitive projects
Education and Experience
● BS degree in Computer Science or related technical field or equivalent practical experience.
● 9+ years of professional software development experience focused on payments and/or billing and customer accounts. Worked with worldwide payments, billing systems, PCI Compliance & payment gateways.
Technical and Functional
● Extensive knowledge of micro service development using Spring, Spring Boot, Java - built on top of Kubernetes and public cloud computing such as AWS, Lambda, S3.
● Experience with relational databases (MySQL, DB2 or Oracle) and NoSQL databases
● Experience with unit testing and test driven development
Technologies at Constant Contact
Working on the Constant Contact platform provides our engineers with an opportunity to produce high impact work inside of our multifaceted platform (Email, Social, SMS, E-Commerce, CRM, Customer Data Platform, MLBased Recommendations & Insights, and more).
As a member of our team, you'll be utilizing the latest technologies and frameworks (React/SPA, JavaScript/TypeScript, Swift, Kotlin, GraphQL, etc) and deploying code to our cloud-first microservice infrastructure (declarative CI/CD, GitOps managed kubernetes) with regular opportunities to level up your skills.
● Past experience of working with and integrating payment gateways and processors, online payment methods, and billing systems.
● Familiar with integrating Stripe/Plaid/PayPal/Adyen/Cybersource or similar systems along with PCI compliance.
● International software development and payments experience is a plus.
● Knowledge of DevOps and CI/CD, automated test and build tools ( Jenkins & Gradle/Maven)
● Experience integrating with sales tax engines is a plus.
● Familiar with tools like Splunk, New relic or similar tools like datadog, elastic elk, amazon
cloudwatch.
● Good to have - Experience with React, Backbone, Marionette or other front end frameworks.
Cultural
● Strong verbal and written communication skills.
● Flexible attitude and willingness to frequently move between different teams, software architectures and priorities.
● Desire to collaborate with our other product teams to think strategically about how to solve problems.
Our team
● We focus on cross-functional team collaboration where engineers, product managers, and designers all work together to solve customer problems and build exciting features.
● We love new ideas and are eager to see what your experiences can bring to help influence our technical and product vision.
● Collaborate/Overlap with the teams in Eastern Standard Time (EST), USA.
About PGAGI:
At PGAGI, we believe in a future where AI and human intelligence coexist in harmony, creating a world that is smarter, faster, and better. We are not just building AI; we are shaping a future where AI is a fundamental and positive force for businesses, societies, and the planet.
Position Overview:
PGAGI Consultancy Pvt. Ltd. is seeking a proactive and motivated DevOps Intern with around 3-6 months of hands-on experience to support our AI model deployment and infrastructure initiatives. This role is ideal for someone looking to deepen their expertise in DevOps practices tailored to AI/ML environments, including CI/CD automation, cloud infrastructure, containerization, and monitoring.
Key Responsibilities:
AI Model Deployment & Integration
- Assist in containerizing and deploying AI/ML models into production using Docker.
- Support integration of models into existing systems and APIs.
Infrastructure Management
- Help manage cloud and on-premise environments to ensure scalability and consistency.
- Work with Kubernetes for orchestration and environment scaling.
CI/CD Pipeline Automation
- Collaborate on building and maintaining automated CI/CD pipelines (e.g., GitHub Actions, Jenkins).
- Implement basic automated testing and rollback mechanisms.
Hosting & Web Environment Management
- Assist in managing hosting platforms, web servers, and CDN configurations.
- Support DNS, load balancer setups, and ensure high availability of web services.
Monitoring, Logging & Optimization
- Set up and maintain monitoring/logging tools like Prometheus and Grafana.
- Participate in troubleshooting and resolving performance bottlenecks.
Security & Compliance
- Apply basic DevSecOps practices including security scans and access control implementations.
- Follow security and compliance checklists under supervision.
Cost & Resource Management
- Monitor resource usage and suggest cost optimization strategies in cloud environments.
Documentation
- Maintain accurate documentation for deployment processes and incident responses.
Continuous Learning & Innovation
- Suggest improvements to workflows and tools.
- Stay updated with the latest DevOps and AI infrastructure trends.
Requirements:
- Around 6 months of experience in a DevOps or related technical role (internship or professional).
- Basic understanding of Docker, Kubernetes, and CI/CD tools like GitHub Actions or Jenkins.
- Familiarity with cloud platforms (AWS, GCP, or Azure) and monitoring tools (e.g., Prometheus, Grafana).
- Exposure to scripting languages (e.g., Bash, Python) is a plus.
- Strong problem-solving skills and eagerness to learn.
- Good communication and documentation abilities.
Compensation
- Joining Bonus: INR 2,500 one-time bonus upon joining.
- Monthly Stipend: Base stipend of INR 8,000 per month, with the potential to increase up to INR 20,000 based on performance evaluations.
- Performance-Based Pay Scale: Eligibility for monthly performance-based bonuses, rewarding exceptional project contributions and teamwork.
- Additional Benefits: Access to professional development opportunities, including workshops, tech talks, and mentoring sessions.
Ready to kick-start your DevOps journey in a dynamic AI-driven environment? Apply now
#Devops #Docker #Kubernetes #DevOpsIntern