50+ Kubernetes Jobs in India
Apply to 50+ Kubernetes Jobs on CutShort.io. Find your next job, effortlessly. Browse Kubernetes Jobs and apply today!

Minimum 5 years of experience in a customer-facing role such as pre-sales, solutions engineering or technical architecture.
- Exceptional communication and presentation skills.
- Proven ability in technical integrations and conducting POCs.
- Proficiency in coding with high-level programming languages (Java, Go, Python).
- Solid understanding of Monitoring, Observability, Log Management, SIEM.
- Background in Engineering/DevOps will be considered an advantage.
- Previous experience in Technical Sales of Log Analytics, Monitoring, APM, RUM, SIEM is desirable.
Technical Expertise :
- In-depth knowledge of Kubernetes, AWS, Azure, GCP, Docker, Prometheus, OpenTelemetry.
- Candidates should have hands-on experience and the ability to integrate these technologies into customer environments, providing tailored solutions that meet diverse operational requirements.
Job description :
Job Title: Java Full Stack Developer - Azure
Experience Level: 6–10 Years
Job Summary:
We are looking for a highly capable and motivated Full Stack Developer with strong expertise in Java (Spring Boot), React.js, Angular, and solid hands-on experience with Microsoft Azure cloud services. The role involves end-to-end application development—frontend to backend—and deploying scalable, secure, and resilient applications on Azure.
Key Responsibilities:
- Develop scalable, full-stack web applications using Java (Spring Boot) on the backend and React.js/Angular on the frontend.
- Create and consume RESTful APIs and integrate them with modern frontend UIs.
- Build responsive and dynamic user interfaces using React and Angular, as per project needs.
- Deploy and manage applications on Microsoft Azure using services like App Services, Azure SQL, Azure Functions, Azure Storage, etc.
- Containerize applications using Docker, and deploy them using Azure Kubernetes Service (AKS).
- Automate infrastructure provisioning using ARM templates or Bicep, and implement CI/CD pipelines with Azure DevOps.
- Ensure security best practices across applications using OAuth 2.0, JWT, and Azure Identity services.
- Collaborate with cross-functional teams in Agile/Scrum methodologies.
- Participate in code reviews, unit testing, and troubleshooting in a cloud-native environment.
Required Skills:
Backend Development:
- Strong experience with Java 8+, Spring Boot, Spring MVC, RESTful services
- ORM tools: JPA/Hibernate
- Experience with SQL and Azure SQL Database
Frontend Development:
- Hands-on experience with React.js and Angular 8+
- Proficient in JavaScript, TypeScript, HTML5, CSS3, and responsive design
- Familiarity with Redux, RxJS, or similar state management libraries
Cloud & DevOps (Azure):
- Deep understanding of Microsoft Azure ecosystem:
- App Services, Azure Functions, Azure SQL, Azure Blob Storage
- Azure Kubernetes Service (AKS) and Container Registry (ACR)
- Azure Key Vault, Azure Monitor, Application Insights
- CI/CD with Azure DevOps, GitHub Actions, or Jenkins
- Docker-based application packaging and deployment
Preferred Skills:
- Experience with micro-services architecture
- NoSQL databases such as Cosmos DB or MongoDB
- Basic scripting knowledge (PowerShell or Bash)
- Exposure to event-driven patterns using Azure Event Hub or Service Bus
Educational qualification:
B.E/B.Tech/MCA

Product company for financial operations automation platform

Mandatory Criteria
- Candidate must have Strong hands-on experience with Kubernetes of at least 2 years in production environments.
- Candidate should have Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI).
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Candidate should have strong Backend experience.
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
About the Role
We are looking for a highly skilled and motivated Cloud Backend Engineer with 4–7 years of experience, who has worked extensively on at least one major cloud platform (GCP, AWS, Azure, or OCI). Experience with multiple cloud providers is a strong plus. As a Senior Development Engineer, you will play a key role in designing, building, and scaling backend services and infrastructure on cloud-native platforms.
# Experience with Kubernetes is mandatory.
Key Responsibilities
- Design and develop scalable, reliable backend services and cloud-native applications.
- Build and manage RESTful APIs, microservices, and asynchronous data processing systems.
- Deploy and operate workloads on Kubernetes with best practices in availability, monitoring, and cost-efficiency.
- Implement and manage CI/CD pipelines and infrastructure automation.
- Collaborate with frontend, DevOps, and product teams in an agile environment.
- Ensure high code quality through testing, reviews, and documentation.
Required Skills
- Strong hands-on experience with Kubernetes of atleast 2 years in production environments (mandatory).
- Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI].
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Solid understanding of distributed systems, microservices, and cloud-native architecture.
- Experience with containerization using Docker and Kubernetes-native deployment workflows.
- Working knowledge of SQL and relational databases.
Preferred Qualifications
- Experience working across multiple cloud platforms.
- Familiarity with infrastructure-as-code tools like Terraform or CloudFormation.
- Exposure to monitoring, logging, and observability stacks (e.g., Prometheus, Grafana, Cloud Monitoring).
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
Nice to Have
- Knowledge of NoSQL databases or event-driven/message-based architectures.
- Experience with serverless services, managed data pipelines, or data lake platforms.
Requirements
- Bachelors/Masters in Computer Science or a related field
- 5-8 years of relevant experience
- Proven track record of Team Leading/Mentoring a team successfully.
- Experience with web technologies and microservices architecture both frontend and backend.
- Java, Spring framework, hibernate
- MySQL, Mongo, Solr, Redis,
- Kubernetes, Docker
- Strong understanding of Object-Oriented Programming, Data Structures, and Algorithms.
- Excellent teamwork skills, flexibility, and ability to handle multiple tasks.
- Experience with API Design, ability to architect and implement an intuitive customer and third-party integration story
- Ability to think and analyze both breadth-wise (client, server, DB, control flow) and depth-wise (threads, sessions, space-time complexity) while designing and implementing services
- Exceptional design and architectural skills
- Experience of cloud providers/platforms like GCP and AWS
Roles & Responsibilities
- Develop new user-facing features.
- Work alongside the product to understand our requirements, and design, develop and iterate, think through the complex architecture.
- Writing clean, reusable, high-quality, high-performance, maintainable code.
- Encourage innovation and efficiency improvements to ensure processes are productive.
- Ensure the training and mentoring of the team members.
- Ensure the technical feasibility of UI/UX designs and optimize applications for maximum speed.
- Research and apply new technologies, techniques, and best practices.
- Team mentorship and leadership.

Product company for financial operations automation platform

Mandatory Criteria :
- Candidate must have Strong hands-on experience with Kubernetes of atleast 2 years in production environments.
- Candidate should have Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI).
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Candidate should have strong Backend experience.
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
About the Role
We are looking for a highly skilled and motivated Cloud Backend Engineer with 4–7 years of experience, who has worked extensively on at least one major cloud platform (GCP, AWS, Azure, or OCI). Experience with multiple cloud providers is a strong plus. As a Senior Development Engineer, you will play a key role in designing, building, and scaling backend services and infrastructure on cloud-native platforms.
# Experience with Kubernetes is mandatory.
Key Responsibilities
- Design and develop scalable, reliable backend services and cloud-native applications.
- Build and manage RESTful APIs, microservices, and asynchronous data processing systems.
- Deploy and operate workloads on Kubernetes with best practices in availability, monitoring, and cost-efficiency.
- Implement and manage CI/CD pipelines and infrastructure automation.
- Collaborate with frontend, DevOps, and product teams in an agile environment.
- Ensure high code quality through testing, reviews, and documentation.
Required Skills
- Strong hands-on experience with Kubernetes of atleast 2 years in production environments (mandatory).
- Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI].
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Solid understanding of distributed systems, microservices, and cloud-native architecture.
- Experience with containerization using Docker and Kubernetes-native deployment workflows.
- Working knowledge of SQL and relational databases.
Preferred Qualifications
- Experience working across multiple cloud platforms.
- Familiarity with infrastructure-as-code tools like Terraform or CloudFormation.
- Exposure to monitoring, logging, and observability stacks (e.g., Prometheus, Grafana, Cloud Monitoring).
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
Nice to Have
- Knowledge of NoSQL databases or event-driven/message-based architectures.
- Experience with serverless services, managed data pipelines, or data lake platforms.
Job description
Job Title: Cloud Migration Consultant – (AWS to Azure)
Experience: 4+ years in application assessment and migration
About the Role
We’re looking for a Cloud Migration Consultant with hands-on experience assessing and migrating complex applications to Azure. You'll work closely with Microsoft business units, participating in Intake & Assessment and Planning & Design phases, creating migration artifacts, and leading client interactions. You’ll also support application modernization efforts in Azure, with exposure to AWS as needed.
Key Responsibilities
- Assess application readiness and document architecture, dependencies, and migration strategy.
- Conduct interviews with stakeholders and generate discovery insights using tools like Azure Migrate, CloudockIt, PowerShell.
- Create architecture diagrams, migration playbooks, and maintain Azure DevOps boards.
- Set up applications both on-premises and in cloud environments (primarily Azure).
- Support proof-of-concepts (PoCs) and advise on migration options.
- Collaborate with application, database, and infrastructure teams to enable smooth transition to migration factory teams.
- Track progress, blockers, and risks, reporting timely status to project leadership.
Required Skills
- 4+ years of experience in cloud migration and assessment
- Strong expertise in Azure IaaS/PaaS (VMs, App Services, ADF, etc.)
- Familiarity with AWS IaaS/PaaS (EC2, RDS, Glue, S3)
- Experience with Java (SpringBoot)/C#, .Net/Python, Angular/React.js, REST APIs
- Working knowledge of Kafka, Docker/Kubernetes, Azure DevOps
- Network infrastructure understanding (VNets, NSGs, Firewalls, WAFs)
- IAM knowledge: OAuth, SAML, Okta/SiteMinder
- Experience with Big Data tools like Databricks, Hadoop, Oracle, DocumentDB
Preferred Qualifications
- Azure or AWS certifications
- Prior experience with enterprise cloud migrations (especially in Microsoft ecosystem)
- Excellent communication and stakeholder management skills
Educational qualification:
B.E/B.Tech/MCA
Experience :
4+ Years
Key Responsibilities
- Assess application readiness and document architecture, dependencies, and migration strategy.
- Conduct interviews with stakeholders and generate discovery insights using tools like Azure Migrate, CloudockIt, PowerShell.
- Create architecture diagrams, migration playbooks, and maintain Azure DevOps boards.
- Set up applications both on-premises and in cloud environments (primarily Azure).
- Support proof-of-concepts (PoCs) and advise on migration options.
- Collaborate with application, database, and infrastructure teams to enable smooth transition to migration factory teams.
- Track progress, blockers, and risks, reporting timely status to project leadership.
Required Skills
- 4+ years of experience in cloud migration and assessment
- Strong expertise in Azure IaaS/PaaS (VMs, App Services, ADF, etc.)
- Familiarity with AWS IaaS/PaaS (EC2, RDS, Glue, S3)
- Experience with Java (SpringBoot)/C#, .Net/Python, Angular/React.js, REST APIs
- Working knowledge of Kafka, Docker/Kubernetes, Azure DevOps
- Network infrastructure understanding (VNets, NSGs, Firewalls, WAFs)
- IAM knowledge: OAuth, SAML, Okta/SiteMinder
- Experience with Big Data tools like Databricks, Hadoop, Oracle, DocumentDB
Preferred Qualifications
- Azure or AWS certifications
- Prior experience with enterprise cloud migrations (especially in Microsoft ecosystem)
- Excellent communication and stakeholder management skills
We are looking for an experienced GCP Cloud Engineer to design, implement, and manage cloud-based solutions on Google Cloud Platform (GCP). The ideal candidate should have expertise in GKE (Google Kubernetes Engine), Cloud Run, Cloud Loadbalancer, Cloud function, Azure DevOps, and Terraform, with a strong focus on automation, security, and scalability.
You will work closely with development, operations, and security teams to ensure robust cloud infrastructure and CI/CD pipelines while optimizing performance and cost.
Key Responsibilities:
1. Cloud Infrastructure Design & Management
· Architect, deploy, and maintain GCP cloud resources via terraform/other automation.
· Implement Google Cloud Storage, Cloud SQL, file store, for data storage and processing needs.
· Manage and configure Cloud Load Balancers (HTTP(S), TCP/UDP, and SSL Proxy) for high availability and scalability.
· Optimize resource allocation, monitoring, and cost efficiency across GCP environments.
2. Kubernetes & Container Orchestration
· Deploy, manage, and optimize workloads on Google Kubernetes Engine (GKE).
· Work with Helm charts, Istio, and service meshes for microservices deployments.
· Automate scaling, rolling updates, and zero-downtime deployments.
3. Serverless & Compute Services
· Deploy and manage applications on Cloud Run and Cloud Functions for scalable, serverless workloads.
· Optimize containerized applications running on Cloud Run for cost efficiency and performance.
4. CI/CD & DevOps Automation
· Design, implement, and manage CI/CD pipelines using Azure DevOps.
· Automate infrastructure deployment using Terraform, Bash and Power shell scripting
· Integrate security and compliance checks into the DevOps workflow (DevSecOps).
Required Skills & Qualifications:
✔ Experience: 8+ years in Cloud Engineering, with a focus on GCP.
✔ Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions).
✔ Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm.
✔ DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation.
✔ Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources.
✔ Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation.
✔ Security & Compliance: Knowledge of cloud security principles, IAM, and compliance standards.

Role Overview:
As a Backend Developer at LearnTube.ai, you will ship the backbone that powers 2.3 million learners in 64 countries—owning APIs that crunch 1 billion learning events & the AI that supports it with <200 ms latency.
What You'll Do:
At LearnTube, we’re pushing the boundaries of Generative AI to revolutionize how the world learns. As a Backend Engineer, your roles and responsibilities will include:
- Ship Micro-services – Build FastAPI services that handle ≈ 800 req/s today and will triple within a year (sub-200 ms p95).
- Power Real-Time Learning – Drive the quiz-scoring & AI-tutor engines that crunch millions of events daily.
- Design for Scale & Safety – Model data (Postgres, Mongo, Redis, SQS) and craft modular, secure back-end components from scratch.
- Deploy Globally – Roll out Dockerised services behind NGINX on AWS (EC2, S3, SQS) and GCP (GKE) via Kubernetes.
- Automate Releases – GitLab CI/CD + blue-green / canary = multiple safe prod deploys each week.
- Own Reliability – Instrument with Prometheus / Grafana, chase 99.9 % uptime, trim infra spend.
- Expose Gen-AI at Scale – Publish LLM inference & vector-search endpoints in partnership with the AI team.
- Ship Fast, Learn Fast – Work with founders, PMs, and designers in weekly ship rooms; take a feature from Figma to prod in < 2 weeks.
What makes you a great fit?
Must-Haves:
- 2+ yrs Python back-end experience (FastAPI)
- Strong with Docker & container orchestration
- Hands-on with GitLab CI/CD, AWS (EC2, S3, SQS) or GCP (GKE / Compute) in production
- SQL/NoSQL (Postgres, MongoDB) + You’ve built systems from scratch & have solid system-design fundamentals
Nice-to-Haves
- k8s at scale, Terraform,
- Experience with AI/ML inference services (LLMs, vector DBs)
- Go / Rust for high-perf services
- Observability: Prometheus, Grafana, OpenTelemetry
About Us:
At LearnTube, we’re on a mission to make learning accessible, affordable, and engaging for millions of learners globally. Using Generative AI, we transform scattered internet content into dynamic, goal-driven courses with:
- AI-powered tutors that teach live, solve doubts in real time, and provide instant feedback.
- Seamless delivery through WhatsApp, mobile apps, and the web, with over 1.4 million learners across 64 countries.
Meet the Founders:
LearnTube was founded by Shronit Ladhani and Gargi Ruparelia, who bring deep expertise in product development and ed-tech innovation. Shronit, a TEDx speaker, is an advocate for disrupting traditional learning, while Gargi’s focus on scalable AI solutions drives our mission to build an AI-first company that empowers learners to achieve career outcomes. We’re proud to be recognised by Google as a Top 20 AI Startup and are part of their 2024 Startups Accelerator: AI First Program, giving us access to cutting-edge technology, credits, and mentorship from industry leaders.
Why Work With Us?
At LearnTube, we believe in creating a work environment that’s as transformative as the products we build. Here’s why this role is an incredible opportunity:
- Cutting-Edge Technology: You’ll work on state-of-the-art generative AI applications, leveraging the latest advancements in LLMs, multimodal AI, and real-time systems.
- Autonomy and Ownership: Experience unparalleled flexibility and independence in a role where you’ll own high-impact projects from ideation to deployment.
- Rapid Growth: Accelerate your career by working on impactful projects that pack three years of learning and growth into one.
- Founder and Advisor Access: Collaborate directly with founders and industry experts, including the CTO of Inflection AI, to build transformative solutions.
- Team Culture: Join a close-knit team of high-performing engineers and innovators, where every voice matters, and Monday morning meetings are something to look forward to.
- Mission-Driven Impact: Be part of a company that’s redefining education for millions of learners and making AI accessible to everyone.
Who we are
At CoinCROWD, were building the next-gen wallet for real-world crypto utility. Our flagship product, CROWD Wallet, is secure, intuitive, gasless, and designed to bring digital currencies into everyday spending from a coffee shop to cross-border payments.
Were redefining the wallet experience for everyday users, combining the best of Web3 + AI to create a secure, scalable, and delightful platform
Were more than just a blockchain company, were an AI-native, crypto-forward startup. We ship fast, think long, and believe in building agentic, self-healing infrastructure that can scale across geographies and blockchains. If that excites you, lets talk.
What Youll Be Doing :
As the DevOps Lead at CoinCROWD, youll own our infrastructure from end to end, designing, deploying, and scaling secure systems to support blockchain transactions, AI agents, and token operations across global users.
You will :
- Lead the CI/CD, infra automation, observability, and multi-region deployments of CoinCROWD products.
- Manage cloud and container infrastructure using GCP, Docker, Kubernetes, Terraform.
- Deploy and maintain scalable, secure blockchain infrastructure using QuickNode, Alchemy, Web3Auth, and other Web3 APIs.
- Implement infrastructure-level AI agents or scripts for auto-scaling, failure prediction, anomaly detection, and alert management (using LangChain, LLMs, or tools like n8n).
- Ensure 99.99% uptime for wallet systems, APIs, and smart contract layers.
- Build and optimize observability across on-chain/off-chain systems using tools like Prometheus,
Grafana, Sentry, Loki, ELK Stack.
- Create auto-healing, self-monitoring pipelines that reduce human ops time via Agentic AI workflows.
- Collaborate with engineering and security teams on smart contract deployment pipelines, token rollouts, and app store release automation.
Agentic Ops : What it means
- Use GPT-based agents to auto-document infra changes or failure logs.
- Run LangChain agents that triage alerts, perform log analysis, or suggest infra optimizations.
- Build CI/CD workflows that self-update or auto-tune based on system usage.
- Integrate AI to detect abnormal wallet behaviors, fraud attempts, or suspicious traffic spikes
What Were Looking For :
- 5 to 10 years of DevOps/SRE experience, with at least 2 to 3 years in Web3, fintech, or high-scale infra.
- Deep expertise with Docker, Kubernetes, Helm, and cloud providers (GCP preferred).
- Hands-on with Terraform, Ansible, GitHub Actions, Jenkins, or similar IAC and pipeline tools.
- Experience maintaining or scaling blockchain infra (EVM nodes, RPC endpoints, APIs).
- Understanding of smart contract CI/CD, token lifecycle (ICO, vesting, etc.), and wallet integrations.
- Familiarity with AI DevOps tools, or interest in building LLM-enhanced internal tooling.
- Strong grip on security best practices, key management, and secrets infrastructure (Vault, SOPS, AWS KMS).
Bonus Points :
- You've built or run infra for a token launch, DEX, or high-TPS crypto wallet.
- You've deployed or automated a blockchain node network at scale.
- You've used AI/LLMs to write ops scripts, manage logs, or analyze incidents.
- You've worked with systems handling real-money movement with tight uptime and security requirements.
Why Join CoinCROWD :
- Equity-first model: Build real value as we scale.
- Be the architect of infrastructure that supports millions of real-world crypto transactions.
- Build AI-powered ops that scale without a 24/7 pager culture
- Work remotely with passionate people who ship fast and iterate faster.
- Be part of one of the most ambitious crossovers of AI + Web3 in 2025.
Job Title: QA Engineer
Location- Bangalore / Hybrid
Desired skills- Java / Node.js, Docker, Kubernetes, Nomad, Grafana, Kibana
Exp range- 5-8 yrs
Job Description:
- Develop automated tests and test frameworks to enhance software quality.
- Conduct functional and non-functional testing, including performance, security, resiliency, and chaos testing.
- Optimize and improve CI/CD pipelines for faster and more efficient deployments.
- Identify and implement quality processes to enhance engineering efficiency.
- Automate various aspects of the software development lifecycle.
- Review code and designs, providing constructive feedback for improvement.
- Continuously upskill and mentor team members to drive growth and excellence.
- Strong automation testing expertise for REST APIs, with coding proficiency in Node.js Or Java.
- Prior software development experience before transitioning into testing and automation is a plus.
- Hands-on experience with testing frameworks and CI/CD tools like Jenkins or GitHub Actions.
- Experience working with Docker, Kubernetes, and Nomad for containerized environments.
- Familiarity with cloud platforms, particularly AWS,
- Experience with observability and monitoring tools like Grafana and Kibana.
- Ensure comprehensive test coverage through automation and manual testing where necessary.
- Research automation tools and infrastructure for improvement
- Review plans and consult junior QAs
- Excellent problem-solving skills and attention to detail.
- Strong communication and interpersonal skills.
- Define E2E testing requirements (scenarios, conditions, testing types, metrics formonitoring)
- Execute E2E manual tests
- Automate E2E regression
Role: Sr. Java Developer
Experience: 6+ Years
Location: Bangalore (Whitefield)
Work Mode: Hybrid (2-3 days WFO)
Shift Timing: Regular Morning Shift
About the Role:
We are looking for a seasoned Java Developer with 6+ years of experience to join our growing engineering team. The ideal candidate should have strong expertise in Java, Spring Boot, Microservices, and cloud-based deployment using AWS or DevOps tools. This is a hybrid role based out of our Whitefield, Bangalore location.
Key Responsibilities:
- Participate in agile development processes and scrum ceremonies.
- Translate business requirements into scalable and maintainable technical solutions.
- Design and develop applications using Java, Spring Boot, and Microservices architecture.
- Ensure robust and reliable code through full-scale unit testing and TDD/BDD practices.
- Contribute to CI/CD pipeline setup and cloud deployments.
- Work independently and as an individual contributor on complex features.
- Troubleshoot production issues and optimize application performance.
Mandatory Skills:
- Strong Core Java and Spring Boot expertise.
- Proficiency in AWS or DevOps (Docker & Kubernetes).
- Experience with relational and/or non-relational databases (SQL, NoSQL).
- Sound understanding of Microservices architecture and RESTful APIs.
- Containerization experience using Docker and orchestration via Kubernetes.
- Familiarity with Linux-based development environments.
- Exposure to modern SDLC tools – Maven, Git, Jenkins, etc.
- Good understanding of CI/CD pipelines and cloud-based deployment.
Soft Skills:
- Self-driven, proactive, and an individual contributor.
- Strong problem-solving and analytical skills.
- Excellent communication and interpersonal abilities.
- Able to plan, prioritize, and manage tasks independently.
Nice-to-Have Skills:
- Exposure to frontend technologies like Angular, JavaScript, HTML5, and CSS.
Looking for Fresher developers
Responsibilities:
- Implement integrations requested by customers
- Deploy updates and fixes
- Provide Level 2 technical support
- Build tools to reduce occurrences of errors and improve customer experience
- Develop software to integrate with internal back-end systems
- Perform root cause analysis for production errors
- Investigate and resolve technical issues
- Develop scripts to automate visualization
- Design procedures for system troubleshooting and maintenance
Requirements and skill:
Knowledge in DevOps Engineer or similar software engineering role
Good knowledge of Terraform, Kubernetes
Working knowledge on AWS, Google Cloud
You can directly contact me on nine three one six one two zero one three two

Job Title : Senior Consultant (Java / NodeJS + Temporal)
Experience : 5 to 12 Years
Location : Bengaluru, Chennai, Hyderabad, Pune, Mumbai, Gurugram, Coimbatore
Work Mode : Remote (Must be open to travel for occasional team meetups)
Notice Period : Immediate Joiners or Serving Notice
Interview Process :
- R1 : Tech Interview (60 mins)
- R2 : Technical Interview
- R3 : (Optional) Interview with Client
Job Summary :
We are seeking a Senior Backend Consultant with strong hands-on expertise in Temporal (BPM/Workflow Engine) and either Node.js or Java.
The ideal candidate will have experience in designing and developing microservices and process-driven applications, as well as orchestrating complex workflows using Temporal.io.
You will work on high-scale systems, collaborating closely with cross-functional teams.
Mandatory Skills :
Temporal.io, Node.js (or Java), React.js, Keycloak IAM, PostgreSQL, Terraform, Kubernetes, Azure, Jest, OpenAPI
Key Responsibilities :
- Design and implement scalable backend services using Node.js or Java.
- Build and manage complex workflow orchestrations using Temporal.io.
- Integrate with IAM solutions like Keycloak for role-based access control.
- Work with React (v17+), TypeScript, and component-driven frontend design.
- Use PostgreSQL for structured data persistence and optimized queries.
- Manage infrastructure using Terraform and orchestrate via Kubernetes.
- Leverage Azure Services like Blob Storage, API Gateway, and AKS.
- Write and maintain API documentation using Swagger/Postman/Insomnia.
- Conduct unit and integration testing using Jest.
- Participate in code reviews and contribute to architectural decisions.
Must-Have Skills :
- Temporal.io – BPMN modeling, external task workers, Operate, Tasklist
- Node.js + TypeScript (preferred) or strong Java experience
- React.js (v17+) and component-driven UI development
- Keycloak IAM, PostgreSQL, and modern API design
- Infrastructure automation with Terraform, Kubernetes
- Experience in using GitFlow, OpenAPI, Jest for testing
Nice-to-Have Skills :
- Blockchain integration experience for secure KYC/identity flows
- Custom Camunda Connectors or exporter plugin development
- CI/CD experience using Azure DevOps or GitHub Actions
- Identity-based task completion authorization enforcement
About the Role
We are looking for a skilled Backend Engineer with strong experience in building scalable microservices, integrating with distributed data systems, and deploying web APIs that serve UI applications in the cloud. You’ll work on high-performance systems involving Kafka, DynamoDB, Redis, and other modern backend technologies.
Responsibilities
- Design, develop, and deploy backend microservices and APIs that power UI applications.
- Implement event-driven architectures using Apache Kafka or similar messaging platforms.
- Build scalable and highly available systems using NoSQL databases (e.g., DynamoDB, MongoDB).
- Optimize backend systems using caching layers like Redis to enhance performance.
- Ensure seamless deployment and operation of services in cloud environments (AWS, GCP, or Azure).
- Write clean, maintainable, and well-tested code; contribute to code reviews and architecture discussions.
- Collaborate closely with frontend, DevOps, and product teams to deliver integrated solutions.
- Monitor and troubleshoot production issues and participate in on-call rotations as needed.
Required Qualifications
- 3–7 years of professional experience in backend development.
- Strong programming skills in one or more languages: Java, Python, Go, Node.js.
- Hands-on experience with microservices architecture and API design (REST/gRPC).
- Practical experience with Kafka, RabbitMQ, or other event streaming/message queue systems.
- Solid knowledge of NoSQL databases, especially DynamoDB or equivalents.
- Experience using Redis or Memcached for caching or pub/sub mechanisms.
- Proficiency with cloud platforms (preferably AWS – e.g., Lambda, ECS, EKS, API Gateway).
- Familiarity with Docker, Kubernetes, and CI/CD pipelines.
Job Description
What does a successful Senior DevOps Engineer do at Fiserv?
This role’s focus will be on contributing and enhancing our DevOps environment within Issuer Solution group, where our cross functional Scrum teams are delivering solutions built on cutting-edge mobile technology and products. You will be expected to support across the wider business unit, leading DevOps practices and initiatives.
What will you do:
• Build, manage, and deploy CI/CD pipelines.
• DevOps Engineer - Helm Chart, Rundesk, Openshift
• Strive for continuous improvement and build continuous integration, continuous development, and constant deployment pipeline.
• Implementing various development, testing, automation tools, and IT infrastructure
• Optimize and automate release/development cycles and processes.
• Be part of and help promote our DevOps culture.
• Identify and implement continuous improvements to the development practice
What you must have:
• 3+ years of experience in devops with hands-on experience in the following:
- Writing automation scripts for deployments and housekeeping using shell scripts (bash) and ansible playbooks
- Building docker images and running/managing docker instances
- Building Jenkins pipelines using groovy scripts
- Working knowledge on kubernetes including application deployments, managing application configurations and persistence volumes
• Has good understanding on infrastructure as code
• Ability to write and update documentation
• Demonstrate a logical, process orientated approach to problems and troubleshooting
• Ability to collaborate with multi development teams
What you are preferred to have:
• 8+ years of development experience
• Jenkins administration experience
• Hands-on experience in building and deploying helm charts
Process Skills:
• Should have worked in Agile Project
Behavioral Skills :
• Good Communication skills
Skills
PRIMARY COMPETENCY : Cloud Infra PRIMARY SKILL : DevOps PRIMARY SKILL PERCENTAGE : 100
About the Role
We are looking for a DevOps Engineer to build and maintain scalable, secure, and high-
performance infrastructure for our next-generation healthcare platform. You will be
responsible for automation, CI/CD pipelines, cloud infrastructure, and system reliability,
ensuring seamless deployment and operations.
Responsibilities
1. Infrastructure & Cloud Management
• Design, deploy, and manage cloud-based infrastructure (AWS, Azure, GCP)
• Implement containerization (Docker, Kubernetes) and microservices orchestration
• Optimize infrastructure cost, scalability, and performance
2. CI/CD & Automation
• Build and maintain CI/CD pipelines for automated deployments
• Automate infrastructure provisioning using Terraform, Ansible, or CloudFormation
• Implement GitOps practices for streamlined deployments
3. Security & Compliance
• Ensure adherence to ABDM, HIPAA, GDPR, and healthcare security standards
• Implement role-based access controls, encryption, and network security best
practices
• Conduct Vulnerability Assessment & Penetration Testing (VAPT) and compliance
audits
4. Monitoring & Incident Management
• Set up monitoring, logging, and alerting systems (Prometheus, Grafana, ELK,
Datadog, etc.)
• Optimize system reliability and automate incident response mechanisms
• Improve MTTR (Mean Time to Recovery) and system uptime KPIs
5. Collaboration & Process Improvement
• Work closely with development and QA teams to streamline deployments
• Improve DevSecOps practices and cloud security policies
• Participate in architecture discussions and performance tuning
Required Skills & Qualifications
• 2+ years of experience in DevOps, cloud infrastructure, and automation
• Hands-on experience with AWS and Kubernetes
• Proficiency in Docker and CI/CD tools (Jenkins, GitHub Actions, ArgoCD, etc.)
• Experience with Terraform, Ansible, or CloudFormation
• Strong knowledge of Linux, shell scripting, and networking
• Experience with cloud security, monitoring, and logging solutions
Nice to Have
• Experience in healthcare or other regulated industries
• Familiarity with serverless architectures and AI-driven infrastructure automation
• Knowledge of big data pipelines and analytics workflows
What You'll Gain
• Opportunity to build and scale a mission-critical healthcare infrastructure
• Work in a fast-paced startup environment with cutting-edge technologies
• Growth potential into Lead DevOps Engineer or Cloud Architect roles
Job Role : Azure DevSecOps Engineer (Security-Focused)
Experience : 12 to 18 Years
Location : Preferably Delhi NCR (Hybrid); Remote possible with 1–2 office visits per quarter (Gurgaon)
Joining Timeline : Max 45 days (Buyout option available)
Work Mode : Full-time | 5 Days Working
About the Role :
We are looking for a highly experienced Azure DevSecOps Engineer with a strong focus on cloud security practices.
This role is 60–70% security-driven, involving threat modeling, secure cloud architecture, and infrastructure security on Azure using Terraform.
Key Responsibilities :
- Architect and maintain secure, scalable Azure cloud infrastructure using Terraform.
- Implement security best practices : IAM, threat modeling, network security, data protection, and compliance (e.g., GDPR).
- Build CI/CD pipelines and automate deployments using Azure DevOps, Jenkins, Prometheus.
- Monitor, analyze, and proactively improve security posture.
- Collaborate with global teams to ensure secure design, development, and operations.
- Stay updated on cloud security trends and lead mitigation efforts.
Mandatory Skills :
Azure, Terraform, DevSecOps, Cloud Security, Threat Modelling, IAM, CI/CD (Azure DevOps), Docker, Kubernetes, Prometheus, Infrastructure as Code (IaC), Compliance Frameworks (GDPR)
Preferred Certifications :
Certified DevSecOps Professional (CDP), Microsoft Azure Certifications

Job Role : DevOps Engineer (Python + DevOps)
Experience : 4 to 10 Years
Location : Hyderabad
Work Mode : Hybrid
Mandatory Skills : Python, Ansible, Docker, Kubernetes, CI/CD, Cloud (AWS/Azure/GCP)
Job Description :
We are looking for a skilled DevOps Engineer with expertise in Python, Ansible, Docker, and Kubernetes.
The ideal candidate will have hands-on experience automating deployments, managing containerized applications, and ensuring infrastructure reliability.
Key Responsibilities :
- Design and manage containerization and orchestration using Docker & Kubernetes.
- Automate deployments and infrastructure tasks using Ansible & Python.
- Build and maintain CI/CD pipelines for streamlined software delivery.
- Collaborate with development teams to integrate DevOps best practices.
- Monitor, troubleshoot, and optimize system performance.
- Enforce security best practices in containerized environments.
- Provide operational support and contribute to continuous improvements.
Required Qualifications :
- Bachelor’s in Computer Science/IT or related field.
- 4+ years of DevOps experience.
- Proficiency in Python and Ansible.
- Expertise in Docker and Kubernetes.
- Hands-on experience with CI/CD tools and pipelines.
- Experience with at least one cloud provider (AWS, Azure, or GCP).
- Strong analytical, communication, and collaboration skills.
Preferred Qualifications :
- Experience with Infrastructure-as-Code tools like Terraform.
- Familiarity with monitoring/logging tools like Prometheus, Grafana, or ELK.
- Understanding of Agile/Scrum practices.

You will:
- Collaborate with the I-Stem Voice AI team and CEO to design, build and ship new agent capabilities
- Develop, test and refine end-to-end voice agent models (ASR, NLU, dialog management, TTS)
- Stress-test agents in noisy, real-world scenarios and iterate for improved robustness and low latency
- Research and prototype cutting-edge techniques (e.g. robust speech recognition, adaptive language understanding)
- Partner with backend and frontend engineers to seamlessly integrate AI components into live voice products
- Monitor agent performance in production, analyze failure cases, and drive continuous improvement
- Occasionally demo our Voice AI solutions at industry events and user forums
You are:
- An AI/Software Engineer with hands-on experience in speech-centric ML (ASR, NLU or TTS)
- Skilled in building and tuning transformer-based speech models and handling real-time audio pipelines
- Obsessed with reliability: you design experiments to push agents to their limits and root-cause every error
- A clear thinker who deconstructs complex voice interactions from first principles
- Passionate about making voice technology inclusive and accessible for diverse users
- Comfortable moving fast in a small team, yet dogged about code quality, testing and reproducibility
We are seeking an experienced and passionate Cloud and DevOps Trainer to join our training and development team. The trainer will be responsible for delivering high-quality, hands-on training in Cloud technologies (such as AWS, Azure, or GCP) and DevOps tools and practices to students or working professionals.


Skill Sets:
- Expertise in ML/DL, model lifecycle management, and MLOps (MLflow, Kubeflow)
- Proficiency in Python, TensorFlow, PyTorch, Scikit-learn, and Hugging Face models
- Strong experience in NLP, fine-tuning transformer models, and dataset preparation
- Hands-on with cloud platforms (AWS, GCP, Azure) and scalable ML deployment (Sagemaker, Vertex AI)
- Experience in containerization (Docker, Kubernetes) and CI/CD pipelines
- Knowledge of distributed computing (Spark, Ray), vector databases (FAISS, Milvus), and model optimization (quantization, pruning)
- Familiarity with model evaluation, hyperparameter tuning, and model monitoring for drift detection
Roles and Responsibilities:
- Design and implement end-to-end ML pipelines from data ingestion to production
- Develop, fine-tune, and optimize ML models, ensuring high performance and scalability
- Compare and evaluate models using key metrics (F1-score, AUC-ROC, BLEU etc)
- Automate model retraining, monitoring, and drift detection
- Collaborate with engineering teams for seamless ML integration
- Mentor junior team members and enforce best practices
What you’ll do
- Tame data → pull, clean, and shape structured & unstructured data.
- Orchestrate pipelines → Airflow / Step Functions / ADF… your call.
- Ship models → build, tune, and push to prod on SageMaker, Azure ML, or Vertex AI.
- Scale → Spark / Databricks for the heavy lifting.
- Automate everything → Docker, Kubernetes, CI/CD, MLFlow, Seldon, Kubeflow.
- Pair up → work with engineers, architects, and business folks to solve real problems, fast.
What you bring
- 3+ yrs hands-on MLOps (4-5 yrs total software experience).
- Proven chops on one hyperscaler (AWS, Azure, or GCP).
- Confidence with Databricks / Spark, Python, SQL, TensorFlow / PyTorch / Scikit-learn.
- You debug Kubernetes in your sleep and treat Dockerfiles like breathing.
- You prototype with open-source first, choose the right tool, then make it scale.
- Sharp mind, low ego, bias for action.
Nice-to-haves
- Sagemaker, Azure ML, or Vertex AI in production.
- Love for clean code, clear docs, and crisp PRs.

Job Title : Python Data Engineer
Experience : 4+ Years
Location : Bangalore / Hyderabad (On-site)
Job Summary :
We are seeking a skilled Python Data Engineer to work on cloud-native data platforms and backend services.
The role involves building scalable APIs, working with diverse data systems, and deploying containerized services using modern cloud infrastructure.
Mandatory Skills : Python, AWS, RESTful APIs, Microservices, SQL/PostgreSQL/NoSQL, Docker, Kubernetes, CI/CD (Jenkins/GitLab CI/AWS CodePipeline)
Key Responsibilities :
- Design, develop, and maintain backend systems using Python.
- Build and manage RESTful APIs and microservices architectures.
- Work extensively with AWS cloud services for deployment and data storage.
- Implement and manage SQL, PostgreSQL, and NoSQL databases.
- Containerize applications using Docker and orchestrate with Kubernetes.
- Set up and maintain CI/CD pipelines using Jenkins, GitLab CI, or AWS CodePipeline.
- Collaborate with teams to ensure scalable and reliable software delivery.
- Troubleshoot and optimize application performance.
Must-Have Skills :
- 4+ years of hands-on experience in Python backend development.
- Strong experience with AWS cloud infrastructure.
- Proficiency in building microservices and APIs.
- Good knowledge of relational and NoSQL databases.
- Experience with Docker and Kubernetes.
- Familiarity with CI/CD tools and DevOps processes.
- Strong problem-solving and collaboration skills.

A leader in telecom, fintech, AI-led marketing automation.

We are looking for a talented MERN Developer with expertise in MongoDB/MySQL, Kubernetes, Python, ETL, Hadoop, and Spark. The ideal candidate will design, develop, and optimize scalable applications while ensuring efficient source code management and implementing Non-Functional Requirements (NFRs).
Key Responsibilities:
- Develop and maintain robust applications using MERN Stack (MongoDB, Express.js, React.js, Node.js).
- Design efficient database architectures (MongoDB/MySQL) for scalable data handling.
- Implement and manage Kubernetes-based deployment strategies for containerized applications.
- Ensure compliance with Non-Functional Requirements (NFRs), including source code management, development tools, and security best practices.
- Develop and integrate Python-based functionalities for data processing and automation.
- Work with ETL pipelines for smooth data transformations.
- Leverage Hadoop and Spark for processing and optimizing large-scale data operations.
- Collaborate with solution architects, DevOps teams, and data engineers to enhance system performance.
- Conduct code reviews, troubleshooting, and performance optimization to ensure seamless application functionality.
Required Skills & Qualifications:
- Proficiency in MERN Stack (MongoDB, Express.js, React.js, Node.js).
- Strong understanding of database technologies (MongoDB/MySQL).
- Experience working with Kubernetes for container orchestration.
- Hands-on knowledge of Non-Functional Requirements (NFRs) in application development.
- Expertise in Python, ETL pipelines, and big data technologies (Hadoop, Spark).
- Strong problem-solving and debugging skills.
- Knowledge of microservices architecture and cloud computing frameworks.
Preferred Qualifications:
- Certifications in cloud computing, Kubernetes, or database management.
- Experience in DevOps, CI/CD automation, and infrastructure management.
- Understanding of security best practices in application development.

We are looking for a Senior Software Engineer with 5+ years of experience in modern C++ development, paired with strong hands-on skills in AWS, Node.js, data processing, and containerized service development. The ideal candidate will be responsible for building scalable systems, maintaining complex data pipelines, and modernizing applications through cloud-native approaches and automation.
This is a high-impact role where engineering depth meets platform evolution, ideal for someone who thrives in system-level thinking, data-driven applications, and full-stack delivery.
Key Responsibilities:
- Design, build, and maintain high-performance systems using modern C++
- Develop and deploy scalable backend services using Node.js and manage dependencies via NPM
- Architect and implement containerized services using Docker, with orchestration via Kubernetes or ECS
- Build, monitor, and maintain data ingestion, transformation, and enrichment pipelines
- Utilize AWS services (Lambda, EC2, S3, CloudWatch, Step Functions) to deliver reliable cloud-native solutions
- Implement and maintain modern CI/CD pipelines, ensuring seamless integration, testing, and delivery
- Participate in system design, peer code reviews, and performance tuning
Required Skills:
- 5+ years of software development experience, with strong command over modern C++
- Solid experience with Node.js, JavaScript, and NPM for backend development
- Deep understanding of cloud platforms (preferably AWS) and hands-on experience in deploying and managing applications in the cloud
- Proficient in building and scaling data processing workflows and working with structured/unstructured data
- Strong hands-on experience with Docker, container orchestration, and microservices architecture
- Working knowledge of CI/CD practices, Git, and build/release tools
- Strong problem-solving, debugging, and cross-functional collaboration skills
Preferred / Nice to Have:
- Exposure to data streaming frameworks (Kafka, Spark, etc.)
- Familiarity with monitoring and observability tools (e.g., Prometheus, Grafana, ELK stack)
- Background in performance profiling, secure coding, or legacy modernization
- Ability to work in agile environments and lead small technical initiatives
Minimum requirements
5+ years of industry software engineering experience (does not include internships nor includes co-ops)
Strong coding skills in any programming language (we understand new languages can be learned on the job so our interview process is language agnostic)
Strong collaboration skills, can work across workstreams within your team and contribute to your peers’ success
Have the ability to thrive on a high level of autonomy, responsibility, and think of yourself as entrepreneurial
Interest in working as a generalist across varying technologies and stacks to solve problems and delight both internal and external users
Preferred Qualifications
Experience with large-scale financial tracking systems
Good understanding and practical knowledge in cloud based services (e.g. gRPC, GraphQL, Docker/Kubernetes, cloud services such as AWS, etc.)
Job Summary:
We are seeking a skilled DevOps Engineer to design, implement, and manage CI/CD pipelines, containerized environments, and infrastructure automation. The ideal candidate should have hands-on experience with ArgoCD, Kubernetes, and Docker, along with a deep understanding of cloud platforms and deployment strategies.
Key Responsibilities:
- CI/CD Implementation: Develop, maintain, and optimize CI/CD pipelines using ArgoCD, GitOps, and other automation tools.
- Container Orchestration: Deploy, manage, and troubleshoot containerized applications using Kubernetes and Docker.
- Infrastructure as Code (IaC): Automate infrastructure provisioning with Terraform, Helm, or Ansible.
- Monitoring & Logging: Implement and maintain observability tools like Prometheus, Grafana, ELK, or Loki.
- Security & Compliance: Ensure best security practices in containerized and cloud-native environments.
- Cloud & Automation: Manage cloud infrastructure on AWS, Azure, or GCP with automated deployments.
- Collaboration: Work closely with development teams to optimize deployments and performance.
Required Skills & Qualifications:
- Experience: 5+ years in DevOps, Site Reliability Engineering (SRE), or Infrastructure Engineering.
- Tools & Tech: Strong knowledge of ArgoCD, Kubernetes, Docker, Helm, Terraform, and CI/CD pipelines.
- Cloud Platforms: Experience with AWS, GCP, or Azure.
- Programming & Scripting: Proficiency in Python, Bash, or Go.
- Version Control: Hands-on with Git and GitOps workflows.
- Networking & Security: Knowledge of ingress controllers, service mesh (Istio/Linkerd), and container security best practices.
Nice to Have:
- Experience with Kubernetes Operators, Kustomize, or FluxCD.
- Exposure to serverless architectures and multi-cloud deployments.
- Certifications in CKA, AWS DevOps, or similar.
Job Title : Java Backend Developer
Experience : 3 to 6 Years
Locations : Bangalore / Gurgaon (Hybrid – 3 Days Work From Office)
Shift Timings : 11:00 AM – 8:00 PM IST
Notice Period : Immediate to 15 Days Only
Job Description :
We are looking for experienced Java Backend Developers with strong expertise in building scalable microservices-based architectures. The ideal candidate should have hands-on experience with Spring Boot, containerized deployments, and DevOps tools.
✅ Must-Have Skills :
- Java – Strong programming skills in core Java.
- Spring Boot (2.x / 3.x) – Deep understanding of microservices architecture and patterns.
- Microservices – Design and implementation experience.
- Kubernetes – Experience deploying and managing microservices
- Jenkins & Maven – Build and CI/CD pipeline experience
- PostgreSQL – Experience with relational database management
✨ Good-to-Have Skills :
- Git – Source control management
- CI/CD Pipeline Tools – End-to-end pipeline automation
- Cloud & DevOps Knowledge – Experience with cloud-based deployments
We are seeking an experienced and highly skilled Technical Lead with a strong background in Java, SaaS architectures, firewalls, and cybersecurity products, including SIEM and SOAR platforms. The ideal candidate will lead technical initiatives, design and implement scalable systems, and drive best practices across the engineering team. This role requires deep technical expertise, leadership abilities, and a passion for building secure and high-performing security solutions.
Key Roles & Responsibilities:
- Lead the design and development of scalable and secure software solutions using Java.
- Architect and build SaaS-based cybersecurity applications, ensuring high availability, performance, and reliability.
- Provide technical leadership, mentoring, and guidance to the development team.
- Ensure best practices in secure coding, threat modeling, and compliance with industry standards.
- Collaborate with cross-functional teams, including Product Management, Security, and DevOps to deliver high-quality security solutions.
- Design and implement security analytics, automation workflows and ITSM integrations.
- Drive continuous improvements in engineering processes, tools, and technologies.
- Troubleshoot complex technical issues and lead incident response for critical production systems.
Basic Qualifications:
- A bachelor’s or master’s degree in computer science, electronics engineering or a related field
- 3-6 years of software development experience, with expertise in Java.
- Strong background in building SaaS applications with cloud-native architectures (AWS, GCP, or Azure).
- In-depth understanding of microservices architecture, APIs, and distributed systems.
- Experience with containerization and orchestration tools like Docker and Kubernetes.
- Knowledge of DevSecOps principles, CI/CD pipelines, and infrastructure as code (Terraform, Ansible, etc.).
- Strong problem-solving skills and ability to work in an agile, fast-paced environment.
- Excellent communication and leadership skills, with a track record of mentoring engineers.
Preferred Qualifications:
- Experience with cybersecurity solutions, including SIEM (e.g., Splunk, ELK, IBM QRadar) and SOAR (e.g., Palo Alto XSOAR, Swimlane).
- Knowledge of zero-trust security models and secure API development.
- Hands-on experience with machine learning or AI-driven security analytics.
Job description
● Design effective, scalable architectures on top of cloud technologies such as AWS and Kubernetes
● Mentor other software engineers, including actively participating in peer code and architecture review
● Participate in all parts of the development lifecycle from design to coding to deployment to maintenance and operations
● Kickstart new ideas, build proof of concepts and jumpstart newly funded projects
● Demonstrate ability to work independently with minimal supervision
● Embed with other engineering teams on challenging initiatives and time sensitive projects
● Collaborate with other engineering teams on challenging initiatives and time sensitive projects
Education and Experience
● BS degree in Computer Science or related technical field or equivalent practical experience.
● 9+ years of professional software development experience focused on payments and/or billing and customer accounts. Worked with worldwide payments, billing systems, PCI Compliance & payment gateways.
Technical and Functional
● Extensive knowledge of micro service development using Spring, Spring Boot, Java - built on top of Kubernetes and public cloud computing such as AWS, Lambda, S3.
● Experience with relational databases (MySQL, DB2 or Oracle) and NoSQL databases
● Experience with unit testing and test driven development
Technologies at Constant Contact
Working on the Constant Contact platform provides our engineers with an opportunity to produce high impact work inside of our multifaceted platform (Email, Social, SMS, E-Commerce, CRM, Customer Data Platform, MLBased Recommendations & Insights, and more).
As a member of our team, you'll be utilizing the latest technologies and frameworks (React/SPA, JavaScript/TypeScript, Swift, Kotlin, GraphQL, etc) and deploying code to our cloud-first microservice infrastructure (declarative CI/CD, GitOps managed kubernetes) with regular opportunities to level up your skills.
● Past experience of working with and integrating payment gateways and processors, online payment methods, and billing systems.
● Familiar with integrating Stripe/Plaid/PayPal/Adyen/Cybersource or similar systems along with PCI compliance.
● International software development and payments experience is a plus.
● Knowledge of DevOps and CI/CD, automated test and build tools ( Jenkins & Gradle/Maven)
● Experience integrating with sales tax engines is a plus.
● Familiar with tools like Splunk, New relic or similar tools like datadog, elastic elk, amazon
cloudwatch.
● Good to have - Experience with React, Backbone, Marionette or other front end frameworks.
Cultural
● Strong verbal and written communication skills.
● Flexible attitude and willingness to frequently move between different teams, software architectures and priorities.
● Desire to collaborate with our other product teams to think strategically about how to solve problems.
Our team
● We focus on cross-functional team collaboration where engineers, product managers, and designers all work together to solve customer problems and build exciting features.
● We love new ideas and are eager to see what your experiences can bring to help influence our technical and product vision.
● Collaborate/Overlap with the teams in Eastern Standard Time (EST), USA.


Requirements
- 7+ years of experience with Python
- Strong expertise in Python frameworks (Django, Flask, or FastAPI)
- Experience with GCP, Terraform, and Kubernetes
- Deep understanding of REST API development and GraphQL
- Strong knowledge of SQL and NoSQL databases
- Experience with microservices architecture
- Proficiency with CI/CD tools (Jenkins, CircleCI, GitLab)
- Experience with container orchestration using Kubernetes
- Understanding of cloud architecture and serverless computing
- Experience with monitoring and logging solutions
- Strong background in writing unit and integration tests
- Familiarity with AI/ML concepts and integration points
Responsibilities
- Design and develop scalable backend services for our AI platform
- Architect and implement complex systems with high reliability
- Build and maintain APIs for internal and external consumption
- Work closely with AI engineers to integrate ML functionality
- Optimize application performance and resource utilization
- Make architectural decisions that balance immediate needs with long-term scalability
- Mentor junior engineers and promote best practices
- Contribute to the evolution of our technical standards and processes
About PGAGI:
At PGAGI, we believe in a future where AI and human intelligence coexist in harmony, creating a world that is smarter, faster, and better. We are not just building AI; we are shaping a future where AI is a fundamental and positive force for businesses, societies, and the planet.
Position Overview:
PGAGI Consultancy Pvt. Ltd. is seeking a proactive and motivated DevOps Intern with around 3-6 months of hands-on experience to support our AI model deployment and infrastructure initiatives. This role is ideal for someone looking to deepen their expertise in DevOps practices tailored to AI/ML environments, including CI/CD automation, cloud infrastructure, containerization, and monitoring.
Key Responsibilities:
AI Model Deployment & Integration
- Assist in containerizing and deploying AI/ML models into production using Docker.
- Support integration of models into existing systems and APIs.
Infrastructure Management
- Help manage cloud and on-premise environments to ensure scalability and consistency.
- Work with Kubernetes for orchestration and environment scaling.
CI/CD Pipeline Automation
- Collaborate on building and maintaining automated CI/CD pipelines (e.g., GitHub Actions, Jenkins).
- Implement basic automated testing and rollback mechanisms.
Hosting & Web Environment Management
- Assist in managing hosting platforms, web servers, and CDN configurations.
- Support DNS, load balancer setups, and ensure high availability of web services.
Monitoring, Logging & Optimization
- Set up and maintain monitoring/logging tools like Prometheus and Grafana.
- Participate in troubleshooting and resolving performance bottlenecks.
Security & Compliance
- Apply basic DevSecOps practices including security scans and access control implementations.
- Follow security and compliance checklists under supervision.
Cost & Resource Management
- Monitor resource usage and suggest cost optimization strategies in cloud environments.
Documentation
- Maintain accurate documentation for deployment processes and incident responses.
Continuous Learning & Innovation
- Suggest improvements to workflows and tools.
- Stay updated with the latest DevOps and AI infrastructure trends.
Requirements:
- Around 6 months of experience in a DevOps or related technical role (internship or professional).
- Basic understanding of Docker, Kubernetes, and CI/CD tools like GitHub Actions or Jenkins.
- Familiarity with cloud platforms (AWS, GCP, or Azure) and monitoring tools (e.g., Prometheus, Grafana).
- Exposure to scripting languages (e.g., Bash, Python) is a plus.
- Strong problem-solving skills and eagerness to learn.
- Good communication and documentation abilities.
Compensation
- Joining Bonus: INR 2,500 one-time bonus upon joining.
- Monthly Stipend: Base stipend of INR 8,000 per month, with the potential to increase up to INR 20,000 based on performance evaluations.
- Performance-Based Pay Scale: Eligibility for monthly performance-based bonuses, rewarding exceptional project contributions and teamwork.
- Additional Benefits: Access to professional development opportunities, including workshops, tech talks, and mentoring sessions.
Ready to kick-start your DevOps journey in a dynamic AI-driven environment? Apply now
#Devops #Docker #Kubernetes #DevOpsIntern
About the Role:
We are looking for a skilled AWS DevOps Engineer to join our Cloud Operations team in Bangalore. This hybrid role is ideal for someone with hands-on experience in AWS and a strong background in application migration from on-premises to cloud environments. You'll play a key role in driving cloud adoption, optimizing infrastructure, and ensuring seamless cloud operations.
Key Responsibilities:
- Manage and maintain AWS cloud infrastructure and services.
- Lead and support application migration projects from on-prem to cloud.
- Automate infrastructure provisioning using Infrastructure as Code (IaC) tools.
- Monitor cloud environments and optimize cost, performance, and reliability.
- Collaborate with development, operations, and security teams to implement DevOps best practices.
- Troubleshoot and resolve infrastructure and deployment issues.
Required Skills:
- 3–5 years of experience in AWS cloud environment.
- Proven experience with on-premises to cloud application migration.
- Strong understanding of AWS core services (EC2, VPC, S3, IAM, RDS, etc.).
- Solid scripting skills (Python, Bash, or similar).
Good to Have:
- Experience with Terraform for Infrastructure as Code.
- Familiarity with Kubernetes for container orchestration.
- Exposure to CI/CD tools like Jenkins, GitLab, or AWS CodePipeline.
Key Responsibilities:
- ☁️ Manage cloud infrastructure and automation on AWS, Google Cloud (GCP), and Azure.
- 🖥️ Deploy and maintain Windows Server environments, including Internet Information Services (IIS).
- 🐧 Administer Linux servers and ensure their security and performance.
- 🚀 Deploy .NET applications (ASP.Net, MVC, Web API, WCF, etc.) using Jenkins CI/CD pipelines.
- 🔗 Manage source code repositories using GitLab or GitHub.
- 📊 Monitor and troubleshoot cloud and on-premises server performance and availability.
- 🤝 Collaborate with development teams to support application deployments and maintenance.
- 🔒 Implement security best practices across cloud and server environments.
Required Skills:
- ☁️ Hands-on experience with AWS, Google Cloud (GCP), and Azure cloud services.
- 🖥️ Strong understanding of Windows Server administration and IIS.
- 🐧 Proficiency in Linux server management.
- 🚀 Experience in deploying .NET applications and working with Jenkins for CI/CD automation.
- 🔗 Knowledge of version control systems such as GitLab or GitHub.
- 🛠️ Good troubleshooting skills and ability to resolve system issues efficiently.
- 📝 Strong documentation and communication skills.
Preferred Skills:
- 🖥️ Experience with scripting languages (PowerShell, Bash, or Python) for automation.
- 📦 Knowledge of containerization technologies (Docker, Kubernetes) is a plus.
- 🔒 Understanding of networking concepts, firewalls, and security best practices.

🔹 Job Title: Full Stack Java Developer
🔹 Experience: 4 to 6 Years
🔹 Location: Gurugram (Hybrid)
📝 Job Description:
Deqode is hiring a skilled Full Stack Java Developer to join our team of technology experts building scalable, enterprise-grade systems. We’re looking for passionate developers who thrive in fast-paced environments and love solving real-world challenges.
Key Responsibilities:
- Design, develop, and deploy scalable Java-based microservices using Spring Boot.
- Develop front-end components using any modern JavaScript framework.
- Build robust APIs and integrate third-party services.
- Work with Quarkus to enhance Java runtime performance (preferred).
- Implement containerized services using Kubernetes and manage deployments.
- Ensure clean, testable, and maintainable code in an Agile environment.
- Collaborate with cross-functional teams to define and deliver high-quality products.
Must-Have Skills:
- 4–6 years of hands-on experience in Java and Spring Boot
- Proven experience in building and deploying Microservices
- Experience with Kubernetes and Kafka
- Proficiency in any modern frontend framework (React, Angular, etc.)
- Exposure to Quarkus is a strong plus
- Strong understanding of API design and RESTful services
- Excellent problem-solving and communication skills
Required Skills:
- Experience in systems administration, SRE or DevOps focused role
- Experience in handling production support (on-call)
- Good understanding of the Linux operating system and networking concepts.
- Demonstrated competency with the following AWS services: ECS, EC2, EBS, EKS, S3, RDS, ELB, IAM, Lambda.
- Experience with Docker containers and containerization concepts
- Experience with managing and scaling Kubernetes clusters in a production environment
- Experience building scalable infrastructure in AWS with Terraform.
- Strong knowledge of Protocol-level such as HTTP/HTTPS, SMTP, DNS, and LDAP
- Experience monitoring production systems
- Expertise in leveraging Automation / DevOps principles, experience with operational tools, and able to apply best practices for infrastructure and software deployment (Ansible).
- HAProxy, Nginx, SSH, MySQL configuration and operation experience
- Ability to work seamlessly with software developers, QA, project managers, and business development
- Ability to produce and maintain written documentation
Job Title: Backend Developer
Location: In-Office, Bangalore, Karnataka, India
Job Summary:
We are seeking a highly skilled and experienced Backend Developer with a minimum of 1 year of experience in product building to join our dynamic and innovative team. In this role, you will be responsible for designing, developing, and maintaining robust backend systems that drive our applications. You will collaborate with cross-functional teams to ensure seamless integration between frontend and backend components, and your expertise will be critical in architecting scalable, secure, and high-performance backend solutions.
Annual Compensation: 6-10 LPA
Responsibilities:
- Design, develop, and maintain scalable and efficient backend systems and APIs using NodeJS.
- Architect and implement complex backend solutions, ensuring high availability and performance.
- Collaborate with product managers, frontend developers, and other stakeholders to deliver comprehensive end-to-end solutions.
- Design and optimize data storage solutions using relational databases (e.g., MySQL) and NoSQL databases (e.g., MongoDB, Redis).
- Promoting a culture of collaboration, knowledge sharing, and continuous improvement.
- Implement and enforce best practices for code quality, security, and performance optimization.
- Develop and maintain CI/CD pipelines to automate build, test, and deployment processes.
- Ensure comprehensive test coverage, including unit testing, and implement various testing methodologies and tools to validate application functionality.
- Utilize cloud services (e.g., AWS, Azure, GCP) for infrastructure deployment, management, and optimization.
- Conduct system design reviews and contribute to architectural discussions.
- Stay updated with industry trends and emerging technologies to drive innovation within the team.
- Implement secure authentication and authorization mechanisms and ensure data encryption for sensitive information.
- Design and develop event-driven applications utilizing serverless computing principles to enhance scalability and efficiency.
Requirements:
- Minimum of 1 year of proven experience as a Backend Developer, with a strong portfolio of product-building projects.
- Extensive experience with JavaScript backend frameworks (e.g., Express, Socket) and a deep understanding of their ecosystems.
- Strong expertise in SQL and NoSQL databases (MySQL and MongoDB) with a focus on data modeling and scalability.
- Practical experience with Redis and caching mechanisms to enhance application performance.
- Proficient in RESTful API design and development, with a strong understanding of API security best practices.
- In-depth knowledge of asynchronous programming and event-driven architecture.
- Familiarity with the entire web stack, including protocols, web server optimization techniques, and performance tuning.
- Experience with containerization and orchestration technologies (e.g., Docker, Kubernetes) is highly desirable.
- Proven experience working with cloud technologies (AWS/GCP/Azure) and understanding of cloud architecture principles.
- Strong understanding of fundamental design principles behind scalable applications and microservices architecture.
- Excellent problem-solving, analytical, and communication skills.
- Ability to work collaboratively in a fast-paced, agile environment and lead projects to successful completion.


Title: Senior Software Engineer – Python (Remote: Africa, India, Portugal)
Experience: 9 to 12 Years
INR : 40 LPA - 50 LPA
Location Requirement: Candidates must be based in Africa, India, or Portugal. Applicants outside these regions will not be considered.
Must-Have Qualifications:
- 8+ years in software development with expertise in Python
- kubernetes is important
- Strong understanding of async frameworks (e.g., asyncio)
- Experience with FastAPI, Flask, or Django for microservices
- Proficiency with Docker and Kubernetes/AWS ECS
- Familiarity with AWS, Azure, or GCP and IaC tools (CDK, Terraform)
- Knowledge of SQL and NoSQL databases (PostgreSQL, Cassandra, DynamoDB)
- Exposure to GenAI tools and LLM APIs (e.g., LangChain)
- CI/CD and DevOps best practices
- Strong communication and mentorship skills
🔥 High Priority – Senior Lead Java Developer (10+ Years) | Bangalore – Onsite
Summary :
We are hiring Senior Lead Java Developers with 10+ years of experience for an onsite role in Bangalore.
If you're a hands-on expert with a strong background in Java, Spring Boot, Microservices, and Kubernetes, this is your opportunity to lead, mentor, and deliver high-quality solutions in a fast-paced environment.
🔹 Position : Senior Lead Java Developer
🔹 Experience : 10+ Years (12+ preferred)
🔹 Location : Bangalore (Onsite)
🔹 Openings : 6+
✅ Must-Have Skills :
- 8+ years of hands-on experience with Core Java & Spring Boot
- Expertise in Multithreading, Dependency Injection, and AOP
- Strong in Microservices Architecture and RESTful services
- Good exposure to SQL & NoSQL databases
- Proficient with Git (GitLab preferred)
- Experience with Kubernetes deployments and APM tools (New Relic preferred)
- Solid understanding of distributed tracing and log analysis
- Proven debugging and performance optimization skills
💼 Responsibilities :
- Design and develop high-quality, scalable microservices
- Act as SME for multiple services or subsystems
- Own service performance, SLAs, and incident resolutions
- Mentor junior developers and conduct technical interviews
- Participate in production war rooms and troubleshooting
- Lead development efforts and drive code quality
🎓 Qualification :
- BE/B.Tech or equivalent degree

Location: Remote (India only)
About Certa At Certa, we’re revolutionizing process automation for top-tier companies, including Fortune 500 and Fortune 1000 leaders, from the heart of Silicon Valley. Our mission? Simplifying complexity through cutting-edge SaaS solutions. Join our thriving, global team and become a key player in a startup environment that champions innovation, continuous learning, and unlimited growth. We offer a fully remote, flexible workspace that empowers you to excel.
Role Overview
Ready to elevate your DevOps career by shaping the backbone of a fast-growing SaaS platform? As a Senior DevOps Engineer at Certa, you’ll lead the charge in building, automating, and optimizing our cloud infrastructure. Beyond infrastructure management, you’ll actively contribute with a product-focused mindset, understanding customer requirements, collaborating closely with product and engineering teams, and ensuring our AWS-based platform consistently meets user needs and business goals.
What You’ll Do
- Own SaaS Infrastructure: Design, architect, and maintain robust, scalable AWS infrastructure, enhancing platform stability, security, and performance.
- Orchestrate with Kubernetes: Utilize your advanced Kubernetes expertise to manage and scale containerized deployments efficiently and reliably.
- Collaborate on Enterprise Architecture: Align infrastructure strategies with enterprise architectural standards, partnering closely with architects to build integrated solutions.
- Drive Observability: Implement and evolve sophisticated monitoring and observability solutions (DataDog, ELK Stack, AWS CloudWatch) to proactively detect, troubleshoot, and resolve system anomalies.
- Lead Automation Initiatives: Champion an automation-first mindset across the organization, streamlining development, deployment, and operational workflows.
- Implement Infrastructure as Code (IaC): Master Terraform to build repeatable, maintainable cloud infrastructure automation.
- Optimize CI/CD Pipelines: Refine and manage continuous integration and deployment processes (currently GitHub Actions, transitioning to CircleCI), enhancing efficiency and reliability.
- Enable GitOps with ArgoCD: Deliver seamless GitOps-driven application deployments, ensuring accuracy and consistency in Kubernetes environments.
- Advocate for Best Practices: Continuously promote and enforce industry-standard DevOps practices, ensuring consistent, secure, and efficient operational outcomes.
- Innovate and Improve: Constantly evaluate and enhance current DevOps processes, tooling, and methodologies to maintain cutting-edge efficiency.
- Product Mindset: Actively engage with product and engineering teams, bringing infrastructure expertise to product discussions, understanding customer needs, and helping prioritize infrastructure improvements that directly benefit users and business objectives.
What You Bring
- Hands-On Experience: 3-5 years in DevOps roles, ideally within fast-paced SaaS environments.
- Kubernetes Mastery: Advanced knowledge and practical experience managing Kubernetes clusters and container orchestration.
- AWS Excellence: Comprehensive expertise across AWS services, infrastructure management, and security.
- IaC Competence: Demonstrated skill in Terraform for infrastructure automation and management.
- CI/CD Acumen: Proven proficiency managing pipelines with GitHub Actions; familiarity with CircleCI highly advantageous.
- GitOps Knowledge: Experience with ArgoCD for effective continuous deployment and operations.
- Observability Skills: Strong capabilities deploying and managing monitoring solutions such as DataDog, ELK, and AWS CloudWatch.
- Python Automation: Solid scripting and automation skills using Python.
- Architectural Awareness: Understanding of enterprise architecture frameworks and alignment practices.
- Proactive Problem-Solving: Exceptional analytical and troubleshooting skills, adept at swiftly addressing complex technical challenges.
- Effective Communication: Strong interpersonal and collaborative skills, essential for remote, distributed teamwork.
- Product Focus: Ability and willingness to understand customer requirements, prioritize tasks that enhance product value, and proactively suggest infrastructure improvements driven by user needs.
- Startup Mindset (Bonus): Prior experience or enthusiasm for dynamic startup cultures is a distinct advantage.
Why Join Us
- Compensation: Top-tier salary and exceptional benefits.
- Work-Life Flexibility: Fully remote, flexible scheduling.
- Growth Opportunities: Accelerate your career in a company poised for significant growth.
- Innovative Culture: Engineering-centric, innovation-driven work environment.
- Team Events: Annual offsites and quarterly Hackerhouse.
- Wellness & Family: Comprehensive healthcare and parental leave.
- Workspace: Premium workstation setup allowance, providing the tech you need to succeed.
Red Hat OpenShift Administrator :
Job Description:
• We are seeking a skilled RedHat OpenShift Administrator to manage and maintain our OpenShift container platform.
• The ideal candidate will have experience in administering Kubernetes-based environments and ensuring the reliability, performance, and security of our containerized applications.
Key Responsibilities:
• Install, configure, and maintain RedHat OpenShift clusters.
• Monitor and manage the health and performance of OpenShift environments.
• Implement and manage security policies and procedures for OpenShift clusters.
• Troubleshoot and resolve issues related to OpenShift and containerized applications.
• Collaborate with development and operations teams to support CI/CD pipelines and application deployments.
• Perform regular updates and patch management for OpenShift components.
• Develop and maintain documentation for OpenShift administration processes and best practices.
• Stay current with industry trends and advancements in container technologies.
Qualifications:
• Proven experience with RedHat OpenShift and Kubernetes administration.
• Strong knowledge of containerization technologies (Docker, CRI-O).
• Experience with Linux system administration and scripting (Bash, Python).
• Familiarity with CI/CD tools and processes (Jenkins, GitLab CI).
• Excellent problem-solving and analytical skills.
• Strong communication and collaboration abilities.
Preferred Certifications:
• Red Hat Certified Specialist in OpenShift Administration
• Certified Kubernetes Administrator (CKA)
Position: Project Manager
Location: Bengaluru, India (Hybrid/Remote flexibility available)
Company: PGAGI Consultancy Pvt. Ltd
About PGAGI
At PGAGI, we are building the future where human and artificial intelligence coexist to solve complex problems, accelerate innovation, and power sustainable growth. We develop and deploy advanced AI solutions across industries, making AI not just a tool but a transformational force for businesses and society.
Position Summary
PGAGI is seeking a dynamic and experienced Project Manager to lead cross-functional engineering teams and drive the successful execution of multiple AI/ML-centric projects. The ideal candidate is a strategic thinker with a solid background in engineering-led product/project management, especially in AI/ML product lifecycles. This role is crucial to scaling our technical operations, ensuring seamless collaboration, timely delivery, and high-impact results across initiatives.
Key Responsibilities
• Lead Engineering Teams Across AI/ML Projects: Manage and mentor cross-functional teams of ML engineers, DevOps professionals, and software developers through agile delivery cycles, ensuring timely and high-quality execution of AI-focused initiatives.
• Drive Agile Project Execution: Define project scope, objectives, timelines, and deliverables using Agile/Scrum methodologies. Ensure continuous sprint planning, backlog grooming, and milestone tracking via tools like Jira or GitHub Projects.
• Manage Multiple Concurrent Projects: Oversee the full lifecycle of multiple high-priority projects—ranging from AI model development and infrastructure integration to client delivery and platform enhancements.
• Collaborate with Technical and Business Stakeholders: Act as the bridge between engineering, research, and client-facing teams, translating complex requirements into actionable tasks and product features.
• Maintain Engineering and Infrastructure Quality: Uphold rigorous engineering standards across deployments. Coordinate testing, model performance validation, version control, and CI/CD operations.
• Budget and Resource Allocation: Optimize resource distribution across teams, track project costs, and ensure effective use of cloud infrastructure and personnel to maximize project ROI.
• Risk Management & Mitigation: Identify risks proactively across technical and operational layers. Develop mitigation plans and troubleshoot issues that may impact timelines or performance.
• Monitor KPIs and Delivery Metrics: Establish and monitor performance indicators such as sprint velocity, deployment frequency, incident response times, and customer satisfaction for each release.
• Support Continuous Improvement: Foster a culture of feedback and iteration. Champion retrospectives and process reviews to continually refine development practices and workflows.
Qualifications:
• Education: Bachelor’s or Master’s in Computer Science, Engineering, or a related technical field.
• Experience: Minimum 5 years of experience as a Project Manager, with at least 2 years managing AI/ML or software engineering teams.
• Tech Expertise: Familiarity with AI/ML lifecycles, cloud platforms (AWS, GCP, or Azure), and DevOps pipelines (Docker, Kubernetes, GitHub Actions, Jenkins).
• Tools: Strong experience with Jira, Confluence, and project tracking/reporting tools.
• Leadership: Proven success leading high-performing engineering teams in a fast-paced, innovative environment.
• Communication: Excellent written and verbal skills to interface with both technical and non-technical stakeholders.
• Certifications (Preferred): PMP, CSM, or certifications in AI/ML project management or cloud technologies.
Why Join PGAGI?
• Lead cutting-edge AI/ML product teams building scalable, impactful solutions.
• Be part of a fast-growing, innovation-driven startup environment.
• Enjoy a collaborative, intellectually stimulating workplace with growth opportunities.
• Competitive compensation and performance-based rewards.
• Access to learning resources, mentoring, and AI/DevOps communities.
Responsibilities
- Collaborate with the development team to maintain, enhance, and scale the product for enterprise use.
- Design and develop scalable, high-performance solutions using cloud technologies and containerization.
- Contribute to all development lifecycle phases, following SOLID principles and best practices.
- Write well-designed, testable, and efficient code with a strong emphasis on Test-Driven Development (TDD), ensuring comprehensive unit, integration, and performance testing.
- Ensure software designs comply with specifications and security best practices.
- Recommend changes to improve application architecture, maintainability, and performance.
- Develop and optimise database queries using T-SQL.
- Build and maintain microservices-based architectures where applicable.
- Prepare and produce software component releases.
- Develop and execute unit, integration, and performance tests.
- Support formal testing cycles and resolve test defects.
- Enhance software efficiency, maintainability, and scalability, with a strong focus on cloud-native solutions.
Skills Required:
Must Have
A. Highly proficient in ASP.NET Core (C#) and full-stack development.
B. Proficiency in front-end technologies (JavaScript, HTML, CSS, Bootstrap, and UI frameworks)
C. Experience developing REST APIs and Microservices.
D. Strong database experience, particularly with T-SQL and relational database design.
E. Experience in cloud platforms (Azure preferred), including scalability and enterprise cloud.
F. Experience with security best practices in web and API development.
G. Advanced understanding of object-oriented programming (OOP) and SOLID principles.
H. Experience with containerization. (Kubernetes & Docker) - Good to have
Shift Timings:
2 PM - 11 PM IST (Remote role)
WHO WE ARE
We are a team of digital practitioners with roots stretching back to the earliest days of online commerce, who dedicate themselves to serving our client companies.
We’ve seen the advancements first-hand over the last 25 years and believe our experiences allow us to innovate. Utilizing cutting-edge technology and providing bespoke, innovative services, we believe we can help you stay ahead of the curve.
We take a holistic view of digital strategy. Our approach to transformation is based on conscious Intent to delight customers through continuous Insight and creative Innovation with an enduring culture of problem-solving.
We bring every element together to create innovative, high-performing commerce experiences for enterprise B2C, B2B, D2C and Marketplace brands across industries. From mapping out business and functional requirements, to developing the infrastructure to optimize traditionally fragmented processes, we help you create integrated, future-proofed commerce solutions.
WHAT YOU’LL BE DOING
As part of our team, you'll play a key role in building and evolving our Integration Platform as a Service (iPaaS) solution. This platform empowers our clients to seamlessly connect systems, automate workflows, and scale integrations with modern cloud-native tools.
Here’s what your day-to-day will look like:
- Designing and Building Integrations
- Collaborate with clients to understand integration needs and build scalable, event-driven solutions using Apache Kafka, AWS Lambda, API Gateway, and EventBridge.
- Cloud-Native Development
- Develop and deploy microservices and serverless functions using TypeScript (Node.js), hosted on Kubernetes (EKS) and fully integrated with core AWS services like S3, SQS, and SNS.
- Managing Data Pipelines
- Build robust data flows and streaming pipelines using Kafka and NoSQL databases like MongoDB, ensuring high availability and fault tolerance.
- Client Collaboration
- Work directly with customers to gather requirements, design integration patterns, and provide guidance on best practices for cloud-native architectures.
- Driving Platform Evolution
- Contribute to the ongoing improvement of our iPaaS platform—enhancing observability, scaling capabilities, and CI/CD processes using modern DevOps practices.
WHAT WE NEED IN YOU
- Solid Experience in Apache Kafka for data streaming and event-driven systems
- Production experience with Kubernetes (EKS) and containerized deployments
- Deep knowledge of AWS, including S3, EC2, SQS, SNS, EventBridge, Lambda
- Proficient in TypeScript (Node.js environment)
- Experience with MongoDB or other NoSQL databases
- Familiarity with microservices architecture, async messaging, and DevOps practices
- AWS Certification (e.g., Solutions Architect or Developer Associate) is a plus
Qualification
- Graduate - BE / Btech or equivalent.
- 5 to 8 years of experience
- Self motivated and quick learner with excellent problem solving skills.
- A good team player with nice communication skills.
- Energy and real passion to work in a startup environment.
Visit our website - https://www.trikatechnologies.com


Requirement:
● Role: Fullstack Developer
● Location: Noida (Hybrid)
● Experience: 1-3 years
● Type: Full-Time
Role Description : We’re seeking a Fullstack Developer to join our fast-moving team at Velto. You’ll be responsible for building robust backend services and user-facing features using a modern tech stack. In this role, you’ll also get hands-on exposure to applied AI, contributing to the development of LLM-powered workflows, agentic systems, and custom fi ne-tuning pipelines.
Responsibilities:
● Develop and maintain backend services using Python and FastAPI
● Build interactive frontend components using React
● Work with SQL databases, design schema, and integrate data models with Python
● Integrate and build features on top of LLMs and agent frameworks (e.g., LangChain, OpenAI, HuggingFace)
● Contribute to AI fi ne-tuning pipelines, retrieval-augmented generation (RAG) setups, and contract intelligence workfl ows
● Profi ciency with unit testing libraries like jest, React testing library and pytest.
● Collaborate in agile sprints to deliver high-quality, testable, and scalable code
● Ensure end-to-end performance, security, and reliability of the stack
Required Skills:
● Proficient in Python and experienced with web frameworks like FastAPI
● Strong grasp of JavaScript and React for frontend development
● Solid understanding of SQL and relational database integration with Python
● Exposure to LLMs, vector databases, and AI-based applications (projects, internships, or coursework count)
● Familiar with Git, REST APIs, and modern software development practices
● Bachelor’s degree in Computer Science or equivalent fi eld
Nice to Have:
● Experience working with LangChain, RAG pipelines, or building agentic workfl ows
● Familiarity with containerization (Docker), basic DevOps, or cloud deployment
● Prior project or internship involving AI/ML, NLP, or SaaS products
Why Join Us?
● Work on real-world applications of AI in enterprise SaaS
● Fast-paced, early-stage startup culture with direct ownership
● Learn by doing—no layers, no red tape
● Hybrid work setup and merit-based growth
Job Description
Position Title: Senior System Engineer
Position Type: Full Time
Department: RSG
Reports to: First Level Manager, Indian Development Centre
Company Background:
Cglia is a software development company building highly available, highly secure, cloud-based enterprise software products that helps speed the research process resulting in new drugs, new devices, and new treatments to improve the health and wellbeing of world population.
At Cglia, our work shows our dedication and passion for innovative quality software products that are intuitive and easy to use and exceeds every aspect of customer expectations.
Cglia, is the place that develops world-class professionals who would like to be innovative, creative, learn continuously, and build a solid foundation to build products that are special and delight the customer.
Job Description:
The Senior System Engineer will have expertise in managing both Linux and Windows environments, along with hands-on experience in containerization technologies such as Kubernetes and Docker. Proficiency in Ansible for automation and configuration management is essential. This role is critical in ensuring the seamless operation, deployment, and maintenance of our IT infrastructure.
The ideal candidate has to oversee and participate with the installation, monitoring, maintenance, support, optimization and documentation of all network hardware and software. This includes managing multiple projects, planning network technology roadmaps and configuring/optimizing network services both internally and those integrated with Internet-based services
Job Responsibilities:
· Manage, maintain, and monitor Linux and Windows servers to ensure high availability and performance.
· Perform system upgrades, patches, and performance tuning for both operating systems and DBA servers.
· Deploy, manage, and troubleshoot containerized applications using Kubernetes and Docker.
· Design and implement Kubernetes clusters to ensure scalability, security, and reliability.
· Develop and maintain Ansible playbooks for automation of repetitive tasks, configuration management,
and system provisioning.
· Implement security best practices for both Linux and Windows environments.
· Set up and manage backup and disaster recovery solutions for critical systems and data.
· Work closely with development teams to support CI/CD pipelines and troubleshoot application issues.
· Manage VM Ware in a high availability environment with Disaster Recovery
· Good experience in RAID & Firewall
· Maintaining and managing SQL database server support
· Experience with scripting languages Unix/Shell, Bash or PowerShell
· Assist Quality Assurance with testing program changes, new releases or user documentation and support
new product release activities that include testing customer flows
· Must have the ability to work a flexible schedule and is required to participate in on-call rotation, which
includes different shift timings, weekends, and holidays
· Work across multiple time zones with remote team members
· Perform other duties as deemed necessary to provide quality service to the clients
Experience and Skills Required:
· Minimum 4+ years of experience in Linux and Windows administration
· 3 years of experience in VM Ware in a high availability environment with Disaster Recovery
· Good experience in RAID & Firewall
· 2+ years of experience in SQL database server support
· Ability to quickly acquire an in-depth knowledge of multiple custom applications
· Experience in setting up IT policies based on best practices and monitoring them
· Experience in shell scripting and automating tasks
· Experience in hardware and software monitoring tools
· Experience in administration and best practices for Apache and Tomcat
· Experience in handling Cisco router and firewall configurations and management
· Working knowledge on SQL Server, Oracle and other RDBMS databases
· Must be proactive and possess strong interpersonal, communication and organization skills
· Must possess excellent written and verbal presentation skills
· Must be self-motivated
· Certification in Linux/Windows administration is preferable.
Academics:
· Bachelor's / Master's degree (or equivalent) in computer science or related field or equivalent experience.
About AiSensy
AiSensy is a WhatsApp based Marketing & Engagement platform helping businesses like Skullcandy, Vivo, Rentomojo, Physicswallah, Cosco grow their revenues via WhatsApp.
- Enabling 100,000+ Businesses with WhatsApp Engagement & Marketing
- 400Crores + WhatsApp Messages done between Businesses and Users via AiSensy per year
- Working with top brands like Delhi Transport Corporation, Vivo, Physicswallah & more
- High Impact as Businesses drive 25-80% Revenues using AiSensy Platform
- Mission-Driven and Growth Stage Startup backed by Marsshot.vc, Bluelotus.vc & 50+ Angel Investors
Now, we’re looking for a DevOps Engineer to help scale our infrastructure and optimize performance for millions of users. 🚀
What You’ll Do (Key Responsibilities)
🔹 CI/CD & Automation:
- Implement, manage, and optimize CI/CD pipelines using AWS CodePipeline, GitHub Actions, or Jenkins.
- Automate deployment processes to improve efficiency and reduce downtime.
🔹 Infrastructure Management:
- Use Terraform, Ansible, Chef, Puppet, or Pulumi to manage infrastructure as code.
- Deploy and maintain Dockerized applications on Kubernetes clusters for scalability.
🔹 Cloud & Security:
- Work extensively with AWS (Preferred) or other cloud platforms to build and maintain cloud infrastructure.
- Optimize cloud costs and ensure security best practices are in place.
🔹 Monitoring & Troubleshooting:
- Set up and manage monitoring tools like CloudWatch, Prometheus, Datadog, New Relic, or Grafana to track system performance and uptime.
- Proactively identify and resolve infrastructure-related issues.
🔹 Scripting & Automation:
- Use Python or Bash scripting to automate repetitive DevOps tasks.
- Build internal tools for system health monitoring, logging, and debugging.
What We’re Looking For (Must-Have Skills)
✅ Version Control: Proficiency in Git (GitLab / GitHub / Bitbucket)
✅ CI/CD Tools: Hands-on experience with AWS CodePipeline, GitHub Actions, or Jenkins
✅ Infrastructure as Code: Strong knowledge of Terraform, Ansible, Chef, or Pulumi
✅ Containerization & Orchestration: Experience with Docker & Kubernetes
✅ Cloud Expertise: Hands-on experience with AWS (Preferred) or other cloud providers
✅ Monitoring & Alerting: Familiarity with CloudWatch, Prometheus, Datadog, or Grafana
✅ Scripting Knowledge: Python or Bash for automation
Bonus Skills (Good to Have, Not Mandatory)
➕ AWS Certifications: Solutions Architect, DevOps Engineer, Security, Networking
➕ Experience with Microsoft/Linux/F5 Technologies
➕ Hands-on knowledge of Database servers
➕ Experience on AWS Sagemaker, AWS Bedrock
Must be:
- Based in Mumbai
- Comfortable with Work from Office
- Available to join immediately
Responsibilities:
- Manage, monitor, and scale production systems across cloud (AWS/GCP) and on-prem.
- Work with Kubernetes, Docker, Lambdas to build reliable, scalable infrastructure.
- Build tools and automation using Python, Go, or relevant scripting languages.
- Ensure system observability using tools like NewRelic, Prometheus, Grafana, CloudWatch, PagerDuty.
- Optimize for performance and low-latency in real-time systems using Kafka, gRPC, RTP.
- Use Terraform, CloudFormation, Ansible, Chef, Puppet for infra automation and orchestration.
- Load testing using Gatling, JMeter, and ensuring fault tolerance and high availability.
- Collaborate with dev teams and participate in on-call rotations.
Requirements:
- B.E./B.Tech in CS, Engineering or equivalent experience.
- 3+ years in production infra and cloud-based systems.
- Strong background in Linux (RHEL/CentOS) and shell scripting.
- Experience managing hybrid infrastructure (cloud + on-prem).
- Strong testing practices and code quality focus.
- Experience leading teams is a plus.
Job Title : Lead Java Developer (Backend)
Experience Required : 8 to 15 Years
Open Positions : 5
Location : Any major metro city (Bengaluru, Pune, Chennai, Kolkata, Hyderabad)
Work Mode : Open to Remote / Hybrid / Onsite
Notice Period : Immediate Joiner/30 Days or Less
About the Role :
- We are looking for experienced Lead Java Developers who bring not only strong backend development skills but also a product-oriented mindset and leadership capability.
- This is an opportunity to be part of high-impact digital transformation initiatives that go beyond writing code—you’ll help shape future-ready platforms and drive meaningful change.
- This role is embedded within a forward-thinking digital engineering team that thrives on co-innovation, lean delivery, and end-to-end ownership of platforms and products.
Key Responsibilities :
- Design, develop, and implement scalable backend systems using Java and Spring Boot.
- Collaborate with product managers, designers, and engineers to build intuitive and reliable digital products.
- Advocate and implement engineering best practices : SOLID principles, OOP, clean code, CI/CD, TDD/BDD.
- Lead Agile-based development cycles with a focus on speed, quality, and customer outcomes.
- Guide and mentor team members, fostering technical excellence and ownership.
- Utilize cloud platforms and DevOps tools to ensure performance and reliability of applications.
What We’re Looking For :
- Proven experience in Java backend development (Spring Boot, Microservices).
- 8+ Years of hands-on engineering experience with at least 2+ years in a Lead role.
- Familiarity with cloud platforms such as AWS, Azure, or GCP.
- Good understanding of containerization and orchestration tools like Docker and Kubernetes.
- Exposure to DevOps and Infrastructure as Code practices.
- Strong problem-solving skills and the ability to design solutions from first principles.
- Prior experience in product-based or startup environments is a big plus.
Ideal Candidate Profile :
- A tech enthusiast with a passion for clean code and scalable architecture.
- Someone who thrives in collaborative, transparent, and feedback-driven environments.
- A leader who takes ownership beyond individual deliverables to drive overall team and project success.
Interview Process
- Initial Technical Screening (via platform partner)
- Technical Interview with Engineering Team
- Client-facing Final Round
Additional Info :
- Targeting profiles from product/startup backgrounds.
- Strong preference for candidates with under 1 month of notice period.
- Interviews will be fast-tracked for qualified profiles.
📍 Position : Java Architect
📅 Experience : 10 to 15 Years
🧑💼 Open Positions : 3+
📍 Work Location : Bangalore, Pune, Chennai
💼 Work Mode : Hybrid
📅 Notice Period : Immediate joiners preferred; up to 1 month maximum
🔧 Core Responsibilities :
- Lead architecture design and development for scalable enterprise-level applications.
- Own and manage all aspects of technical development and delivery.
- Define and enforce best coding practices, architectural guidelines, and development standards.
- Plan and estimate the end-to-end technical scope of projects.
- Conduct code reviews, ensure CI/CD, and implement TDD/BDD methodologies.
- Mentor and lead individual contributors and small development teams.
- Collaborate with cross-functional teams, including DevOps, Product, and QA.
- Engage in high-level and low-level design (HLD/LLD), solutioning, and cloud-native transformations.
🛠️ Required Technical Skills :
- Strong hands-on expertise in Java, Spring Boot, Microservices architecture
- Experience with Kafka or similar messaging/event streaming platforms
- Proficiency in cloud platforms – AWS and Azure (must-have)
- Exposure to frontend technologies (nice-to-have)
- Solid understanding of HLD, system architecture, and design patterns
- Good grasp of DevOps concepts, Docker, Kubernetes, and Infrastructure as Code (IaC)
- Agile/Lean development, Pair Programming, and Continuous Integration practices
- Polyglot mindset is a plus (Scala, Golang, Python, etc.)
🚀 Ideal Candidate Profile :
- Currently working in a product-based environment
- Already functioning as an Architect or Principal Engineer
- Proven track record as an Individual Contributor (IC)
- Strong engineering fundamentals with a passion for scalable software systems
- No compromise on code quality, craftsmanship, and best practices
🧪 Interview Process :
- Round 1: Technical pairing round
- Rounds 2 & 3: Technical rounds with panel (code pairing + architecture)
- Final Round: HR and offer discussion