50+ CI/CD Jobs in India
Apply to 50+ CI/CD Jobs on CutShort.io. Find your next job, effortlessly. Browse CI/CD Jobs and apply today!
Job Title: QA Tester – FinTech (Manual + Automation Testing)
Location: Bangalore, India
Job Type: Full-Time
Experience Required: 3 Years
Industry: FinTech / Financial Services
Function: Quality Assurance / Software Testing
About the Role:
We are looking for a skilled QA Tester with 3 years of experience in both manual and automation testing, ideally in the FinTech domain. The candidate will work closely with development and product teams to ensure that our financial applications meet the highest standards of quality, performance, and security.
Key Responsibilities:
- Analyze business and functional requirements for financial products and translate them into test scenarios.
- Design, write, and execute manual test cases for new features, enhancements, and bug fixes.
- Develop and maintain automated test scripts using tools such as Selenium, TestNG, or similar frameworks.
- Conduct API testing using Postman, Rest Assured, or similar tools.
- Perform functional, regression, integration, and system testing across web and mobile platforms.
- Work in an Agile/Scrum environment and actively participate in sprint planning, stand-ups, and retrospectives.
- Log and track defects using JIRA or a similar defect management tool.
- Collaborate with developers, BAs, and DevOps teams to improve quality across the SDLC.
- Ensure test coverage for critical fintech workflows like transactions, KYC, lending, payments, and compliance.
- Assist in setting up CI/CD pipelines for automated test execution using tools like Jenkins, GitLab CI, etc.
Required Skills and Experience:
- 3+ years of hands-on experience in manual and automation testing.
- Solid understanding of QA methodologies, STLC, and SDLC.
- Experience in testing FinTech applications such as digital wallets, online banking, investment platforms, etc.
- Strong experience with Selenium WebDriver, TestNG, Postman, and JIRA.
- Knowledge of API testing, including RESTful services.
- Familiarity with SQL to validate data in databases.
- Understanding of CI/CD processes and basic scripting for automation integration.
- Good problem-solving skills and attention to detail.
- Excellent communication and documentation skills.
Preferred Qualifications:
- Exposure to financial compliance and regulatory testing (e.g., PCI DSS, AML/KYC).
- Experience with mobile app testing (iOS/Android).
- Working knowledge of test management tools like TestRail, Zephyr, or Xray.
Location: Bangalore
Experience: 2–5 years
Type: Full-time | On-site
Open Roles: 1
Start: Immediate
Why this role exists
Most engineering teams choose between speed and stability.
We need both.
Today:
- Deployments carry risk
- Cloud costs are higher than they should be
- Compliance is reactive, not built-in
This role exists to build a platform where:
- We can deploy fast without breaking production
- We can scale without runaway cost
- We can pass enterprise InfoSec reviews without firefighting
What you’ll do
You will not just manage infrastructure.
You will build the platform that engineering runs on.
1. Drive cloud cost efficiency
- Reduce Azure compute spend by 40%
- Implement:
- Reserved Instances / savings plans
- Right-sizing of workloads
- Scheduling for non-critical workloads
- Continuously monitor and optimize cost vs performance
2. Build zero-downtime deployment systems
- Ship a deployment pipeline that supports:
- 5+ production deployments per week
- Zero customer-visible downtime
- Implement:
- Blue-green / canary deployments
- Automated health checks
- Safe rollout strategies
3. Enable fast and safe releases
- Reduce time-to-launch significantly
- Ensure:
- High reliability in every release
- Ability to rollback instantly if something breaks
- Create systems where:
- Scaling up is seamless when things go right
- Failures are contained when they don’t
4. Build disaster recovery and compliance readiness
- Create DR/BCP systems that pass enterprise audits from:
- HDFC Life, SBI Life
- Ensure:
- Backup and recovery processes are defined and tested
- Failover strategies are documented and executable
- Build compliance as part of the system, not an afterthought
5. Embed security into the pipeline
- Integrate:
- SAST (Static Application Security Testing)
- DAST (Dynamic Application Security Testing)
- SCA (Software Composition Analysis)
- Secret scanning
- Container scanning
- IaC scanning
- Ensure vulnerabilities are caught before deployment
6. Enforce policy-as-code
- Implement:
- OPA / Gatekeeper
- Azure Policy
- Prevent non-compliant infrastructure from being deployed
- Ensure consistency across environments
7. Build a scalable platform layer
- Create systems that:
- Support increasing deployment frequency
- Maintain reliability under scale
- Work closely with backend and SRE teams to:
- Improve system stability
- Reduce operational overhead
What success looks like
- Cloud costs reduce by ≥ 40%
- Deployments are:
- Frequent
- Safe
- Invisible to customers
- Rollbacks are instant and reliable
- DR/BCP passes enterprise audits in the first attempt
- Security is embedded in the pipeline, not patched later
- Engineering teams ship faster with confidence
Who you are
- You have 2-5 years of experience in DevOps / Platform Engineering
- You have worked with:
- Cloud platforms (Azure preferred)
- CI/CD systems
- Infrastructure as Code
- You think in:
- Systems
- Trade-offs (speed vs reliability vs cost)
- You are comfortable owning:
- Production infrastructure
- Deployment systems
What will make you stand out
- Experience with:
- High-frequency deployment systems
- Cost optimization at scale
- Security-first pipelines
- Strong understanding of:
- Kubernetes / container orchestration
- Monitoring and observability
- Distributed system reliability
- Experience passing enterprise security/compliance audits
Why join
- You will define how engineering ships and scales
- Your work directly impacts:
- Reliability
- Cost
- Deployment velocity
- You will build a platform that moves from:
- Fragile → predictable and scalable
What this role is not
- Not manual infra management
- Not reactive firefighting
- Not limited to CI/CD maintenance
What this role is
- A builder of deployment systems
- A driver of cost efficiency
- A guardian of reliability and compliance
One question to self-evaluate
Can you build a platform where we deploy faster, spend less, and never break production?
Job Title : Backend Engineer (AI-First | FinTech/Crypto)
Experience : 3 to 6 Years
Location : Gurugram (Sector 49)
Working Hours : 10:00 AM – 6:00 PM
Work Mode : On-site | 6 Days Working
Employment Type : Full-time
Role Overview :
This is not a typical ticket-based engineering role. You will take end-to-end ownership of systems—designing architecture, building scalable solutions, and solving real-world performance challenges.
We operate in an AI-first engineering environment, leveraging advanced tools and automation workflows to build high-performance distributed systems.
Mandatory Skills :
Java/Spring Boot or Node.js, System Design (HLD/LLD), Distributed Systems, Event-Driven Architecture (Kafka/RabbitMQ), Low-Latency APIs, PostgreSQL/MongoDB, CI/CD, Docker/Kubernetes, AI-assisted development (Copilot/Cursor/Claude)
Key Responsibilities :
- Design and build scalable backend systems (Java/Spring Boot, Node.js, or similar).
- Architect and implement event-driven systems (Kafka, RabbitMQ, pub/sub).
- Develop secure and reliable financial systems with strong data integrity.
- Solve scalability and performance challenges in fintech/crypto environments.
- Own features end-to-end: design → development → deployment → monitoring.
- Work with real-time data pipelines (WebSockets, streaming, event sourcing).
- Define service contracts and optimize system architecture.
AI-First Engineering (Must-Have Mindset) :
You will :
- Use tools like GitHub Copilot, Cursor, and Claude in daily development
- Follow spec-driven development using structured instructions
- Review, validate, and ship AI-generated code with strong engineering judgment
Core Requirements :
- 3+ years of backend development experience.
- Strong expertise in Java (Spring Boot) or Node.js.
- Solid understanding of System Design (HLD/LLD, Distributed Systems).
- Experience with event-driven architectures (Kafka, RabbitMQ, async pipelines).
- Hands-on experience building low-latency, high-throughput systems.
- Strong database knowledge (PostgreSQL, MongoDB, etc.).
- Understanding of security, performance optimization, and reliability.
- Experience with CI/CD, Git, Docker, Kubernetes.
- Exposure to React / React Native is a plus.
Good to Have (Differentiators) :
- Experience in FinTech / Crypto / Web3 / Blockchain.
- Built systems for trading, payments, or real-time financial data.
- Experience with AI agents, automation pipelines, or agent-based systems.
- Exposure to parallel AI workflows (coding / testing / refactoring).
- Contributions to open source or technical blogs.
- Experience handling production-scale systems.
Company Description
Rootflo is an innovative AI-driven company transforming finance back offices through advanced revenue intelligence and workflow automation solutions. Specializing in empowering financial institutions, Rootflo partners with leading banks, lending platforms, and insurance companies to revolutionize their operations. Leveraging the power of artificial intelligence, Rootflo provides efficient and intelligent systems that drive business outcomes. The company is dedicated to delivering impactful advancements for modern financial processes.
About the Role
We are looking for a motivated and technically sound Junior Cloud / Infrastructure Engineer with approximately 1 year of hands-on experience in designing, deploying, and managing cloud infrastructure on AWS, GCP, or Azure using Kubernetes. The candidate will work closely with senior engineers to support and maintain scalable, secure, and highly available cloud environments while gaining exposure across the full cloud stack.
Key Responsibilities
Cloud Infrastructure
- Assist in provisioning and managing cloud resources (VMs, storage, networking, databases) on AWS / GCP / Azure in Kubernetes
- Support infrastructure-as-code (IaC) implementation using Terraform
- Monitor cloud resource usage, costs, and performance; raise alerts for anomalies
- Set up log monitoring and other tracking
- Implementing cost optimizations and efficient resource utilisation
Networking & Security
- Configure VPCs, subnets, security groups, IAM roles, and access policies
- Assist in implementing firewalls, SSL/TLS certificates, and VPN connectivity
- Support compliance and security best practices (CIS benchmarks, least privilege)
CI/CD & DevOps Support
- Work with CI/CD pipelines (Jenkins, GitHub Actions, GitLab CI, or similar)
- Assist in containerization and orchestration using Docker and Kubernetes (EKS / GKE / AKS)
- Support deployment automation and version control workflows
Monitoring & Incident Response
- Use monitoring tools such as CloudWatch, Stackdriver, Azure Monitor, Grafana, or Datadog
- Respond to infrastructure alerts and assist in root cause analysis (RCA)
- Participate in on-call rotations and incident triage as required
Collaboration
- Work closely with development, QA, and product teams to support release cycles
- Maintain documentation for infrastructure, runbooks, and architecture diagrams
- Participate in code reviews for infrastructure scripts and configurations
Required Skills & Qualifications
Cloud Platforms (Any One Primary)
- AWS: EC2, S3, RDS, Lambda, VPC, IAM, ECS/EKS, CloudWatch, Route 53
- GCP: Compute Engine, GCS, Cloud SQL, GKE, Cloud Functions, VPC, IAM, Stackdriver
- Azure: VMs, Azure Storage, Azure SQL, AKS, Azure Functions, VNet, AAD, Azure Monitor
Infrastructure & Automation
- Hands-on with Terraform and/or CloudFormation / ARM Templates / GCP Deployment Manager
- Proficiency in Linux/Unix system administration
- Scripting skills in Bash, Python, or PowerShell
Containers & DevOps
- Working knowledge of Docker (build, push, run, Compose)
- Basic understanding of Kubernetes concepts (pods, deployments, services, ingress)
- Familiarity with at least one CI/CD tool
Networking Fundamentals
- Understanding of DNS, HTTP/HTTPS, TCP/IP, load balancers, CDNs
- Experience with cloud-native networking (VPCs, peering, NAT gateways)
Soft Skills
- Strong problem-solving ability and attention to detail
- Good written and verbal communication
- Eagerness to learn and adapt in a fast-paced environment
Good to Have
- Cloud certification: AWS Solutions Architect Associate / GCP ACE / Azure AZ-900 or AZ-104
- Experience with GitOps tools like ArgoCD or Flux
- Exposure to service mesh (Istio, Linkerd)
- Knowledge of Ansible or Chef/Puppet for configuration management
- Familiarity with observability tools: Prometheus, Grafana, ELK Stack
- Experience with cost optimization and FinOps practices
About BootLabs:
BootLabs is a boutique tech consulting and digital engineering company that partners with leading enterprises to build and support scalable, cloud-native, and data-driven platforms. We specialise in delivering high-impact solutions across cloud, AI/GenAI, platform engineering, and enterprise application support. Our teams work closely with global clients across highly regulated industries, including BFSI, Healthcare, E-commerce, and other domains, ensuring reliability, security, and operational excellence for mission-critical systems.
JD:
Extensive experience in deploying, and supporting enterprise workloads across Azure IaaS, PaaS, and SaaS environments.
Experience managing enterprise Landing Zones & Azure Policies.
Exposure to Azure Data Lake and Azure Databricks (basic to intermediate).
Knowledge of Terraform and any DevOps CI/CD pipelines.
Hybrid connectivity experience (VPN Gateway / ExpressRoute).
Strong networking fundamentals (NSG, UDR, Peering).
OS-level troubleshooting (Windows & Linux).
Experience with enterprise security tools such as Prisma / Cortex or similar vulnerability & endpoint security tools
Experience in production support, governance, and security guardrails.
Experience handling production incidents, RCA preparation, and performance tuning. Change management & CAB coordination.
Hands-on experience with native Azure security services like Azure Firewall, Azure Application Gateway with WAF etc
Cost optimization ,Monitoring & alerting implementation
Job Title : DevOps Engineer
Experience : 3+ Years
Location : Indiranagar, Bengaluru (Work From Office – 5 Days)
Employment Type : Full-Time
Work Timings : 11:00 AM to 7:00 PM IST
Notice Period : Immediate Joiners Preferred
Role Overview :
We are seeking a skilled DevOps Engineer with 3+ years of experience in building and managing scalable cloud-native infrastructure.
The ideal candidate will have strong expertise in Kubernetes and Helm, along with hands-on experience in deploying and maintaining production-grade systems on cloud platforms.
This role offers an opportunity to work in a high-growth startup environment, contributing to both existing systems and new infrastructure development.
Key Responsibilities :
- Design, deploy, and manage scalable infrastructure using Kubernetes.
- Build and maintain CI/CD pipelines for efficient and automated deployments.
- Manage and optimize cloud environments (preferably GCP).
- Implement Infrastructure as Code using Helm/Terraform.
- Monitor system performance and ensure high availability and reliability.
- Handle bug fixes, system improvements, and performance optimization.
- Collaborate with engineering teams to design scalable microservices architecture.
- Implement logging, monitoring, and alerting solutions.
- Ensure security best practices including IAM, secrets management, and network policies.
Mandatory Skills :
- Strong hands-on experience with Kubernetes.
- Expertise in Helm Charts.
- Experience with Google Cloud Platform (GCP).
- Hands-on experience with ArgoCD or similar CI/CD tools.
- Knowledge of CI/CD tools like Jenkins, GitHub Actions, GitLab CI.
- Experience in database hosting and scaling.
Nice to Have :
- Exposure to other cloud platforms (AWS/Azure).
- Experience with modern DevOps and automation tools.
- Ability to quickly learn and adapt to new technologies.
Team & Work Scope :
- No dedicated DevOps team currently – high ownership role.
- Work on both existing systems (maintenance & improvements) and new system builds (greenfield projects).
- Opportunity to shape DevOps practices and infrastructure from scratch.
Preferred Candidate Profile :
- 3+ years of relevant DevOps experience.
- Strong problem-solving and debugging skills.
- Experience working in fast-paced startup environments.
- Understanding of scalability, security, and performance optimization.
- Good communication and collaboration skills.
Hiring Process :
- Profile Screening
- GT Assessment
- Technical Interview – Round 1
- Technical Interview – Round 2
- Final Round (if required with US team)
Job Summary:
We are looking for a highly motivated and skilled Software Engineer to join our team.
This role requires a strong understanding of the software development lifecycle, proficiency in coding, and excellent communication skills.
The ideal candidate will be responsible for production monitoring, resolving minor technical issues, collecting client information, providing effective client interactions, and supporting our development team in resolving challenges
Key Responsibilities:
Client Interaction: Serve as the primary point of contact for client queries, provide excellent communication, and ensure timely issue resolution.
Issue Resolution: Troubleshoot and resolve minor issues related to software applications in a timely manner.
Information Collection: Gather detailed technical information from clients, understand the problem context, and relay the information to the development leads for further action.
Collaboration: Work closely with development leads and cross-functional teams to provide timely support and resolution for customer issues.
Documentation: Document client issues, actions taken, and resolutions for future reference and continuous improvement.
Software Development Lifecycle: Be involved in maintaining, supporting, and optimizing software through its lifecycle, including bug fixes and enhancements.
Automating Redundant Support Tasks: (good to have) Should be able to automate the redundant repetitive tasks Required Skills and Qualifications:
Mandatory Skills:
Expertise in at least one Object Oriented Programming language (Python, Java, C#, C++, Reactjs, Nodejs).
Good knowledge on Data Structure and their correct usage.
Open to learn any new software development skill if needed for the project.
Alignment and utilization of the core enterprise technology stacks and integration capabilities throughout the transition states.
Participate in planning, definition, and high-level design of the solution and exploration of solution alternatives.
Define, explore, and support the implementation of enablers to evolve solution intent, working directly with Agile teams to implement them.
Good knowledge on the implications.
Experience architecting & estimating deep technical custom solutions & integrations.
Added advantage:
You have developed software using web technologies.
You have handled a project from start to end.
You have worked in an Agile Development project and have experience of writing and estimating User Stories
Communication Skills: Excellent verbal and written communication skills, with the ability to clearly explain technical issues to non-technical clients.
Client-Facing Experience: Strong ability to interact with clients, gather necessary information, and ensure a high level of customer satisfaction.
Problem-Solving: Quick-thinking and proactive in resolving minor issues, with a focus on providing excellent user experience.
Team Collaboration: Ability to collaborate with development leads, engineering teams, and other stakeholders to escalate complex issues or gather additional technical support when required.
Preferred Skills:
Familiarity with Cloud Platforms and Cyber Security tools: Knowledge of cloud computing platforms and services (AWS, Azure, Google Cloud) and Cortex XSOAR, SIEM, SOAR, XDR tools is a plus.
Automation and Scripting: Experience with automating processes or writing scripts to support issue resolution is an advantage.
Job Title : UI Software Engineer
Experience : 5+ Years
Location : Pune / Hyderabad
Work Mode : Hybrid
Employment Type : Contractual (6 Months + Extendable)
🚀 About the Role :
We are looking for a UI Software Engineer to build scalable, high-performance web applications using modern frontend technologies. You will work closely with product, design, and backend teams to deliver seamless user experiences while leveraging AI-assisted development tools.
🔥 Mandatory Skills :
React + TypeScript, GraphQL (Apollo Client), Jest/React Testing Library (MSW), Webpack/CI-CD, AI tools (GitHub Copilot/Cursor), microservices integration.
🔧 Key Responsibilities :
- Develop and deliver user-facing features using React & TypeScript.
- Build reusable, scalable UI components aligned with design systems.
- Integrate frontend with GraphQL (Apollo Client) and microservices.
- Collaborate with designers to implement pixel-perfect UI from Figma.
- Ensure performance, accessibility (WCAG), and cross-browser compatibility.
- Write unit/component tests using Jest & React Testing Library.
- Maintain frontend tooling (Webpack, ESLint, Prettier, CI/CD).
- Use AI tools (GitHub Copilot, Cursor AI) to improve productivity.
- Contribute to AI-driven automation (UI testing, component generation, etc.).
🧠 Required Skills :
- Strong experience in React + TypeScript.
- Hands-on with GraphQL (Apollo Client).
- Solid knowledge of HTML, CSS, JavaScript (ES6+).
- Experience with frontend testing (Jest, RTL, MSW).
- Familiarity with microservices architecture.
- Exposure to AI-assisted development tools.
⭐ Good to Have :
- Experience with NestJS backend integration.
- Knowledge of Cypress / Playwright.
- Exposure to LLMs / AI automation / agent workflows.
- Understanding of performance optimization & design systems.
Profile - Databricks Developer
Experience- 5+ years
Location- Bangalore (On site)
PF & BGV is Mandatory
Job Description: -
* Design, build, and optimize data pipelines and ETL/ELT workflows using Databricks and
Apache Spark (PySpark).
* Develop scalable, high performance data solutions using Spark distributed processing.
* Lead engineering initiatives focused on automation, performance tuning, and platform
modernization.
* Implement and manage CI/CD pipelines using Git-based workflows and tools such as
GitHub Actions or Jenkins.
* Collaborate with cross-functional teams to translate business needs into technical
solutions.
* Ensure data quality, governance, and security across all processes.
* Troubleshoot and optimize Spark jobs, Databricks clusters, and workflows.
* Participate in code reviews and develop reusable engineering frameworks.
* Should have knowledge of utilizing AI tools to improve productivity and support daily
engineering activities.
* Strong knowledge and hands-on experience in Databricks Genie, including prompt
engineering, workspace usage, and automation.
Required Skills & Experience:
* 5+ years of experience in Data Engineering or related fields.
* Strong hands-on expertise in Databricks (notebooks, Delta Lake, job orchestration).
* Deep knowledge of Apache Spark (PySpark, Spark SQL, optimization techniques).
* Strong proficiency in Python for data processing, automation, and framework
development.
* Strong proficiency in SQL, including complex queries, performance tuning, and analytical
functions.
* Strong knowledge of Databricks Genie and leveraging it for engineering workflows.
* Strong experience with CI/CD and Git-based development workflows.
* Proficiency in data modeling and ETL/ELT pipeline design.
* Experience with automation frameworks and scheduling tools.
* Solid understanding of distributed systems and big data concepts
Key Responsibilities:-
- Design, build, and enhance Salesforce applications using Apex, Lightning Web Components (LWC), Visualforce, and SOQL.
- Implement integrations with external systems using REST APIs and event-driven messaging (e.g., Kafka).
- Collaborate with architects and business analysts to translate requirements into scalable, maintainable solutions.
- Establish and follow engineering best practices, including source control (Git), code reviews, branching strategies, CI/CD pipelines, automated testing, and environment management.
- Establish and maintain Azure DevOps-based workflows (repos, pipelines, automated testing) for Salesforce engineering.
- Ensure solutions follow Salesforce security, data modeling, and performance guidelines.
- Participate in Agile ceremonies, providing technical expertise and leadership within sprints and releases.
- Optimize workflows, automations, and data processes across Sales Cloud, Service Cloud, and custom Salesforce apps.
- Provide technical mentoring and knowledge sharing when required.
- Support production environments, troubleshoot issues, and drive root-cause analysis for long-term reliability.
- Stay current on Salesforce platform updates, releases, and new features, recommending adoption where beneficial.
Required Qualifications:-
- Bachelor’s degree in Computer Science, Engineering, or related field (or equivalent experience).
- 4+ years of Salesforce development experience with strong knowledge of Apex, Lightning Web Components, and Salesforce APIs.
- Proven experience with Salesforce core clouds (Sales Cloud, Service Cloud, or equivalent).
- Strong hands-on experience with API integrations (REST/SOAP) and event-driven architectures (Kafka, Pub/Sub).
- Solid understanding of engineering practices: Git-based source control (Salesforce DX/metadata), branching strategies, CI/CD, automated testing, and deployment management.
- Familiarity with Azure DevOps repositories and pipelines.
- Strong knowledge of Salesforce data modeling, security, and sharing rules.
- Excellent problem-solving skills and ability to collaborate across teams.
Preferred Qualifications:-
- Salesforce Platform Developer II certification (or equivalent advanced credentials).
- Experience with Health Cloud, Financial Services Cloud, or other industry-specific Salesforce products.
- Experience implementing logging, monitoring, and observability within Salesforce and integrated systems.
- Background in Agile/Scrum delivery with strong collaboration skills.
- Prior experience establishing or enforcing engineering standards across Salesforce teams.
Role - IT Cloud Engineer/ Devops
- Proficient in Linux.
- Hands on experience with AWS cloud or Google Cloud.
- Knowledge of container technology like Docker.
- Expertise in scripting languages. (Shell scripting or Python scripting)
- Working knowledge of LAMP/LEMP stack, networking and version control system like Gitlab or Github.
Job Description:
The incumbent would be responsible for:
- Deployment of various infrastructures on Cloud platforms like AWS, GCP, Azure, OVH etc.
- Server monitoring, analysis and troubleshooting.
- Deploying multi-tier architectures using microservices.
- Integration of Container technologies like Docker, Kubernetes etc as per application requirement.
- Automating workflow with python or shell scripting.
- CI and CD integration for application lifecycle management.
- Hosting and managing websites on Linux machines.
- Frontend, backend and database optimization.
- Protecting operations by keeping information confidential.
- Providing information by collecting, analyzing, summarizing development & service issues.
- Prepares & installs solutions by determining and designing system specifications, standards & programming.
Dear Candidates,
We have an urgent requirement for a Technical Lead – Full Stack role based in Bangalore. Please find the details below:
Work Location (WFO):
Nagar, Bengaluru, Karnataka
Interview Process:
L1 Interview – Face-to-Face at Office
Experience Required:
4-6 Years (Minimum1+ years in Technical Leadership role)
Role Overview:
The candidate will lead the technical vision and architecture of a compliance platform by designing scalable, secure, and high-performance systems. The role involves driving full-stack development across .NET and open-source technologies, enabling unified AI Agent capabilities, Single Authentication (SSO), and a One-UI experience.
Key Responsibilities:
- Define and own end-to-end architecture including micro-frontends, .NET services, FastAPI APIs, and microservices
- Lead full-stack development using .NET and modern open-source technologies
- Modernize legacy systems (ASP.NET, .NET Core, MS SQL Server) to cloud-native architecture
- Design and implement AI Agents, SSO, and unified UI experiences
- Manage sprint planning, backlogs, and collaborate with Product Owners
- Implement CI/CD pipelines using Jenkins, GitHub Actions
- Drive containerization and orchestration using Docker & Kubernetes
- Ensure secure deployments and cloud infrastructure management
- Establish engineering best practices, code reviews, and architecture governance
- Mentor teams on Clean Architecture, SOLID principles, and DevOps practices
Required Skills:
- ReactJS, FastAPI, Python, REST/GraphQL
- ASP.NET, MVC, .NET Core, Entity Framework, MS SQL Server
- Strong experience in Microservices Architecture
- DevOps: CI/CD, Jenkins, GitOps, Docker, Kubernetes
- Cloud Platforms: AWS / Azure / GCP
- AI/ML & LLM tools: OpenAI, Llama, LangChain, etc.
- Security: RBAC, API security, secrets management
Qualifications:
- BE / BTech in Computer Science
Summary:
We are seeking a highly skilled Python Backend Developer with proven expertise in FastAPI to join our team as a full-time contractor for 12 months. The ideal candidate will have 5+ years of experience in backend development, a strong understanding of API design, and the ability to deliver scalable, secure solutions. Knowledge of front-end technologies is an added advantage. Immediate joiners are preferred. This role requires full-time commitment—please apply only if you are not engaged in other projects.
Job Type:
Full-Time Contractor (12 months)
Location:
Remote
Experience:
3+ years in backend development
Key Responsibilities:
- Design, develop, and maintain robust backend services using Python and FastAPI.
- Implement and manage Prisma ORM for database operations.
- Build scalable APIs and integrate with SQL databases and third-party services.
- Deploy and manage backend services using Azure Function Apps and Microsoft Azure Cloud.
- Collaborate with front-end developers and other team members to deliver high-quality web applications.
- Ensure application performance, security, and reliability.
- Participate in code reviews, testing, and deployment processes.
Required Skills:
- Expertise in Python backend development with strong experience in FastAPI.
- Solid understanding of RESTful API design and implementation.
- Proficiency in SQL databases and ORM tools (preferably Prisma)
- Hands-on experience with Microsoft Azure Cloud and Azure Function Apps.
- Familiarity with CI/CD pipelines and containerization (Docker).
- Knowledge of cloud architecture best practices.
Added Advantage:
- Front-end development knowledge (React, Angular, or similar frameworks).
- Exposure to AWS/GCP cloud platforms.
- Experience with NoSQL databases.
Eligibility:
- Minimum 3 years of professional experience in backend development.
- Available for full-time engagement.
- Please excuse if you are currently engaged in other projects—we require dedicated availability.
Job Title: Snowflake Platform Administrator
Duration: 6-12 months contract (could be extended upon performance)
Mode: Remote
About the Role
We are looking for a Snowflake Administrator to join our Snowflake Center of Excellence (COE) to manage, secure, and optimize the enterprise Snowflake data platform. The role will focus on platform administration, security governance, and automation while enabling data engineering, analytics, and business teams to effectively leverage Snowflake capabilities.
Key Responsibilities
•Administer and maintain the Snowflake platform, including warehouses, databases, schemas, users, roles, and resource monitors.
•Implement and manage Snowflake security and access governance including RBAC, network policies, and network rules.
•Manage identity and access integration with Azure Active Directory (Azure AD), including role mapping with Azure AD groups.
•Monitor platform performance, usage, and cost to ensure efficient and reliable operations.
•Manage key Snowflake capabilities including data sharing (consumer and provider), cloning, data recovery, integrations (storage/API/notification), and performance optimization.
•Develop automation scripts using SQL and Python for administrative and operational tasks.
•Create and maintain CI/CD workflows using GitHub Actions for Snowflake deployments.
•Collaborate with data engineers, analysts, and architects to ensure secure and scalable data platform usage.
•Stay up to date with Snowflake product releases, new features, and platform best practices, and proactively evaluate their applicability to the organization.
•Contribute to standards, best practices, and governance frameworks within the Snowflake COE.
General Business
•Explore opportunities to leverage AI to improve platform automation and productivity.
Required Experience & Skills:
•5-8 years of relevant experience in Snowflake Administration and platform management.
•Solid understanding of Snowflake architecture, security, features, and performance optimization.
•Experience implementing RBAC, Network Policies, and Network Rules in Snowflake.
•Experience with Snowflake integration with Azure AD for role and access management via AD groups.
•Proficiency in SQL and Python scripting.
•Experience with GitHub and GitHub Actions/Workflow creation.
•Strong analytical and problem-solving skills.
•Functional Domain: FMCH (Fast Moving Consumer Health).
Preferred Additional Skills:
•AI enthusiast and Automation expertise
•Understanding of modern data architectures including data lakes and real-time processing
•Familiarity with BI tools such as Power BI, Tableau, Looker
Education & Languages
•Bachelor’s degree in computer science, Information Technology, or similar quantitative field of study.
•Fluent in English.
•Function effectively within teams of varied cultural backgrounds and expertise sources.
The DevOps Engineer will play a critical role in operationalizing artificial intelligence across Bell Techlogix client environments. This role focuses on building and supporting cloud infrastructure, CI/CD pipelines, and automation frameworks that power AI and machine learning workloads. The ideal candidate has experience supporting AI platforms such as Azure AI, Azure Machine Learning, Azure OpenAI, and ServiceNow or conversational AI platforms, and understands the operational requirements of production AI systems, including reliability, scalability, and security.
Key Responsibilities
•Design, build, and operate cloud infrastructure and platform services that support AI and machine learning workloads in production, SLA-driven managed services environments
•Implement CI/CD and MLOps pipelines to enable automated training, testing, deployment, and rollback of AI and ML models
•Develop and maintain Infrastructure as Code to provision AI-ready environments consistently across dev/test/prod
•Support AI platform operations including monitoring model health, pipeline execution, compute utilization, and data dependencies
•Partner with Machine Learning Engineers and Data Engineers to standardize deployment patterns for AI services and LLM-based solutions
•Enable secure and scalable AI integrations using APIs, messaging, and event-driven architectures
•Implement observability solutions for AI platforms, including logging, metrics, alerting, and drift detection integrations
•Troubleshoot AI platform incidents, perform root cause analysis, and implement remediation to improve reliability and automation coverage
•Apply security best practices for AI environments including secrets management, identity and access controls, network isolation, and policy enforcement
•Support AI-driven automation use cases across platforms such as Microsoft Copilot, ServiceNow, and conversational AI tools
•Collaborate with service desk, security, and architecture teams to continuously improve AI service delivery and operational maturity
Required Qualifications
•Bachelor’s degree in Computer Science, Engineering, or equivalent practical experience
•5+ years of experience in DevOps, cloud engineering, or platform operations, with exposure to AI or data workloads
•Hands-on experience with Microsoft Azure, including compute, networking, storage, and monitoring services
•Experience building CI/CD pipelines using Azure DevOps, GitHub Actions, or similar tools
•Working knowledge of Infrastructure as Code (Terraform and/or Bicep/ARM)
•Scripting experience using PowerShell and/or Python
•Experience supporting production platforms with incident management, change control, and root cause analysis
•Understanding of cloud security fundamentals and enterprise governance requirements
Preferred Qualifications
•Experience with Azure Machine Learning, Azure AI Services, Azure OpenAI, or MLOps frameworks
•Exposure to containerization and orchestration technologies (Docker, Kubernetes, AKS)
•Experience supporting data pipelines or feature stores used by machine learning systems
•Familiarity with ServiceNow, AI-driven ITSM workflows, or automation platforms
•Experience with observability tools
•Knowledge of Responsible AI, data governance, and compliance considerations for AI systems
•Relevant certifications (Microsoft Azure Administrator, Azure DevOps Engineer, Azure AI Engineer)
👉 Job Title: Senior Backend Developer
🌟 Experience: 5-7 Years
💡Location: Bangalore
👉 Notice Period :- Immediate Joiners
💡 Work Mode - 5 Days work from Office
( Candidate Serving notice period are preffered)
Role Summary
We are seeking a Senior Backend Developer with strong expertise in Python and FastAPI to build scalable, high-performance backend systems integrated with LLM technologies on Azure. The role involves designing distributed systems, optimizing data pipelines, and ensuring secure, enterprise-grade applications.
Key Responsibilities
- Develop backend services using Python & FastAPI (async, middleware)
- Build high-concurrency, scalable systems and microservices
- Work with Azure services and event-driven architectures
- Optimize MongoDB & Redis for performance
- Integrate LLM APIs (OpenAI, Gemini, Claude)
- Implement security (JWT, encryption, API management)
Mandatory Skills (Top 3)
- Strong Python backend development with FastAPI
- Hands-on experience with Microsoft Azure cloud
- Experience in building scalable distributed/microservices systems
Good to Have
- Docker, Kubernetes, CI/CD
- LLM frameworks (LangChain, vector DBs)
- Monitoring tools and real-time data processing
Role: Backend Developer
Experience: 3-6 years
Duration: 1–3 months
Tech Stack: .NET Core, MongoDB
Domain: Clinical/Healthcare (Preferred)
Responsibilities
- Develop and maintain REST APIs using .NET Core
- Design scalable backend services and microservices
- Implement data storage and retrieval using MongoDB
- Optimize database queries and application performance
- Integrate with external systems (APIs, FHIR etc.)
- Ensure data security and compliance (GxP/FDA preferred)
- Support debugging and production issue resolution
Required Skills
- Strong experience in .NET Core/ASP.NET Core
- Hands-on experience with MongoDB or NoSQL databases
- Experience in REST API development
- Knowledge of microservices architecture
- Familiarity with Git and CI/CD pipelines
Preferred
- Healthcare/Clinical Trial domain exposure
- Experience with Azure cloud deployments
- Understanding of FHIR or Healthcare Data Standards
👉 Job Title: Backend Developer
🌟 Experience: 5-7 Years
💡Location: Bangalore
👉 Notice Period :- Immediate Joiners
💡 Work Mode - 5 Days work from Office
( Candidate Serving notice period are preffered)
Role Summary
We are looking for a Backend Engineer to join the Platform Implementation Team, responsible for building scalable, secure, and high-performance backend systems for a multi-cloud Data & AI platform. You will design microservices, develop REST APIs, and enable seamless data integration across enterprise systems like CRM and ERP.
💫 Key Responsibilities
✅ Design and develop scalable microservices and RESTful APIs
✅ Build event-driven architectures for asynchronous processing
✅ Integrate backend systems with cloud platforms (GCP/Azure)
✅ Ensure secure, reliable, and optimized data handling
✅ Collaborate with cross-functional teams (UI, Data, Platform)
✅ Follow best practices in coding, testing, CI/CD, and containerization
💫 Mandatory Skills (Top 3)
✅ Strong backend programming experience (Python / Node.js / Java)
✅ Expertise in API development & Microservices architecture
✅ Hands-on experience with Cloud platforms (GCP or Azure)

A boutique software product engineering services company
Job Description :
Responsibilities :
- Design and develop Python-based microservices
- Build and operate gRPC / Protobuf-based APIs
- Implement asynchronous processing, concurrency, and job orchestration
- Design systems with retries, idempotency, and fault tolerance
- Work with and integrate native C/C++ components with Python services
- Design and optimise PostgreSQL schemas and queries
- Contribute to React-based frontend applications (TypeScript/JavaScript)
- Own features end-to-end : design, development, deployment, and monitoring
- Debug issues across application, system, and performance layers
- Build and maintain CI/CD pipelines and automated tests
Requirements :
- Strong experience in Python backend development (sync + async)
- Hands-on experience with gRPC / Protobuf-based APIs
- Experience with FastAPI / Flask / Django
- Strong understanding of microservices and distributed systems
- Experience with PostgreSQL and data modeling
- Exposure to React / JavaScript / TypeScript
- Knowledge of concurrency, multi-threading, and system design
- Strong understanding of Linux systems and debugging
- Experience in production environments (performance tuning, issue resolution)
- Exposure to C/C++ or Python-native integrations (preferred)
Qualification :
- Engineering Graduates from Tier 1 & Tier 2 Colleges/Deemed Universities only
- Open for candidates from outstation.
- Experience range : 12 - 15 years (but not more than 15 years of experience)
- Strictly Individual contributors, Handson coding. 80% Individual Contributor & 20% Architecture, Design, Systems

at Prismberry Technologies Pvt Ltd (acquired by eYantra Ventures Limited)
Role Overview
We are seeking a highly skilled Senior Java Developer to join our engineering team. In this role, you will be responsible for designing and developing high-performance, scalable, and resilient microservices. You will work at the intersection of complex backend logic, real-time data streaming, and cloud infrastructure to deliver seamless user experiences.
Key Responsibilities
- System Design: Architect and develop robust, scalable, and maintainable backend services using Java and Spring Boot.
- Scalability: Build distributed systems capable of handling high traffic and large datasets with low latency.
- Database Management: Design and optimize complex schemas in both Relational (SQL) and NoSQL databases, ensuring data integrity and performance.
- Event-Driven Architecture: Implement real-time messaging and data pipelines using Apache Kafka.
- Cloud Infrastructure: Deploy and manage services on cloud platforms (AWS or GCP), leveraging managed services to improve system reliability.
- Collaboration: Work closely with cross-functional teams to define requirements, participate in code reviews, and mentor junior developers.
Technical Requirements
- Core Java: Deep expertise in Java (8 or higher), including concurrency, multithreading, and JVM tuning.
- Frameworks: Strong experience with Spring Boot, Spring Cloud, and Hibernate/JPA.
- Messaging: Proven experience with Apache Kafka for event streaming and asynchronous processing.
- Cloud: Proficiency in AWS (EC2, S3, RDS, Lambda) or GCP (GCE, GCS, Cloud SQL, Pub/Sub).
- Databases: Solid knowledge of PostgreSQL, MySQL, or Oracle, alongside NoSQL experience (e.g., MongoDB, Cassandra, or Redis).
- DevOps & Tools: Familiarity with Docker, Kubernetes, and CI/CD pipelines (Jenkins, GitLab CI, or GitHub Actions).
Preferred Qualifications
- Experience with Microservices Architecture and Domain-Driven Design (DDD).
- Understanding of distributed caching strategies and load balancing.
- Strong problem-solving skills and a "clean code" mentality.
Role & Responsibilities
- Develop and deliver automation software to build and improve platform functionality
- Ensure reliability, availability, and manageability of applications and cloud platforms
- Champion adoption of Infrastructure as Code (IaC) practices
- Design and build self-service, self-healing, monitoring, and alerting platforms
- Automate development and testing workflows through CI/CD pipelines (Git, Jenkins, SonarQube, Artifactory, Docker containers)
- Build and manage container hosting platforms using Kubernetes
Requirements
- Strong experience deploying and maintaining GCP cloud infrastructure
- Well-versed in service-oriented and cloud-based architecture design patterns
- Knowledge of cloud services including compute, storage, networking, messaging, and automation tools (e.g., CloudFormation/Terraform equivalents)
- Experience with relational and NoSQL databases (Postgres, Cassandra)
- Hands-on experience with automation/configuration tools (Puppet, Chef, Ansible, Terraform)
Additional Skills
- Strong Linux system administration and troubleshooting skills
- Programming/scripting exposure (Bash, Python, Core Java, or Scala)
- CI/CD pipeline experience (Jenkins, Git, Maven, etc.)
- Experience integrating solutions in multi-region environments
- Familiarity with Agile/Scrum/DevOps methodologies
Profile - Databricks Developer
Experience- 5+ years
Location- Bangalore (On site)
PF & BGV is Mandatory
Job Description: -
* Design, build, and optimize data pipelines and ETL/ELT workflows using Databricks and Apache Spark (PySpark).
* Develop scalable, high performance data solutions using Spark distributed processing.
* Lead engineering initiatives focused on automation, performance tuning, and platform modernization.
* Implement and manage CI/CD pipelines using Git-based workflows and tools such as GitHub Actions or Jenkins.
* Collaborate with cross-functional teams to translate business needs into technical solutions.
* Ensure data quality, governance, and security across all processes.
* Troubleshoot and optimize Spark jobs, Databricks clusters, and workflows.
* Participate in code reviews and develop reusable engineering frameworks.
* Should have knowledge of utilizing AI tools to improve productivity and support daily engineering activities.
* Strong knowledge and hands-on experience in Databricks Genie, including prompt engineering, workspace usage, and automation
. Required Skills & Experience:
* 5+ years of experience in Data Engineering or related fields.
* Strong hands-on expertise in Databricks (notebooks, Delta Lake, job orchestration).
* Deep knowledge of Apache Spark (PySpark, Spark SQL, optimization techniques).
* Strong proficiency in Python for data processing, automation, and framework development.
* Strong proficiency in SQL, including complex queries, performance tuning, and analytical functions.
* Strong knowledge of Databricks Genie and leveraging it for engineering workflows.
* Strong experience with CI/CD and Git-based development workflows. * Proficiency in data modeling and ETL/ELT pipeline design.
* Experience with automation frameworks and scheduling tools.
* Solid understanding of distributed systems and big data concepts
Senior Project Owner / Project Manager Technology
Department - Technology / Software Development
Work Mode - Work From Home (WFH), Full Time
Experience - Minimum 10 Years (Development Background)
Time Zone - Candidate should be comfortable working in US time zone overlap and attending client calls accordingly.
ROLE SUMMARY
We are looking for a seasoned Senior Project Owner / Project Manager with a strong development foundation to lead our technology initiatives. This role bridges client management and technical execution you will own endto-end delivery of multiple concurrent projects while supporting a high-performing remote team.
KEY RESPONSIBILITIES
Project & Delivery Management
- Own and manage multiple concurrent technology projects from initiation to production release
- Define project scope, timelines, milestones, and resource allocation plans
- Distribute tasks effectively across a team of developers, QA, and support engineers
- Track assigned work daily, follow up on progress, and proactively remove blockers
- Ensure all projects meet deadlines and quality benchmarks without compromise
- Participate actively in production activities and take full accountability for live deployments
US Client Management
- Serve as the Technology single point of contact for all assigned US clients
- Attend and lead client calls that are focused on an ARDEM Technical Solution. This may include discussions related to future clients or existing clients (US time zone overlap required)
- Resolve client queries, manage escalations, and ensure high client satisfaction
- Showcase company-developed applications and software demos confidently to clients
- Translate complex client requirements into clear technical deliverables for the team
Team Leadership
- Lead, mentor, and performance-manage a distributed remote team of technical members
- Foster accountability, ownership, and a high-delivery culture within the team
- Conduct sprint planning, stand-ups, retrospectives, and performance reviews
- Identify skill gaps and work with HR/training teams to bridge them
Process & Operations
- Deeply understand ARDEM's internal processes and align project execution accordingly
- Ensure development standards and best practices are followed across all projects
- Manage crisis situations with composure, identify root causes and drive swift resolution
- Coordinate with cross-functional teams including HR, Operations, Training, and QA
- Maintain project documentation, status reports, and risk registers
REQUIRED EXPERIENCE
- 10+ years of total experience in software development and project management
- 5–7 years of hands-on coding experience in one or more technologies listed below
- 2–3 years in a team management or tech lead role overseeing 5+ members
- Proven experience managing multiple simultaneous projects in a remote/WFH environment
- Prior experience working with US-based clients strong understanding of US work culture and expectations
TECHNICAL SKILLS
- Python: scripting, automation, data processing, backend services
- JavaScript / Node.js: server-side development, REST APIs, async workflows
- NET Core: enterprise application development and service integration
- SQL Databases: query optimization, schema design, stored procedures
- Familiarity with CI/CD pipelines, Git workflows, and deployment processes
- Ability to review code, understand architectural decisions, and guide the team technically
SKILLS & COMPETENCIES
- Exceptional verbal and written communication skills in English client-facing confidence is a must
- Strong crisis management and conflict resolution ability under tight deadlines
- Highly organized with a structured approach to planning, prioritization, and execution
- Self-driven and accountable capable of operating independently in a remote environment
- Strong presentation skills able to demo software to non-technical stakeholders
- Empathetic leadership style with the ability to motivate and align diverse team members
QUALIFICATIONS
- Bachelor's or master's degree in computer science
- PMP Certification: Preferred (candidates without PMP must demonstrate equivalent project management rigor)
- Agile / Scrum certifications (CSM, PMI-ACP) are an added advantage
LOCATION PREFERENCE
- Candidates must be based in a Tier-1 city: Mumbai, Delhi NCR, Bengaluru, Hyderabad, Chennai, Pune, or Kolkata
- This is a full-time Work From Home role: reliable internet, a dedicated workspace, and availability during US business hours are mandatory
ABOUT ARDEM
ARDEM Incorporated is a leading Business Process Outsourcing (BPO) and Automation company serving US based clients across diverse industries. Our Technology Team builds and maintains in-house applications that power data processing pipelines, automation workflows, internal platforms, and domain-specific training modules all engineered to deliver operational excellence at scale. To our clients, we provide cloud-based platforms to assist in their day-to-day business analytics. Our cloud services focus on finance, logistics and utility management.
Company Overview:
Planview has one mission: to build the future of connected work with market-leading portfolio management and work management solutions. Planview is a recognized innovator and industry leader, our solutions enable organizations to connect the business from ideas to impact, empowering companies to accelerate the achievement of what matters most. Our solutions span every class of work, resource, and organization to address the varying needs of diverse and distributed teams, departments, and enterprises.
As a Sr CloudOps Engineer II, you will oversee teams of Engineers and be a champion for configuration management, technologies in the cloud, and continuous improvement. You will work closely with global leaders to ensure that our applications, infrastructure, and processes are scalable, secure, and supportable. By leveraging your production experience and development skills you will work hand in hand with Engineers (Dev, DevOps, DBOps) to design and implement solutions that improve delivery of value to customers, reduce costs, and eliminate toil.
Responsibilities (What you will do):
- Guide the professional development of Engineers and support the teams to accomplish business goals
- Work closely with leaders in the Israel to align on priorities and architect, deliver, and manage our products
- Build systems that are secure, scalable, and self-healing.
- Manage and improve deployment pipelines.
- Triage and remediate production issues.
- Participate in on-call rotations for escalations.
Qualifications (What you will bring):
- Bachelor's degree is CS or equivalent experience in related field.
- 2+ years managing Engineering teams.
- 8+ years of experience as a site reliability or platform engineer, preferably in a fast-scaling environment
- 5+ years administering Linux and Windows environments.
- 3+ years programming / scripting experience (e.g., Python, JavaScript, PowerShell)
- Strong technical knowledge in OS’s (Linux and Windows), virtualizations, storage systems, networking, and firewall implementations
- Maintaining production environments in the On Premise (90%) and Cloud (10%) (e.g., AWS, Google Cloud, Azure)
- Solid understanding of networking principles and how it applies to data flow and security.
- Automating deployments of cloud based available services (e.g., AWS EC2 / RDS, Docker, Kubernetes)
- Experience managing CI/CD infrastructures, with a strong proficiency in platforms like bitbucket and Jenkins to streamline deployment pipelines and ensure efficient software delivery.
- Management of resources using Infrastructure as Code tools (e.g., CloudFormation, Terraform, Chef)
- Knowledge of observability tools such as LogicMonitor, New Relic, Prometheus, and Coralogix, as well as their implementation.
- Worked within Agile and Lean software development teams.
- Experience working in globally distributed teams.
- Ability to look on the big picture and manage risks.
What are we looking for?
- You have a good understanding and work experience in AKS, Kubernetes, and EKS.
- You are able to manage multi region clusters for disaster recovery.
- You have a good understanding of AWS stack.
- You have experience of production level in Kubernetes.
- You are comfortable coding/programming and can do so whenever required.
- You have worked with programmable infrastructure in some way - Built a CI/CD pipeline, Provisioned infrastructure programmatically or Provisioned monitoring and logging infrastructure for large sets of machines.
- You love automating things, sometimes even what seems like you can’t automate - such as one of our engineers used Ansible to set up the Ubuntu workstation and runs a playbook every time something has to be installed.
- You don’t throw around words such as “high availability” or “resilient systems” without understanding at least their basics. Because you know that words are easy to talk about but there is a fair amount of work to build such a system in practice.
- You love coaching people - about the 12-factor apps or the latest tool that reduced your time of doing a task by X times and so on. You lead by example when it comes to technical work and community.
- You understand the areas you have worked on very well but, you are curious about many systems that you may not have worked on and want to fiddle with them.
- You know that understanding applications and the runtime technologies gives you a better perspective - you never looked at them as two different things.
What you will be learning and doing?
- You will be working with customers trying to transform their applications and adopt cloud-native technologies. The technologies used will be Kubernetes, Prometheus, Service Mesh, Distributed tracing and public cloud technologies or on-premise infrastructure.
- The problems and solutions are continuously evolving in space but fundamentally you will be solving problems with simplest and scalable automation.
- You will be building open source tools for problems that you think are common across customers and industry. No one ever benefited from re-inventing the wheel, did they?
- You will be hacking around open source projects, understand their capabilities, limitations and apply the right tool for the right job.
- You will be educating the customers - from their operations engineers to developers on scalable ways to build and operate applications in modern cloud-native infrastructure.
🔹 Role: Python Engineer – Python & MLOps
📍 Location: Bellandur, Bangalore
🕐 Work Timings: 01:30 PM – 10:30 PM
🏢 Work Mode: Monday (WFH), Tuesday–Friday (WFO)
📅 Experience: 8-12 Years (Ideal: 8-10 Years)
🔹 Role Overview
This role focuses on building and maintaining a production-grade AI/ML platform. You will work on scalable Python systems, MLOps pipelines, APIs, and CI/CD workflows in an enterprise environment.
🔹 Key Responsibilities
✔ Develop production-grade Python applications using OOP principles
✔ Build and enhance MLOps pipelines (training, validation, deployment)
✔ Design and optimize REST APIs with OpenAI/Swagger
✔ Implement async programming for high-performance systems
✔ Work on CI/CD pipelines (Azure Pipelines / GitHub Actions)
✔ Ensure clean, testable, and maintainable code (PyTest, TDD)
🔹 Required Skills
✔ Strong Python (OOP, modular design)
✔ MLOps & CI/CD pipeline experience
✔ REST API development
✔ Async programming (async/await, concurrency)
✔ Pandas / Polars & Scikit-learn
✔ JSON Schema–driven development
✔ Testing using PyTest
🔹 Nice to Have
➕ Azure ML SDK
➕ Pydantic
➕ Azure Cosmos DB
➕ Experience with large enterprise platforms
Must-Have Skills:
- Hands-on experience with airgap Kubernetes clusters, ideally in regulated industries (finance, healthcare, etc.).
- Strong expertise in CI/CD pipelines, programmable infrastructure, and automation.
- Proficiency in Linux troubleshooting, observability (Prometheus, Grafana, ELK), and multi-region disaster recovery.
- Security & compliance knowledge for regulated industries.
- Preferred: Experience with GKE, RKE, Rook-Ceph and certifications like CKA, CKAD.
Who You Are
- A Kubernetes expert who thrives on scalability, automation, and security.
- Passionate about optimizing infrastructure, CI/CD, and high-availability systems.
- Comfortable troubleshooting Linux, improving observability, and ensuring disaster recovery readiness.
- A problem solver who simplifies complexity and drives cloud-native adoption.
What You’ll Do
- Architect & automate Kubernetes solutions for airgap and multi-region clusters.
- Optimize CI/CD pipelines & cloud-native deployments.
- Work with open-source projects, selecting the right tools for the job.
- Educate & guide teams on modern cloud-native infrastructure best practices.
- Solve real-world scaling, security, and infrastructure automation challenges.
Why Join Us?
- Work on high-impact Kubernetes projects in regulated industries.
- Solve real-world automation & infrastructure challenges with cutting-edge tools.
- Grow in a team that values learning, open-source contributions, and innovation.
Job role: Systems Engineer (L2)
Location: Remote/Bengaluru
Experience: 3-6 years
About the Role:
We are looking for a Systems Engineer (L2) to join our growing infrastructure team. You will be responsible for managing, optimizing, and scaling our cloud communication platform that handles billions of messages and voice calls annually.
Key Responsibilities:
— Design, deploy, and maintain scalable cloud infrastructure — AWS/GCP/Azure.
— Manage and optimize networking components — routers, switches, firewalls, load balancers.
— Handle incident response — monitor systems, identify issues, resolve production problems.
— Implement DevOps best practices — CI/CD pipelines, automation, containerization.
— Collaborate with backend and product teams on system architecture.
— Performance tuning — ensure high availability and reliability of platform.
— Security management — implement security protocols and compliance standards.
Required Skills:
Technical:
- Linux/Unix administration — strong fundamentals
- Networking — TCP/IP, DNS, BGP, VoIP protocols
- Cloud platforms — AWS/GCP/Azure — minimum 2 years
- DevOps tools — Docker, Kubernetes, Jenkins, CI/CD
- Monitoring tools — Grafana, Prometheus, Kibana, Datadog
- Scripting — Python, Bash, Shell
- Databases — MySQL, PostgreSQL, Redis
Soft skills:
- Strong problem-solving under pressure
- Good communication — English written and verbal
- Team player — collaborative mindset
Good to Have:
- Experience in telecom/CPaaS/cloud communications industry
- Knowledge of VoIP, SIP, RTP protocols
- AI/ML operations experience
- CCNA/AWS certifications
Location: PAN India
💼 Employment Type: Full-Time / Contract
👨💻 Experience: 5–10 Years
🏢 Work Mode: Hybrid / Onsite
🔍 Job Overview
We are looking for a skilled SDET – Mobile Automation Engineer with strong expertise in Appium, Selenium, and Java-based automation frameworks. The ideal candidate will be responsible for building scalable automation solutions for mobile (Android & iOS) and web applications in an Agile environment.
🔑 Key Responsibilities
Design, develop, and maintain automation frameworks for mobile and web applications
Perform mobile automation testing using Appium (Android & iOS)
Implement UI and API automation using Selenium & Rest Assured
Work on BDD frameworks (Cucumber, Gherkin) for in-sprint automation
Integrate automation suites with CI/CD pipelines (Jenkins, Git)
Execute tests across multiple browsers and devices
Collaborate with cross-functional teams in Agile/SAFe environment
Debug, troubleshoot, and optimize test automation scripts
🛠 Must-Have Skills
Strong experience in Core Java & Maven
Hands-on experience with Selenium WebDriver & Selenium Grid
Expertise in Appium for Android & iOS automation
Understanding of XCTest & Espresso frameworks
Experience in testing hybrid & native mobile applications
Hands-on with Cucumber (BDD), JUnit, Rest Assured
Experience with Git, JIRA, Jenkins
Strong knowledge of REST APIs, HTTP, JSON, API workflows
Experience with BrowserStack / SauceLabs / AWS Device Farm
Strong communication & problem-solving skills
⭐ Good to Have
Domain knowledge in Retail Banking
Experience in responsive UI testing
Knowledge of API tools (Postman, Rest Client, Anypoint)
Understanding of Service Virtualization / Stubs
📩 Interested candidates can share their profiles for immediate consideration!
JobTitle : Mission Control Software Engineer – Java Backend
Experience : 4+ Years
Location : Bengaluru, Karnataka (On-site)
Employment Type : Full-time
Industry : Space Technology
🧠 Role Overview :
As a Mission Control Software Engineer – Java Backend, you will be a founding member of the Mission Control Software team. You will design and develop the backend core of a distributed, high-reliability system responsible for streaming, processing, and routing satellite data.
This role offers high ownership and the opportunity to directly influence system architecture as it evolves from a single-node deployment to a scalable multi-node system.
🛠️ Tech Stack :
Java 11+, gRPC, REST APIs, PostgreSQL, Docker, Linux, WebSockets, Git, CI/CD
⚙️ Key Responsibilities :
- Design, develop, and maintain backend services in Java (11+) for data ingestion, processing, and control workflows
- Build and enhance gRPC and REST APIs for internal and external system integrations
- Develop modular plugins and middleware for routing, persistence, and automation
- Work with PostgreSQL and other databases for structured and time-series data
- Collaborate with frontend teams to support real-time dashboards via WebSockets/event streaming
- Ensure system scalability, reliability, and observability through logging, metrics, and testing
- Contribute to containerized deployments (Docker) and CI/CD pipelines
- Write clean, maintainable, and well-documented code with strong test coverage
🧩 Required Skills & Experience :
- 4+ years of experience in backend or distributed systems development.
- Strong proficiency in Java (11+), including concurrency and networking concepts.
- Experience with frameworks like Spring Boot, Vert.x, or similar.
- Hands-on experience with gRPC and REST API design.
- Strong knowledge of SQL databases (PostgreSQL preferred) and schema design.
- Experience working in Linux environments, with Git and Docker.
- Familiarity with event-driven systems / messaging queues (Kafka, RabbitMQ, MQTT, etc.).
- Ability to design scalable architectures and write automated tests.
⭐ Good to Have :
- Experience with real-time systems, telemetry, IoT, or satellite systems
- Exposure to distributed systems / microservices / high-availability architectures
- Working knowledge of Python for scripting or integrations
- Experience with monitoring tools like Prometheus and Grafana
- Interest in hardware-software integrated systems
Job Title : Senior QA Engineer (Crypto Exchange Platform)
Experience : 2+ Years
Location : Gurugram & Vadodara
Employment Type : Full-Time
About the Company :
We are a fast-growing crypto exchange platform building secure, scalable, and high-performance trading systems with real-time data and wallet infrastructure.
Role Overview :
We are looking for a Senior QA Engineer to ensure the quality, reliability, and security of our platform. You’ll work on web, mobile, and backend systems, focusing on APIs, trading engines, and real-time systems in a fast-paced agile environment.
Mandatory Skills :
2+ years in QA with strong manual & automation testing, experience in Selenium/Cypress/Playwright, API testing (Postman/REST Assured), CI/CD (Jenkins/GitHub Actions), SQL, and real-time/WebSocket testing.
Key Responsibilities :
- Create and execute test plans, cases, and strategies
- Perform functional, regression, integration & API testing
- Build and maintain automation frameworks
- Test trading systems, wallets, and real-time data (WebSockets)
- Track bugs using Jira and collaborate with teams
- Integrate testing into CI/CD pipelines
- Ensure performance, stability, and security
Required Skills :
- Strong experience in automation + functional testing
- Hands-on with Selenium/Cypress/Playwright
- Good knowledge of API testing & microservices
- Experience with CI/CD tools
- Strong SQL & database validation skills
- Understanding of Agile & SDLC
Good to Have :
- Experience in crypto/fintech/trading platforms
- Knowledge of blockchain, wallets, smart contracts
- Performance testing (JMeter, K6)
- Basic security testing knowledge
What We’re Looking For :
- Strong problem-solving skills
- Attention to detail
- Ability to work in a fast-paced environment
- Good communication & ownership mindset
Role Overview
We are looking for a skilled Systems Engineer with strong expertise in Ansible automation, Kubernetes, and system-level scripting. This role focuses on building scalable automation and managing GPU-aware infrastructure.
Key Responsibilities
- Develop and maintain automation workflows using Ansible
- Write and optimize Ansible playbooks for infrastructure and application deployment
- Automate system-level tasks using Bash scripting
- Debug and test automation workflows for reliability and scalability
- Manage Kubernetes clusters, including:
- Pod lifecycle management
- Networking and firewall configurations
- GPU resource mapping and scheduling
- Custom Resource Definitions (CRDs)
- Collaborate with engineering teams to integrate automation into CI/CD pipelines
- Ensure secure and scalable infrastructure for AI workloads
Requirements
- Strong hands-on experience with Ansible and automation workflows
- Solid Linux system administration experience
- Proficiency in Bash scripting
- Deep understanding of Kubernetes (pods, networking, CRDs, GPU scheduling)
- Experience in debugging and optimizing automation systems
- Strong problem-solving and ownership mindset
Nice to Have
- Experience with GPU workloads or HPC environments
- Exposure to monitoring tools like Prometheus and Grafana
Role: Sr. Azure Data Engineer
Experience: 8–10 Years
Work Timings: 1:30 PM – 10:30 PM IST
Location: Bellandur Bengaluru (Work from Office)
Company: Chevron
Employment Type: 6- 12 months Contract
Role Overview
We are seeking an experienced Senior Data Engineer to design and deliver scalable cloud data solutions on Azure. The ideal candidate will have strong expertise in Databricks, PySpark, and modern data architectures, with exposure to energy domain standards like OSDU.
Key Responsibilities
- Architect and design robust Azure-based data solutions using Databricks, ADLS, and PaaS services
- Define and implement scalable data Lakehouse architectures aligned with OSDU standards
- Build and manage end-to-end data pipelines for batch and real-time processing using PySpark
- Establish data governance frameworks including metadata, lineage, security, and access control
- Implement DevOps best practices (CI/CD, Azure Pipelines, GitHub, automated deployments)
- Collaborate with stakeholders to translate business needs into technical solutions
- Develop and maintain architecture documentation, solution patterns, and standards
- Provide technical leadership and mentorship to engineering teams
- Optimize solutions for performance, cost, reliability, and security
- Ensure alignment with enterprise architecture and compliance standards
- Drive adoption of modular and reusable cloud data components
Required Skills & Qualifications
Core Technical Skills
- Azure Databricks, Apache Spark (PySpark), Delta Lake, Unity Catalog
- Azure Data Lake Storage (ADLS), Azure Data Factory, Synapse Analytics
- Strong experience in Python-based data engineering
- Data pipeline development (batch + real-time)
Architecture & Advanced Skills
- Data Lakehouse architecture and distributed systems
- Microservices, APIs, and integration frameworks
- OSDU (Open Subsurface Data Universe) or similar energy data models
DevOps & Tools
- CI/CD tools: Azure Pipelines, GitHub Actions
- Infrastructure as Code: Terraform or similar
Other Skills
- Data governance, security, compliance, and cost optimization
- Strong analytical and problem-solving skills
- Excellent communication and stakeholder management
Job Title : AWS Data Engineer
Experience : 4+ Years
Location : Bengaluru (HSR – Hybrid, 3 Days WFO)
Notice Period : Immediate Joiner
💡 Role Overview :
We are looking for a skilled AWS Data Engineer to design, build, and scale modern data platforms. The role involves working with AWS-native services, Python, Spark, and DBT to deliver secure, scalable, and high-performance data solutions in an Agile environment.
🔥 Mandatory Skills :
Python, SQL, Spark, AWS (S3, Glue, EMR, Redshift, Athena, Lambda), DBT, ETL/ELT pipeline development, Airflow/Step Functions, Data Lake (Parquet/ORC/Iceberg), Terraform & CI/CD, Data Governance & Security
🚀 Key Responsibilities :
- Design, build, and optimize ETL/ELT pipelines using Python, DBT, and AWS services
- Develop and manage scalable data lakes on S3 using formats like Parquet, ORC, and Iceberg
- Build end-to-end data solutions using Glue, EMR, Lambda, Redshift, and Athena
- Implement data governance, security, and metadata management using Glue Data Catalog, Lake Formation, IAM, and KMS
- Orchestrate workflows using Airflow, Step Functions, or AWS-native tools
- Ensure reliability and automation via CloudWatch, CloudTrail, CodePipeline, and Terraform
- Collaborate with data analysts and data scientists to deliver actionable insights
- Work in an Agile environment to deliver high-quality data solutions
✅ Mandatory Skills :
- Strong Python (including AWS SDKs), SQL, Spark
- Hands-on experience with AWS data stack (S3, Glue, EMR, Redshift, Athena, Lambda)
- Experience with DBT and ETL/ELT pipeline development
- Workflow orchestration using Airflow / Step Functions
- Knowledge of data lake formats (Parquet, ORC, Iceberg)
- Exposure to DevOps practices (Terraform, CI/CD)
- Strong understanding of data governance and security best practices
- Minimum 4–7 years in Data Engineering (3+ years on AWS)
➕ Good to Have :
- Understanding of Data Mesh architecture
- Experience with platforms like Data.World
- Exposure to Hadoop / HDFS ecosystems
🤝 What We’re Looking For :
- Strong problem-solving and analytical skills
- Ability to work in a collaborative, cross-functional environment
- Good communication and stakeholder management skills
- Self-driven and adaptable to fast-paced environments
📝 Interview Process :
- Online Assessment
- Technical Interview
- Fitment Round
- Client Round
Required Skills
- 8+ years of DevOps / Cloud Engineering experience
- Strong hands-on experience with AWS services (EC2, S3, RDS, IAM, VPC, etc.)
- Expertise in Kubernetes (deployment, scaling, cluster management)
- Strong experience in PostgreSQL and AWS RDS administration
- Proficiency in Terraform for infrastructure automation
- Experience building and maintaining CI/CD pipelines (Jenkins, GitLab CI, etc.)
- Strong knowledge of Java (mandatory) and application deployment lifecycle
- Experience with Docker and containerization
- Solid understanding of networking, security, and system architecture
- Strong troubleshooting and problem-solving skills
Responsibilities
- Technical Leadership & Architecture: Serve as a technical authority, define cloud-native standards, and own end-to-end system design for complex, distributed, and high-scale solutions.
- Engineering Excellence: Lead modernization initiatives, transition legacy systems to microservices, and ensure performance, scalability, and security.
- Platform & Cloud Enablement: Champion GCP-native services (GKE, Cloud Run, Pub/Sub, BigQuery, Cloud SQL, Spanner), influence CI/CD, and drive FinOps strategies.
- Mentorship & Organizational Impact: Mentor Lead and Senior Engineers, foster engineering rigor, and collaborate with Product Management to translate business strategy into technical solutions.
Required Skills, Experience & Background
- 12+ years of professional experience building enterprise-grade software systems.
- Proven experience operating at a Lead or Principal Engineer level.
- Strong hands-on experience with Java or another major programming language (e.g., Go, Python, C#).
- Deep expertise in microservices architecture, RESTful APIs, and distributed systems.
- Strong experience with SQL and NoSQL databases and data modeling.
- Expertise in cloud-native design principles (stateless services, scalability, resiliency).
- Strong experience building and operating systems on Google Cloud Platform (GCP).
- Proficiency with containers and orchestration (Docker, Kubernetes, GKE).
- Experience with CI/CD pipelines, Git-based workflows, and observability.
- Experience with Big Data and data engineering technologies (e.g., BigQuery, Spark).
Qualifications
- Bachelor’s degree in computer science or related field (required).
- Master’s degree in computer science or related field (preferred).
EasySLR is pioneering the future of systematic literature reviews through AI and innovative technologies. Our platform, recognized by industry leaders and academic communities alike, redefines the way researchers conduct reviews, making the process faster, smarter, and more intuitive. We've been at the forefront of AI-driven research, presenting at major conferences and setting new standards in evidence synthesis. If you are a visionary leader with a passion for technology and a drive to make a significant impact, we want you to join our mission to transform the research landscape.
Responsibilities :
- Lead and mentor a team of talented engineers, fostering a culture of innovation, collaboration, and continuous learning.
- Architect and oversee the development of a scalable, high-performance platform that integrates cutting-edge AI technologies and industry best practices.
- Drive the engineering strategy, ensuring alignment with our product vision and business goals.
- Collaborate closely with cross-functional teams, including product, design, and AI experts, to deliver a world-class product experience.
- Ensure the robustness, security, and scalability of our infrastructure, leveraging your deep expertise in cloud computing and full-stack development.
- Stay ahead of emerging technologies, incorporating the latest advancements into our platform and maintaining our competitive edge.
- Cultivate a high-performing engineering team through effective hiring, coaching, and professional development opportunities.
Requirements :
- 4+ years of experience in software engineering, with a proven track record of leading high-performing engineering teams.
- Expertise in full-stack development, with hands-on experience in Python, Node.js, and frameworks like Next.js.
- Extensive experience with cloud platforms, particularly AWS, and familiarity with tools like AWS Lambda, AWS CDK, and containerization technologies.
- Strong background in designing and scaling complex, distributed systems with a focus on performance and security.
- Experience in AI/ML-driven product development is a significant plus.
- Exceptional problem-solving skills, with a strategic mindset and the ability to make data-driven decisions.
- Excellent communication skills, with the ability to articulate complex technical concepts to non-technical stakeholders.
What We Offer :
- The opportunity to lead a cutting-edge platform at the intersection of AI and systematic literature reviews.
- Competitive compensation and a clear path to executive leadership.
- A vibrant, inclusive work culture that values diversity, innovation, and work-life balance.
- The chance to make a meaningful impact in a fast-growing, AI-first SaaS company shaping the future of research.
Ready to lead the engineering efforts that will drive the next generation of AI-driven systematic reviews? Join us at EasySLR and be part of a team that's revolutionizing the research process. Apply now and embark on an exciting journey at the forefront of technology and innovation
DevOps Engineer
Location: Bangalore office
About Peliqan
Peliqan is an all-in-one data platform combining ELT/ETL pipelines, a built-in data warehouse, SQL and low-code Python transformations, reverse ETL, and AI-powered data activation. We connect 250+ data sources and serve enterprise teams, consultants, and SaaS companies. SOC 2 Type II certified and GDPR compliant.
The Role
Own and evolve the infrastructure powering Peliqan's multi-tenant data platform. You'll manage Kubernetes clusters, cloud resources, CI/CD pipelines, and monitoring — keeping everything reliable, secure, and scalable. You'll be the go-to person for infrastructure support across the engineering team.
Responsibilities
Manage and optimise Kubernetes clusters running production workloads — data pipelines, APIs, and customer-facing services.
Maintain Docker-based local development environments for the engineering team.
Administer cloud infrastructure on AWS and Google Cloud (compute, storage, networking, managed databases).
Build and maintain CI/CD pipelines for automated testing, building, and deploying across staging and production.
Set up and manage monitoring, alerting, and logging for platform health and incident response.
Manage release processes — deployments, rollbacks, and release strategies.
- Maintain infrastructure-as-code using Helm charts.
- Support security hardening and compliance efforts (SOC 2, GDPR).
Requirements
3+ years in a DevOps, SRE, or Infrastructure Engineering role.
Strong hands-on experience with Kubernetes and Helm charts.
Deep familiarity with Docker for containerisation and local dev workflows.
Production experience with AWS and/or Google Cloud.
- Proficiency in Python and Bash scripting for automation and tooling.
- Solid grasp of DevOps principles: infrastructure-as-code, GitOps, observability, continuous delivery.
- Experience with CI/CD platforms (GitHub Actions, GitLab CI, or similar).
Nice to Have
- Experience supporting multi-tenant SaaS platforms or data infrastructure at scale.
- Knowledge of PostgreSQL, MySQL, or cloud-managed database administration.
- Exposure to security compliance frameworks (SOC 2, ISO 27001, GDPR).
QA Tester
Location: Bangalore office
About Peliqan
Peliqan is an all-in-one data platform combining ELT/ETL pipelines, a built-in data warehouse, SQL and low-code Python transformations, reverse ETL, and AI-powered data activation. We connect 250+ data sources and serve enterprise teams, consultants, and SaaS companies. SOC 2 Type II certified and GDPR compliant.
The Role
Own quality end-to-end across Peliqan's platform — from the frontend UI and data apps to backend pipelines, connectors, and APIs. You'll design test strategies, build automated test suites, and work closely with developers to ship reliable software. Comfort using AI tools to accelerate test creation is essential.
Responsibilities
Design and maintain test plans covering manual and automated testing for features, regressions, and releases.
Write unit tests for the frontend (Jest) and backend (pytest).
- Build and extend end-to-end test suites using Playwright across critical user flows — connector setup, data transformations, data app publishing, API creation.
- Use AI tools (Copilot, Claude, etc.) to rapidly generate and refine test cases and test data.
- Triage, document, and track defects with clear reproduction steps.
- Integrate automated tests into CI/CD pipelines.
- Conduct exploratory testing across data pipelines, the query engine, and user-facing interfaces.
Requirements
2+ years in a QA or SDET role, ideally in SaaS or data products.
Hands-on experience with Jest, pytest, and Playwright.
- Comfortable with Python for scripting and test automation.
- Demonstrated use of AI tools to craft and scale test suites.
- Familiarity with CI/CD pipelines and REST API testing.
- Strong analytical mindset and clear written communication.
Nice to Have
- Experience testing ETL/ELT pipelines or database-heavy applications.
- Familiarity with SQL and data validation testing.
- Exposure to Docker and containerised environments.

Job Title: Consultant – Enterprise Application Development
Location: Bengaluru (Hybrid / On-site)
Engagement: Full-Time
Experience: 10 – 15 years preferred
About Us: Introducing VTT, a comprehensive mobility service provider catering to diverse multinational sectors like IT/ITES, KPO/BPO, Financial, Pharma, and more across Indian cities. Our “Managed Mobility Program” includes Fleet Management, Technology, Resource Management, Car Rentals, Logistics, and Special Services (Ambulance and PWD vehicles). Trusted by Fortune companies such as Cisco, Morgan Stanley, Wells Fargo, Google, PWC, and others, we pride ourselves on leveraging expertise and cutting-edge technology for safe, efficient, and uninterrupted service delivery. With a commitment to excellence, we ensure best-in-class standards for all our clients. Trip to school is now timely, comfy and secure! Our well maintained f leet is here to enrich your child’s commute, keeping students punctual and safe thanks to GPS tracking paired with well-trained drivers. Our routes are carefully planned, our drivers attentive, and everything hassle-free.
Role Overview
We are looking for a seasoned Consultant with comprehensive expertise in enterprise-level application development across backend, frontend, mobile, DevOps, and cloud. The role demands a strong architectural mindset combined with hands-on execution. The Consultant will also play a critical role in understanding the current system architecture end-to-end, driving technical improvements, building the tech team foundation, and establishing structured technical documentation.
Key Responsibilities
• Understand the complete architecture of the existing systems, including web, mobile, backend services, and cloud environment.
• Provide hands-on leadership across backend, frontend, mobile, DevOps, and cloud infrastructure.
• Architect and optimize enterprise-grade applications for scalability, security, performance, and reliability.
• Conduct technical due diligence on current systems and propose improvements or refactoring plans.
• Build the foundation for the internal engineering team including hiring support, role definitions, and best-practice processes.
• Drive engineering workflows including coding standards, branching strategy, CI/CD, monitoring, and release management.
• Create comprehensive technical documentation covering system architecture, API specs, deployment playbooks, and SOPs.
• Review code and provide mentorship to engineering resources.
• Coordinate with product and business teams to translate requirements into technical design and actionable development roadmap.
• Troubleshoot and resolve deep-stack issues during development or production.
Technical Expertise Required
Backend
• Java / Spring Boot
• Node.js
•Microservices architecture
• REST / GraphQL
Frontend
• React js
• Responsive UI, component-based architecture, state management
Mobile
• Flutter
• React Native
Cloud & DevOps
• AWS (ECS / EKS / EC2 / RDS / Lambda / S3 / IAM / CloudWatch etc.)
• CI/CD pipelines (GitHub Actions / Jenkins / GitLab CI or equivalent)
• Docker / Kubernetes
• Infrastructure-as-code (Terraform / CloudFormation)
Database
• MongoDB
• Knowledge of PostgreSQL / MySQL is an added advantage
Professional Attributes
• Strong architectural thinking with the ability to simplify complex systems.
• Excellent communication and stakeholder management skills.
• Ability to work independently without constant supervision.
• Capability to mentor, lead, and build an engineering team from scratch.
• Process-driven mindset with a focus on best practices and documentation.
Deliverables
• Architectural understanding and documentation of current systems.
• Recommendations and implementation plan for system upgrades or restructuring.
• Establishment of core engineering processes and standards.
• Hiring support and technical evaluation of developers.
Role Summary
The Service Delivery & Change Manager is responsible for ensuring smooth day-to-day support operations, structured execution of change requests, and controlled production releases across client accounts and internal systems. This is a hybrid role requiring a mix of delivery management, technical coordination, stakeholder communication, and operational discipline.
Key Responsibilities
- Oversee day-to-day production support activities across projects or client accounts
- Manage incidents, service requests, and operational issues to ensure timely resolution and SLA compliance
- Coordinate with internal technical teams for troubleshooting, escalation, and closure of support cases
- Maintain support records, trackers, and operational documentation
- Support business users with system-related requests, user access coordination, and issue follow-up where needed
- Manage the complete lifecycle of change requests from requirement gathering to closure
- Coordinate with business stakeholders, clients, and technical teams to define scope, effort, timelines, and priorities
- Ensure proper approvals, documentation, planning, testing, deployment readiness, and closure of all changes
- Coordinate production deployments involving application and database changes
- Support release planning, deployment communication, release notes, and post-release follow-up
- Act as the primary coordination point between clients, business teams, and delivery teams
- Lead or support review meetings, status calls, follow-ups, and issue discussions
- Provide regular updates on support status, change progress, risks, and delivery timelines
- Ensure adherence to internal processes, governance standards, and audit requirements
- Support compliance-related activities, reporting needs, and documentation for management or client review
- Prepare weekly and monthly reports covering incidents, support activities, change requests, deployment status, and team utilization
- Track open items, pending approvals, delivery risks, and recurring support trends
- Support management with structured data for invoicing inputs, audit preparation, and operational analysis
- Contribute to process improvements that strengthen service quality, delivery control, and communication flow
- Manage resource planning, team allocation, leave schedules, and backup coverage to ensure smooth delivery continuity.
- Support onboarding, knowledge transfer, and internal coordination for smooth service continuity
- Identify, support, and help drive AI-led initiatives that improve operational efficiency, service quality, reporting, or delivery processes
Required Skills / Knowledge
- Strong experience in application support, service delivery, change management, or technical project coordination
- Good understanding of incident management, service request handling, and change control processes
- Experience in production release coordination and deployment planning
- Strong stakeholder management and client communication skills
- Ability to work across business, technical, and operational teams
- Strong reporting, documentation, and follow-up discipline
- Familiarity with ticketing tools, release processes, and enterprise support environments
- Good analytical and problem-solving skills
- Ability to manage multiple priorities in a structured and calm manner
- Working knowledge of Java-based enterprise applications, including Spring Framework and Hibernate
- Familiarity with cloud platforms such as AWS
- Basic working knowledge of Linux environments
- Understanding of CI/CD tools and version control platforms such as Jenkins, GitHub, and Bitbucket
- Ability to coordinate effectively with development and DevOps teams on application changes, deployments, and environment-related issues
- Strong interest in emerging technologies and willingness to contribute to AI initiatives within the organization
Preferred Qualifications
- Bachelor’s degree in Engineering, IT, Computer Science, or related field
- 5+ years of relevant experience in service delivery, application support, or technical project management
- Experience in client-facing delivery environments
- Exposure to audit, compliance, or governance-heavy projects is an advantage
- Experience in the Energy sector and/or Ecommerce will be an added advantage
Job Title : Golang Backend Developer
Experience : 3+ Years
Location : Bangalore (Work From Office)
Notice Period : Immediate to 15 Days (Strict)
🚀 About the Role :
We are looking for a Backend Developer with strong Golang expertise to build scalable, high-performance systems. You will play a key role in designing microservices, handling concurrent workloads, and developing robust backend architectures for production-scale applications.
🔥 Mandatory Skills :
Strong hands-on experience in Golang, Microservices Architecture, REST APIs, Concurrency (goroutines & channels), PostgreSQL/MySQL, Redis, Messaging Systems (Kafka/RabbitMQ/SQS), AWS/GCP, Docker & Kubernetes, and CI/CD pipelines.
🛠️ Key Responsibilities :
- Design, develop, and maintain scalable backend services using Golang.
- Build high-performance REST APIs and microservices.
- Develop concurrent and distributed systems using goroutines and channels.
- Implement event-driven and asynchronous architectures.
- Optimize system performance, latency, and database efficiency.
- Integrate messaging systems and caching layers for scalability.
- Collaborate with cross-functional teams for end-to-end delivery.
- Ensure high code quality, testing, and system reliability.
- Monitor, debug, and enhance production systems.
Required Skills & Qualifications :
- Strong hands-on experience in Golang (must-have).
- Solid understanding of Concurrency in Go (goroutines, channels, worker pools).
- Experience with Microservices Architecture.
- Strong knowledge of RESTful API development.
- Proficiency in Databases : PostgreSQL / MySQL / MongoDB.
- Hands-on experience with Redis (caching).
- Experience with Messaging Systems: Kafka / RabbitMQ / SQS.
- Hands-on experience with AWS or GCP.
- Experience with Docker & Kubernetes.
- Familiarity with CI/CD pipelines (GitHub Actions, Jenkins, etc.).
- Strong understanding of Data Structures, Algorithms, and System Design.
Good to Have :
- Experience with gRPC-based microservices.
- Familiarity with monitoring tools like Prometheus, Grafana.
- Exposure to high-scale distributed systems.
- Experience with event-driven architectures.
- Knowledge of security practices (JWT, OAuth2, RBAC).
What We’re Looking For :
- Strong problem-solving and debugging skills.
- Ownership mindset with end-to-end feature delivery.
- Ability to write clean, efficient, and maintainable code.
- Comfortable working in a fast-paced, high-growth environment.
You can also register yourself on the below platform to proceed further :
What You’ll Be Doing
Design and implement a Cypress + TypeScript automation framework
Define and drive the automation strategy across teams
Build and scale:
Smoke tests
Regression suites
API and end-to-end automation
Integrate automation into CI/CD pipelines
Introduce AI-powered testing approaches (including TestRail AI)
Mentor QA engineers and collaborate with external suppliers
🎯 What We’re Looking For
6+ years in QA Automation
Strong experience with Cypress and TypeScript
Proven track record building automation frameworks from scratch
Experience defining test strategies in complex environments
Solid API + UI automation experience
Strong stakeholder and communication skills
💡 Nice to Have
Experience with TestRail / AI-driven testing
Exposure to enterprise-scale transformation programmes
CI/CD tools (Azure DevOps, GitHub Actions)
Python Embedded Engineer
Location: Chennai
Experience : 3+ years
Budget : 1.2 LPM
We are looking for
Python Engineer who has experience in embedded system is must. Please look for the candidates who are trustworthy and good in communication.
We are looking for a skilled Python Embedded Engineer with 3+ years of experience and exposure to embedded systems.
The ideal candidate will work closely with cross-functional teams to design, develop, and optimize software solutions that interact with hardware and embedded platforms.
Key Responsibilities
Design, develop, and maintain robust Python-based applications and tools.
Work on embedded system software development and integration.
Collaborate with hardware, firmware, and system teams for end-to-end solution development.
Optimize software performance for embedded environments.
Debug, troubleshoot, and resolve system-level issues.
Participate in code reviews and ensure adherence to coding standards.
Contribute to testing, validation, and product release activities.
Required Skills & Qualifications
3+ years of professional experience in Python development.
Strong understanding of software development fundamentals.
Experience or exposure to embedded systems is preferred.
Knowledge of hardware-software interaction and communication protocols.
Familiarity with Git or other version control systems.
Good analytical and problem-solving skills.
Strong communication and teamwork abilities.
Good to Have
Experience with C/C++.
Knowledge of RTOS or Linux-based embedded systems.
Experience with microcontrollers, sensors, or device drivers.
Exposure to CI/CD and automated testing frameworks.
About the Role
As an SDET II, you'll own significant parts of our test infrastructure and drive quality strategy across the engineering team. You'll design testing approaches for complex features, mentor junior engineers, and make architectural decisions that impact how we approach automation at scale.
What You'll Do
- Architect and implement test frameworks and infrastructure
- Design testing strategies for new features and platform capabilities
- Mentor SDET I engineers and conduct technical code reviews
- Refactor and optimize existing test suites for maintainability and performance
- Make architectural decisions about test design patterns and abstractions
- Build and manage AWS-based test environments and infrastructure
- Integrate testing earlier in the development lifecycle through cross-team collaboration
- Optimize CI/CD pipeline performance and test execution times
- Develop custom tooling and reporting to surface quality insights
Technical Requirements
Core Skills:
- Advanced TypeScript expertise: generics, decorators, advanced typing patterns, type inference
- Deep understanding of asynchronous programming, concurrency, and race condition prevention
- Strong software design principles with domain-driven design (DDD) approach
- Extensive experience with Playwright including deep knowledge of fixtures architecture
- Expert-level Git, GitHub, and distributed version control workflows
- Layered architecture design: understanding PCOM (Page Component Object Model) and POM patterns
- Object-oriented design in test frameworks—building scalable abstractions over linear scripts
- API testing and orchestration (REST/GraphQL integration with UI workflows)
Infrastructure & DevOps:
- AWS: EC2 configuration, CloudWatch log analysis, debugging cloud environments
- Terraform for infrastructure as code (plus)
- Docker: containerization, docker-compose, image management
- CI/CD debugging: analyzing pipeline failures, optimizing execution
- Advanced reporting: Allure configuration, Playwright HTML reports, custom reporting solutions
Additional Experience:
- Test infrastructure development and framework architecture
- Design patterns implementation (Factory, Builder, Facade, Composite)
- Performance optimization at scale
- npm ecosystem and package management
About the Role
As an SDET II, you'll own significant parts of our test infrastructure and drive quality strategy across the engineering team. You'll design testing approaches for complex features, mentor junior engineers, and make architectural decisions that impact how we approach automation at scale.
What You'll Do
- Architect and implement test frameworks and infrastructure
- Design testing strategies for new features and platform capabilities
- Mentor SDET I engineers and conduct technical code reviews
- Refactor and optimize existing test suites for maintainability and performance
- Make architectural decisions about test design patterns and abstractions
- Build and manage AWS-based test environments and infrastructure
- Integrate testing earlier in the development lifecycle through cross-team collaboration
- Optimize CI/CD pipeline performance and test execution times
- Develop custom tooling and reporting to surface quality insights
Technical Requirements
Core Skills:
- Advanced TypeScript expertise: generics, decorators, advanced typing patterns, type inference
- Deep understanding of asynchronous programming, concurrency, and race condition prevention
- Strong software design principles with domain-driven design (DDD) approach
- Extensive experience with Playwright including deep knowledge of fixtures architecture
- Expert-level Git, GitHub, and distributed version control workflows
- Layered architecture design: understanding PCOM (Page Component Object Model) and POM patterns
- Object-oriented design in test frameworks—building scalable abstractions over linear scripts
- API testing and orchestration (REST/GraphQL integration with UI workflows)
Infrastructure & DevOps:
- AWS: EC2 configuration, CloudWatch log analysis, debugging cloud environments
- Terraform for infrastructure as code (plus)
- Docker: containerization, docker-compose, image management
- CI/CD debugging: analyzing pipeline failures, optimizing execution
- Advanced reporting: Allure configuration, Playwright HTML reports, custom reporting solutions
Additional Experience:
- Test infrastructure development and framework architecture
- Design patterns implementation (Factory, Builder, Facade, Composite)
- Performance optimization at scale
- npm ecosystem and package management is a good to have
About Us
We believe the future of software development is AI-native — where engineers operate at a higher level of abstraction and quality remains non-negotiable.
Incubyte is a software craft consultancy where the “how” of building software matters as much as the “what”.
We partner with companies of all sizes, from helping enterprises build, scale, and modernize to early-stage founders bring their ideas to life.
Our engineers operate in an AI-native development model, using AI as a collaborator across the SDLC to accelerate development while upholding the discipline of software craftsmanship. Guided by Software Craftsmanship and Extreme Programming practices, we build reliable, maintainable, and scalable systems with speed, without compromising quality. If this way of building software resonates with you, we’d like to talk.
Our Guiding Principles
These principles define how we work at Incubyte. They are non-negotiable.
Relentless Pursuit of Quality with Pragmatism
We build high-quality systems without losing sight of delivery.
Extreme Ownership
We take responsibility end-to-end for decisions, execution, and outcomes.
Proactive Collaboration
We collaborate closely, challenge each other, and solve problems together.
Active Pursuit of Mastery
We continuously improve our craft and raise our bar.
Invite, Give, and Act on Feedback
We seek, give, and act on feedback to get better every day.
Ensuring Client Success
We act as trusted partners and focus on real outcomes, not just output.
Job Description
This is a remote position.
Experience Level
This role is ideal for engineers with 5+ years of full-stack software development experience, with a strong developer mindset and hands-on experience building and maintaining production systems.
An experienced individual contributor who thrives in ownership-driven environments, can operate effectively with minimal oversight, and is comfortable working in the absence of a traditional Product Manager while demonstrating a strong product mindset.
What You’ll Do as a Software Craftsperson
- Design and build high-quality, maintainable systems using disciplined engineering practices such as TDD, Test Pyramid, and a strong focus on overall software quality
- Operate in an AI-powered development environment, leveraging AI as a collaborator to explore solutions, accelerate development, and improve outcomes while maintaining strong engineering judgment
- Take end-to-end ownership of tasks, from understanding requirements to execution and delivery, with minimal supervision
- Work extensively on modern JavaScript/TypeScript-based systems as a core part of the development stack
- Develop and maintain backend systems using Ruby on Rails, ensuring scalability, performance, and reliability
- Build and manage DevOps workflows, including creating and maintaining CI/CD pipelines using tools such as CircleCI and GitHub Actions
- Set up and monitor systems using tools like Datadog dashboards to ensure system reliability and observability
- Make thoughtful technical decisions that balance simplicity, scalability, and maintainability
- Work in a highly autonomous environment, defining problems, interacting with customers when needed, and driving solutions independently
- Contribute to improving engineering practices, developer experience, and system quality across the team
- Contribute to mobile development efforts using React Native or Flutter (where applicable)
Requirements
What You’ll Bring
- 5+ years of hands-on full-stack software development experience with a strong developer mindset
- Expert-level proficiency in JavaScript and/or TypeScript
- Strong experience or familiarity with Ruby on Rails (Intermediate level or above preferred)
- Solid understanding of Software Craftsmanship principles, including TDD, Test Pyramid, and quality-first development
- Experience or familiarity with DevOps practices (CI/CD pipelines, GitHub Actions, CircleCI) at an intermediate level
- Familiarity with monitoring and observability tools such as Datadog
- Exposure to or interest in AI Engineering (beginner level acceptable) and the ability to work effectively in AI-assisted environments
- Experience with mobile development- Flutter
- Strong critical thinking and problem-solving skills, with the ability to independently define and execute tasks
- Comfort working in a high-ownership, low-structure environment without a dedicated Product Manager
What We Value
- Extreme ownership – you take responsibility for outcomes and drive work end-to-end
- Ability to independently execute and define problems, not just solve assigned tasks
- A strong product mindset – willingness to talk to customers, understand problems deeply, and define direction when needed
- Deep care about developer experience and engineering quality
- High standards, autonomy, and a bias toward action
Life at Incubyte
We are a remote-first company with structured flexibility. Teams commit to shared rhythms during core hours, ensuring smooth collaboration while maintaining autonomy. Twice a year, we come together in person for a co-working sprint and once a year for a retreat - with all travel expenses covered.
Our environment is built for crafters: pairing, refactoring, experimenting with AI, and pushing the boundaries of software excellence. We are all lifelong learners, and our work is our passion.
Perks
- Dedicated learning & development budget.
- Sponsorship for conference talks.
- Comprehensive medical & term insurance.
- Employee-friendly leave policies.
- Home Office fund
- Medical Insurance
Job Summary:
We are looking for a skilled ReactJS Developer with 4–7 years of experience to build scalable, high-performance web applications. The ideal candidate should have strong expertise in modern JavaScript frameworks, state management, and frontend architecture, along with exposure to backend integration and deployment practices.
Key Responsibilities:
- Develop and maintain responsive web applications using React (v17+)
- Build reusable components and front-end libraries for future use
- Implement state management using Redux and React Query
- Integrate APIs (REST/GraphQL) efficiently with frontend applications
- Optimize applications for maximum performance and scalability
- Collaborate with cross-functional teams (Design, Backend, DevOps)
- Ensure code quality through best practices, testing, and reviews
- Manage routing and navigation using React Router
- Work with CI/CD pipelines and deployment processes
- Participate in architecture discussions and technical decision-making
Required Skills & Qualifications:
- Strong experience in ReactJS (v17+)
- Proficient in JavaScript (ES6+) and TypeScript
- Hands-on experience with Redux and React Query
- Solid understanding of HTML5, CSS3, and CSS preprocessors / CSS Modules
- Experience with React Router for navigation
- Strong knowledge of REST APIs / GraphQL integration
- Familiarity with Webpack and frontend build tools
- Experience with Docker and containerized environments
- Knowledge of CI/CD pipelines and deployment processes
- Proficient in Git version control
About NonStop io Technologies:
NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment.
We're seeking an experienced Senior Backend Engineer to join our team. As a senior backend engineer, you will be responsible for designing, developing, and deploying scalable backends for the products we build at NonStop. This includes APIs, databases, and server-side logic.
Responsibilities:
● Design, develop, and deploy backend systems, including APIs, databases, and server-side logic
● Write clean, efficient, and well-documented code that adheres to industry standards and best practices
● Code Quality: Ensure code quality through code reviews, adherence to best practices, and continuous improvement
● Mentorship: Guide and mentor team members, fostering growth and innovation
● Collaboration: Work closely with stakeholders to align technical goals with business objectives
● Problem-Solving: Analyze and resolve technical challenges promptly ● Innovation: Stay updated with the latest technology trends and integrate them into solutions
Requirements:
● At least 7+ years of experience building scalable and reliable backend systems
● Strong expertise in NodeJS/NestJS, Express, PostgreSQL
● Experience with microservices architecture and distributed systems
● Proficiency in database design (SQL and NoSQL)
● Knowledge of cloud platforms (AWS, Azure, or GCP) and CI/CD pipelines
● Deep understanding of design patterns, data structures, and algorithms
● Hands-on experience with containerization technologies like Docker and orchestration tools like Kubernetes
● Exceptional communication and leadership skills
● Strong understanding of object-oriented programming principles and design patterns
● Familiarity with automated testing frameworks and methodologies
● Excellent problem-solving skills and attention to detail
● Strong communication skills and ability to effectively lead and maintain a collaborative team environment
About NonStop io Technologies:
NonStop io Technologies is a value-driven company with a strong focus on process-oriented software engineering. We specialize in Product Development and have a decade's worth of experience in building web and mobile applications across various domains. NonStop io Technologies follows core principles that guide its operations and believes in staying invested in a product's vision for the long term. We are a small but proud group of individuals who believe in the 'givers gain' philosophy and strive to provide value in order to seek value. We are committed to and specialize in building cutting-edge technology products and serving as trusted technology partners for startups and enterprises. We pride ourselves on fostering innovation, learning, and community engagement. Join us to work on impactful projects in a collaborative and vibrant environment. We're seeking an experienced Backend Software Engineer to join our team. As a backend engineer, you will be responsible for designing, developing, and deploying scalable backends for the products we build at NonStop. This includes APIs, databases, and server-side logic.
Responsibilities:
● Design, develop, and deploy backend systems, including APIs, databases, and server-side logic
● Write clean, efficient, and well-documented code that adheres to industry standards and best practices
● Participate in code reviews and contribute to the improvement of the codebase
● Debug and resolve issues in the existing codebase
● Develop and execute unit tests to ensure high code quality
● Work with DevOps engineers to ensure seamless deployment of software changes
● Monitor application performance, identify bottlenecks, and optimize systems for better scalability and efficiency
● Stay up-to-date with industry trends and emerging technologies; advocate for best practices and new ideas within the team
● Collaborate with cross-functional teams to identify and prioritize project requirements
Requirements:
● At least 3+ years of experience building scalable and reliable backend systems
● Strong expertise in NodeJS/NestJS, Express, PostgreSQL
● Experience with microservices architecture and distributed systems
● Proficiency in database design (SQL and NoSQL)
● Knowledge of cloud platforms (AWS, Azure, or GCP) and CI/CD pipelines
● Deep understanding of design patterns, data structures, and algorithms
● Hands-on experience with containerization technologies like Docker and orchestration tools like Kubernetes
● Exceptional communication and leadership skills
● Strong understanding of object-oriented programming principles and design patterns
● Familiarity with automated testing frameworks and methodologies
● Excellent problem-solving skills and attention to detail
● Strong communication skills and ability to effectively lead and maintain a collaborative team environment























