
Overview
adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job Description
We are looking for an experienced Senior Software Developer for a local IT Services Management Team. In this role, you will be responsible for the application management of a B2C application to meet the agreed Service Level Agreements (SLAs) and fulfil customer expectations.
You will act as a on-call-duty team in the time between 6 pm to 8 am, 365 days a year.
We are seeking a hands-on Software Developer with strong experience in DevOps and Backend
Responsibilities:
Problem Management & Incident Management activities: Identifying and resolving technical issues and errors that arise during application usage.
Release and Update Coordination: Planning and executing software updates, new versions, or system upgrades to keep applications up to date.
Change Management: Responsible for implementing and coordinating changes to the application, considering the impact on ongoing operations.
Requirements:
Education and Experience: A Bachelor’s or Master’s degree in a relevant field, with a minimum of 5 years of professional experience or equivalent work experience.
Skills & Expertise:
Proficient in ITIL service management frameworks.
Strong analytical and problem-solving abilities.
Experienced in project management methodologies (Agile, Kanban).
Communication: Excellent communication, presentation, and interpersonal skills, with the ability to engage and collaborate with stakeholders.
Language: English on a C2 Level.
Skills & Requirements
Spring Boot 3.x, Spring Security, Spring Data JPA, Testing, Unit Test, Spring Boot Test, Keycloak, Java 21 or lower version.

Similar jobs
Location: Bangalore
Experience: 2–5 years
Type: Full-time | On-site
Open Roles: 1
Start: Immediate
Why this role exists
Most engineering teams choose between speed and stability.
We need both.
Today:
- Deployments carry risk
- Cloud costs are higher than they should be
- Compliance is reactive, not built-in
This role exists to build a platform where:
- We can deploy fast without breaking production
- We can scale without runaway cost
- We can pass enterprise InfoSec reviews without firefighting
What you’ll do
You will not just manage infrastructure.
You will build the platform that engineering runs on.
1. Drive cloud cost efficiency
- Reduce Azure compute spend by 40%
- Implement:
- Reserved Instances / savings plans
- Right-sizing of workloads
- Scheduling for non-critical workloads
- Continuously monitor and optimize cost vs performance
2. Build zero-downtime deployment systems
- Ship a deployment pipeline that supports:
- 5+ production deployments per week
- Zero customer-visible downtime
- Implement:
- Blue-green / canary deployments
- Automated health checks
- Safe rollout strategies
3. Enable fast and safe releases
- Reduce time-to-launch significantly
- Ensure:
- High reliability in every release
- Ability to rollback instantly if something breaks
- Create systems where:
- Scaling up is seamless when things go right
- Failures are contained when they don’t
4. Build disaster recovery and compliance readiness
- Create DR/BCP systems that pass enterprise audits from:
- HDFC Life, SBI Life
- Ensure:
- Backup and recovery processes are defined and tested
- Failover strategies are documented and executable
- Build compliance as part of the system, not an afterthought
5. Embed security into the pipeline
- Integrate:
- SAST (Static Application Security Testing)
- DAST (Dynamic Application Security Testing)
- SCA (Software Composition Analysis)
- Secret scanning
- Container scanning
- IaC scanning
- Ensure vulnerabilities are caught before deployment
6. Enforce policy-as-code
- Implement:
- OPA / Gatekeeper
- Azure Policy
- Prevent non-compliant infrastructure from being deployed
- Ensure consistency across environments
7. Build a scalable platform layer
- Create systems that:
- Support increasing deployment frequency
- Maintain reliability under scale
- Work closely with backend and SRE teams to:
- Improve system stability
- Reduce operational overhead
What success looks like
- Cloud costs reduce by ≥ 40%
- Deployments are:
- Frequent
- Safe
- Invisible to customers
- Rollbacks are instant and reliable
- DR/BCP passes enterprise audits in the first attempt
- Security is embedded in the pipeline, not patched later
- Engineering teams ship faster with confidence
Who you are
- You have 2-5 years of experience in DevOps / Platform Engineering
- You have worked with:
- Cloud platforms (Azure preferred)
- CI/CD systems
- Infrastructure as Code
- You think in:
- Systems
- Trade-offs (speed vs reliability vs cost)
- You are comfortable owning:
- Production infrastructure
- Deployment systems
What will make you stand out
- Experience with:
- High-frequency deployment systems
- Cost optimization at scale
- Security-first pipelines
- Strong understanding of:
- Kubernetes / container orchestration
- Monitoring and observability
- Distributed system reliability
- Experience passing enterprise security/compliance audits
Why join
- You will define how engineering ships and scales
- Your work directly impacts:
- Reliability
- Cost
- Deployment velocity
- You will build a platform that moves from:
- Fragile → predictable and scalable
What this role is not
- Not manual infra management
- Not reactive firefighting
- Not limited to CI/CD maintenance
What this role is
- A builder of deployment systems
- A driver of cost efficiency
- A guardian of reliability and compliance
One question to self-evaluate
Can you build a platform where we deploy faster, spend less, and never break production?
True to its name, is on a mission to unlock $100+ billion of trapped working capital in the economy by creating India’s largest marketplace for invoice discounting to solve the day-to-day. problems faced by businesses. Founded by ex-BCG and ISB / IIM alumni, and backed by SAIF Partners, CashFlo helps democratize access to credit in a fair and transparent manner. Awarded Supply Chain Finance solution of the year in 2019, CashFlo creates a win-win ecosystem for Buyers, suppliers
and financiers through its unique platform model. CashFlo shares its parentage with HCS Ltd., a 25 year old, highly reputed financial services company has raised over Rs. 15,000 Crores in the market till date,
for over 200 corporate clients.
Our leadership team consists of ex-BCG, ISB / IIM alumni with a team of industry veterans from financial services serving as the advisory board. We bring to the table deep insights in the SME lending
space, based on 100+ years of combined experience in Financial Services. We are a team of passionate problem solvers, and are looking for like-minded people to join our team.
The challenge
Solve a complex $300+ billion problem at the cutting edge of Fintech innovation, and make a tangible difference to the small business landscape in India.Find innovative solutions for problems in a yet to be discovered market.
Key Responsibilities
As an early team member, you will get a chance to set the foundations of our engineering culture. You will help articulate our engineering
principles and help set the long-term roadmap. Making decisions on the evolution of CashFlo's technical architectureBuilding new features end to end, from talking to customers to writing code.
Our Ideal Candidate Will Have
3+ years of full-time DevOps engineering experience
Hands-on experience working with AWS services
Deep understanding of virtualization and orchestration tools like Docker, ECS
Experience in writing Infrastructure as Code using tools like CDK, Cloud formation, Terragrunt or Terraform
Experience using centralized logging & monitoring tools such as ELK, CloudWatch, DataDog
Built monitoring dashboards using Prometheus, Grafana
Built and maintained code pipelines and CI/CD
Thorough knowledge of SDLC
Been part of teams that have maintained large deployments
About You
Product-minded. You have a sense for great user experience and feel for when something is off. You love understanding customer pain points
and solving for them.
Get a lot done. You enjoy all aspects of building a product and are comfortable moving across the stack when necessary. You problem solve
independently and enjoy figuring stuff out.
High conviction. When you commit to something, you're in all the way. You're opinionated, but you know when to disagree and commit.
Mediocrity is the worst of all possible outcomes.
Whats in it for me
Gain exposure to the Fintech space - one of the largest and fastest growing markets in India and globally
Shape India’s B2B Payments landscape through cutting edge technology innovation
Be directly responsible for driving company’s success
Join a high performance, dynamic and collaborative work environment that throws new challenges on a daily basis
Fast track your career with trainings, mentoring, growth opportunities on both IC and management track
Work-life balance and fun team events
Experience with Linux infrastructure, opensource databases (mysql, postgres etc), CI/CD tools Jenkins, Gitlab, Nexus Repository, Jira, Agile workflow and Kanban.
- Solid understanding of IP network and TCP/IP
- Minimum of 7+ years’ work experience in IT
- Minimum of 4+ years experience as DevOps Enginneer
- Good interpersonal skills and communication written and oral.
Head Kubernetes
Docker, Kubernetes Engineer – Remote (Pan India)
ABOUT US
Established in 2009, Ashnik is a leading open-source solutions and consulting company in South East Asia and India, headquartered in Singapore. We enable digital transformation for large enterprises through our design, architecting, and solution skills. Over 100 large enterprises in the region have acknowledged our expertise in delivering solutions using key open-source technologies. Our offerings form critical part of Digital transformation, Big Data platform, Cloud and Web acceleration and IT modernization. We represent EDB, Pentaho, Docker, Couchbase, MongoDB, Elastic, NGINX, Sysdig, Redis Labs, Confluent, and HashiCorp as their key partners in the region. Our team members bring decades of experience in delivering confidence to enterprises in adopting open source software and are known for their thought leadership.
THE POSITION
Ashnik is looking for talented and passionate Technical consultant to be part of the training team and work with customers on DevOps Solution. You will be responsible for implementation and consultations related work for customers across SEA and India. We are looking for personnel with personal qualities like -
- Passion for working for different customers and different environments.
- Excellent communication and articulation skills
- Aptitude for learning new technology and willingness to understand technologies which he/she is not directly working on.
- Willingness to travel within and outside the country.
- Ability to independently work at the customer site and navigate through different teams.
SKILLS AND EXPERIENCE
ESSENTIAL SKILLS
- 3+ years of experience with B.E/ B.Tech, MCA, Graduation Degree with higher education in the technical field.
- Must have prior experience with docker containers, swarm and/or Kubernetes.
- Understanding of Operating System, processes, networking, and containers.
- Experience/exposure in writing and managing Dockerfile or creating container images.
- Experience of working on Linux platform.
- Ability to perform installations and system/software configuration on Linux.
- Should be aware of networking and SSL/TLS basics.
- Should be aware of tools used in complex enterprise IT infra e.g. LDAP, AD, Centralized Logging solution etc.
A preferred candidate will have
- Prior experience with open source Kubernetes or another container management platform like OpenShift, EKS, ECS etc.
- CNCF Kubernetes Certified Administrator/Developer
- Experience in container monitoring using Prometheus, Datadog etc.
- Experience with CI/CD tools and their usage
- Knowledge of scripting language e.g., shell scripting
RESPONSIBILITIES
First 2 months:
- Get an in-depth understanding of containers, Docker Engine and runtime, Swarm and Kubernetes.
- Get hands-on experience with various features of Mirantis Kubernetes Engine, and Mirantis Secure Registry
After 2 months:
- Work with Ashnik’s pre-sales and solution architects to deploy Mirantis Kubernetes Engine for customers
- Work with customer teams and provide them guidance on how Kubernetes Platform can be integrated with other tools in their environment (CI/CD, identity management, storage etc)
- Work with customers to containerize their applications.
- Write Dockerfile for customers during the implementation phase.
- Help customers design their network and security policy for effective management of applications deployed using Swarm and/or Kubernetes.
- Help customer design their deployments either with Swarm services or Kubernetes Deployments.
- Work with pre-sales and sales team to help customers during their evaluation of Mirantis Kubernetes platform
- Conduct workshops for customers as needed for technical hand-holding and technical handover.
Package: upto30 L
Office: Mumbai
YOUR ‘OKR’ SUMMARY
OKR means Objective and Key Results.
As a Cloud Engineer, you will understand the overall movement of data in the entire platform, find bottlenecks,define solutions, develop key pieces, write APIs and own deployment of those. You will work with internal and external development teams to discover these opportunities and to solve hard problems. You will also guide engineers in solving the complex problems, developing your acceptance tests for those and reviewing the work and
the test results.
What you will do
- End to End Complete RHOSP Deployment (Under and Overcloud) considered as a NFVi Deployment
- Installing Red Hat’s OpenStack technology using the OSP-director
- Deploying a Red Hat OpenStack Platform based on Red Hat's reference architecture
- Deploying managed hosts with required OpenStack parameters
- Deploying Three (3) Node Highly available (HA) Controller Hosts using Pacemaker and HAP Roxy
- Deploying all the supplied compute hosts that will be hosting multiple VNFs (SRIOV & DPDK)
- Implementing CEPH
- Integrate Software-defined storage (CEPH) with RHOSP, Red Hat OpenStack Platform Operation Tools as
per Industry standard & best practices
- Detailed Network configuration and implementation using Neutron networking - (VXLAN) network type
and Modular Layer 2 (ML2) Open Switch plugin
- Integrating Monitoring Solution with RHOSP
- Design and deployment of common Alarm and Performance Management solution
- Red Hat OpenStack management & monitoring
- VM Alarm and Performance management
- Cloud Management Platform will be configured for day-to-day operational tools to measure uponCPU/Mem/Network utilization etc. from VM level
- Baseline Security Standard) & VA (Vulnerability Assessment) for RHOSP.
Additional Advantage:
- Deep understanding of technology and passionate about what you do.
- Background in designing high performant scalable software systems with strong focus to optimize hardware cost.
- Solid collaborative and interpersonal skills, specifically a proven ability to effectively guide and influence within a dynamic environment.
- Strong commitment to get the most performance out of a system being worked on.
- Prior development of a large software project using service-oriented architecture operating with real time constraints.
What's In It for You
- You will get a chance to work on cloud-native and hyper-scale products
- You will be working with industry leaders in cloud.
- You can expect a steep learning curve.
- You will get the experience of solving real time problems, eventually you become a problem solver.
Benefits & Perks
- Competitive Salary
- Health Insurance
- Open Learning - 100% Reimbursement for online technical courses.
- Fast Growth - opportunities to grow quickly and surely
- Creative Freedom + Flat hierarchy
- Sponsorship to all those employees who represent company in events and meet ups.
- Flexible working hours
- 5 days week
- Hybrid Working model (Office and WFH)
Our Hiring Process
Candidates for this position can expect the hiring process as follows (subject to successful clearing of every round)
- Initial Resume screening call with our Recruiting team
- Next, candidates will be invited to solve coding exercises.
- Next, candidates will be invited for first technical interview
- Next, candidates will be invited for final technical interview
- Finally, candidates will be invited for Culture Plus interview with HR
- Candidates may be asked to interview with the Leadership team
- Successful candidates will subsequently be made an offer via email
As always, the interviews and screening call will be conducted via a mix of telephonic and video call.
So, if you are looking at an opportunity to really make a difference- make it with us…
Coredge.io provides equal employment opportunities to all employees and applicants for employment and prohibits
discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability
status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other
characteristic protected by applicable central, state or local laws.
Job Description
• Minimum 3+ yrs of Experience in DevOps with AWS Platform
• Strong AWS knowledge and experience
• Experience in using CI/CD automation tools (Git, Jenkins, Configuration deployment tools ( Puppet/Chef/Ansible)
• Experience with IAC tools Terraform
• Excellent experience in operating a container orchestration cluster (Kubernetes, Docker)
• Significant experience with Linux operating system environments
• Experience with infrastructure scripting solutions such as Python/Shell scripting
• Must have experience in designing Infrastructure automation framework.
• Good experience in any of the Setting up Monitoring tools and Dashboards ( Grafana/kafka)
• Excellent problem-solving, Log Analysis and troubleshooting skills
• Experience in setting up centralized logging for system (EKS, EC2) and application
• Process-oriented with great documentation skills
• Ability to work effectively within a team and with minimal supervision

Platform Services Engineer
DevSecOps Engineer
- Strong Systems Experience- Linux, networking, cloud, APIs
- Scripting language Programming - Shell, Python
- Strong Debugging Capability
- AWS Platform -IAM, Network,EC2, Lambda, S3, CloudWatch
- Knowledge on Terraform, Packer, Ansible, Jenkins
- Observability - Prometheus, InfluxDB, Dynatrace,
- Grafana, Splunk • DevSecOps-CI/CD - Jenkins
- Microservices
- Security & Access Management
- Container Orchestration a plus - Kubernetes, Docker etc.
- Big Data Platforms knowledge EMR, Databricks. Cloudera a plus
About the Company
- 💰 Early-stage, ed-tech, funded, growing, growing fast
- 🎯 Mission Driven: Make Indonesia competitive on a global scale
- 🥅 Build the best educational content and technology to advance STEM education
- 🥇 Students-First approach
- 🇮🇩 🇮🇳 Teams in India and Indonesia
Skillset 🧗🏼♀️
- You primarily identify as a DevOps/Infrastructure engineer and are comfortable working with systems and cloud-native services on AWS
- You can design, implement, and maintain secure and scalable infrastructure delivering cloud-based services
- You have experience operating and maintaining production systems in a Linux based public cloud environment
- You are familiar with cloud-native concepts - Containers, Lambdas, Orchestration (ECS, Kubernetes)
- You’re in love with system metrics and strive to help deliver improvements to systems all the time
- You can think in terms of Infrastructure as Code to build tools for automating deployment, monitoring, and operations of the platform
- You can be on-call once every few weeks to provide application support, incident management, and troubleshooting
- You’re fairly comfortable with GIT, AWS CLI, python, docker CLI, in general, all things CLI. Oh! Bash scripting too!
- You have high integrity, and you are reliable
What you can expect from us 👌🏼
☮️ Mentorship, growth, great work culture
- Mentorship and continuous improvement are a part of the team’s DNA. We have a battle-tested robust growth framework. You will have people to look up to and people looking up to you
- We are a people-first, high-trust, high-autonomy team
- We live in the TDD, Pair Programming, First Principles world
🌏 Remote done right
- Distributed does not mean working in isolation, feeling alone, being buried in Zoom calls
- Our leadership team has been WFH for 10+ years now and we know how remote teams work. This will be a place to belong
- A good balance between deep focussed work and collaborative work ⚖️
🖥️ Friendly, humane interview process
- 30-minute alignment check and screening call
- A short take-home coding assignment, no more than 2-3 hours. Time is precious
- Pair programming interview. Collaborate, work together. No sitting behind a desk and judging
- In-depth engineering discussion around your skills and career so far
- System design and architecture interview for seniors
What we ask from you👇🏼
- Bring your software engineering — both individual brilliance and collaborative skills
- Bring your good nature — we're building a team that supports each other
- Be vested or interested in the company vision









