Cutshort logo
NAT Jobs in Chennai

11+ NAT Jobs in Chennai | NAT Job openings in Chennai

Apply to 11+ NAT Jobs in Chennai on CutShort.io. Explore the latest NAT Job opportunities across top companies like Google, Amazon & Adobe.

icon
enterprise-grade, streaming integration with intelligence pl

enterprise-grade, streaming integration with intelligence pl

Agency job
via Jobdost by Mamatha A
Chennai
2 - 5 yrs
₹5L - ₹15L / yr
DevOps
skill iconDocker
skill iconKubernetes
CI/CD
skill iconGo Programming (Golang)
+13 more

Striim (pronounced “stream” with two i’s for integration and intelligence) was founded in 2012 with a simple goal of helping companies make data useful the instant it’s born.

Striim’s enterprise-grade, streaming integration with intelligence platform makes it easy to build continuous, streaming data pipelines – including change data capture (CDC) – to power real-time cloud integration, log correlation, edge processing, and streaming analytics

2 - 5 Years of Experience in any Programming any language (Polyglot Preferred ) & System Operations • Awareness of Devops & Agile Methodologies • Proficient in leveraging CI and CD tools to automate testing and deployment . • Experience in working in an agile and fast paced environment . • Hands on knowledge of at least one cloud platform (AWS / GCP / Azure). • Cloud networking knowledge: should understand VPC, NATs, and routers. • Contributions to open source is a plus. • Good written communication skills are a must. Contributions to technical blogs / whitepapers will be an added advantage.

Read more
Global Digital Transformation Solutions Provider

Global Digital Transformation Solutions Provider

Agency job
via Peak Hire Solutions by Dhara Thakkar
Bengaluru (Bangalore), Chennai, Hyderabad, Kochi (Cochin), Noida, Pune, Thiruvananthapuram
7 - 10 yrs
₹21L - ₹30L / yr
Perforce
DevOps
skill iconGit
skill iconGitHub
skill iconPython
+7 more

JOB DETAILS:

* Job Title: Specialist I - DevOps Engineering

* Industry: Global Digital Transformation Solutions Provider

* Salary: Best in Industry

* Experience: 7-10 years

* Location: Bengaluru (Bangalore), Chennai, Hyderabad, Kochi (Cochin), Noida, Pune, Thiruvananthapuram

 

Job Description

Job Summary:

As a DevOps Engineer focused on Perforce to GitHub migration, you will be responsible for executing seamless and large-scale source control migrations. You must be proficient with GitHub Enterprise and Perforce, possess strong scripting skills (Python/Shell), and have a deep understanding of version control concepts.

The ideal candidate is a self-starter, a problem-solver, and thrives on challenges while ensuring smooth transitions with minimal disruption to development workflows.

 

Key Responsibilities:

  • Analyze and prepare Perforce repositories — clean workspaces, merge streams, and remove unnecessary files.
  • Handle large files efficiently using Git Large File Storage (LFS) for files exceeding GitHub’s 100MB size limit.
  • Use git-p4 fusion (Python-based tool) to clone and migrate Perforce repositories incrementally, ensuring data integrity.
  • Define migration scope — determine how much history to migrate and plan the repository structure.
  • Manage branch renaming and repository organization for optimized post-migration workflows.
  • Collaborate with development teams to determine migration points and finalize migration strategies.
  • Troubleshoot issues related to file sizes, Python compatibility, network connectivity, or permissions during migration.

 

Required Qualifications:

  • Strong knowledge of Git/GitHub and preferably Perforce (Helix Core) — understanding of differences, workflows, and integrations.
  • Hands-on experience with P4-Fusion.
  • Familiarity with cloud platforms (AWS, Azure) and containerization technologies (Docker, Kubernetes).
  • Proficiency in migration tools such as git-p4 fusion — installation, configuration, and troubleshooting.
  • Ability to identify and manage large files using Git LFS to meet GitHub repository size limits.
  • Strong scripting skills in Python and Shell for automating migration and restructuring tasks.
  • Experience in planning and executing source control migrations — defining scope, branch mapping, history retention, and permission translation.
  • Familiarity with CI/CD pipeline integration to validate workflows post-migration.
  • Understanding of source code management (SCM) best practices, including version history and repository organization in GitHub.
  • Excellent communication and collaboration skills for cross-team coordination and migration planning.
  • Proven practical experience in repository migration, large file management, and history preservation during Perforce to GitHub transitions.

 

Skills: Github, Kubernetes, Perforce, Perforce (Helix Core), Devops Tools

 

Must-Haves

Git/GitHub (advanced), Perforce (Helix Core) (advanced), Python/Shell scripting (strong), P4-Fusion (hands-on experience), Git LFS (proficient)

Read more
Zolvit (formerly Vakilsearch)

at Zolvit (formerly Vakilsearch)

1 video
2 recruiters
Lakshmi J
Posted by Lakshmi J
Chennai
2 - 4 yrs
₹10L - ₹16L / yr
DevOps
Linux administration
Unix administration
Shell Scripting
CI/CD
+5 more

We are looking for a passionate DevOps Engineer who can support deployment and monitor our Production, QE, and Staging environments performance. Applicants should have a strong understanding of UNIX internals and should be able to clearly articulate how it works. Knowledge of shell scripting & security aspects is a must. Any experience with infrastructure as code is a big plus. The key responsibility of the role is to manage deployments, security, and support of business solutions. Having experience in database applications like Postgres, ELK, NodeJS, NextJS & Ruby on Rails is a huge plus. At VakilSearch. Experience doesn't matter, passion to produce change matters



Responsibilities and Accountabilities:

  • As part of the DevOps team, you will be responsible for configuration, optimization, documentation, and support of the infra components of VakilSearch’s product which are hosted in cloud services & on-prem facility
  • Design, build tools and framework that support deploying and managing our platform & Exploring new tools, technologies, and processes to improve speed, efficiency, and scalability
  • Support and troubleshoot scalability, high availability, performance, monitoring, backup, and restore of different Env 
  • Manage resources in a cost-effective, innovative manner including assisting subordinates ineffective use of resources and tools
  • Resolve incidents as escalated from Monitoring tools and Business Development Team
  • Implement and follow security guidelines, both policy and technology to protect our data
  • Identify root cause for issues and develop long-term solutions to fix recurring issues and Document it
  • Strong in performing production operation activities even at night times if required
  • Ability to automate [Scripts] recurring tasks to increase velocity and quality
  • Ability to manage and deliver multiple project phases at the same time

I Qualification(s): 

  • Experience in working with Linux Server, DevOps tools, and Orchestration tools 
  • Linux, AWS, GCP, Azure, CompTIA+, and any other certification are a value-add 

II Experience Required in DevOps Aspects:

  • Length of Experience: Minimum 1-4 years of experience
  • Nature of Experience: 
  • Experience in Cloud deployments, Linux administration[ Kernel Tuning is a value add ], Linux clustering, AWS, virtualization, and networking concepts [ Azure, GCP value add ]
  • Experience in deployment solutions CI/CD like Jenkins, GitHub Actions [ Release Management is a value add ]
  • Hands-on experience in any of the configuration management IaC tools like Chef, Terraform, and CloudFormation [ Ansible & Puppet is a value add ]
  • Administration, Configuring and utilizing Monitoring and Alerting tools like Prometheus, Grafana, Loki, ELK, Zabbix, Datadog, etc
  • Experience with Containerization and orchestration tools like Docker, and Kubernetes [ Docker swarm is a value add ]Good Scripting skills in at least one interpreted language - Shell/bash scripting or Ruby/Python/Perl
  • Experience in Database applications like PostgreSQL, MongoDB & MySQL [DataOps]
  • Good at Version Control & source code management systems like GitHub, GIT
  • Experience in Serverless [ Lambda/GCP cloud function/Azure function ]
  • Experience in Web Server Nginx, and Apache
  • Knowledge in Redis, RabbitMQ, ELK, REST API [ MLOps Tools is a value add ]
  • Knowledge in Puma, Unicorn, Gunicorn & Yarn
  • Hands-on VMWare ESXi/Xencenter deployments is a value add
  • Experience in Implementing and troubleshooting TCP/IP networks, VPN, Load Balancing & Web application firewalls
  • Deploying, Configuring, and Maintaining Linux server systems ON premises and off-premises
  • Code Quality like SonarQube is a value-add
  • Test Automation like Selenium, JMeter, and JUnit is a value-add
  • Experience in Heroku and OpenStack is a value-add 
  • Experience in Identifying Inbound and Outbound Threats and resolving it
  • Knowledge of CVE & applying the patches for OS, Ruby gems, Node, and Python packages  
  • Documenting the Security fix for future use
  • Establish cross-team collaboration with security built into the software development lifecycle 
  • Forensics and Root Cause Analysis skills are mandatory 
  • Weekly Sanity Checks of the on-prem and off-prem environment 

 

III Skill Set & Personality Traits required:

  • An understanding of programming languages such as Ruby, NodeJS, ReactJS, Perl, Java, Python, and PHP
  • Good written and verbal communication skills to facilitate efficient and effective interaction with peers, partners, vendors, and customers


IV Age Group: 21 – 36 Years


V Cost to the Company: As per industry standards


Read more
Deqode

at Deqode

1 recruiter
Shubham Das
Posted by Shubham Das
Mumbai, Chennai, Gurugram
6 - 9 yrs
₹12L - ₹17L / yr
skill iconAmazon Web Services (AWS)
Windows Azure
Google Cloud Platform (GCP)
DevOps
helm
+3 more

We are looking for a highly skilled DevOps/Cloud Engineer with over 6 years of experience in infrastructure automation, cloud platforms, networking, and security. If you are passionate about designing scalable systems and love solving complex cloud and DevOps challenges—this opportunity is for you.

Key Responsibilities

  • Design, deploy, and manage cloud-native infrastructure using Kubernetes (K8s), Helm, Terraform, and Ansible
  • Automate provisioning and orchestration workflows for cloud and hybrid environments
  • Manage and optimize deployments on AWS, Azure, and GCP for high availability and cost efficiency
  • Troubleshoot and implement advanced network architectures including VPNs, firewalls, load balancers, and routing protocols
  • Implement and enforce security best practices: IAM, encryption, compliance, and vulnerability management
  • Collaborate with development and operations teams to improve CI/CD workflows and system observability

Required Skills & Qualifications

  • 6+ years of experience in DevOps, Infrastructure as Code (IaC), and cloud-native systems
  • Expertise in Helm, Terraform, and Kubernetes
  • Strong hands-on experience with AWS and Azure
  • Solid understanding of networking, firewall configurations, and security protocols
  • Experience with CI/CD tools like Jenkins, GitHub Actions, or similar
  • Strong problem-solving skills and a performance-first mindset

Why Join Us?

  • Work on cutting-edge cloud infrastructure across diverse industries
  • Be part of a collaborative, forward-thinking team
  • Flexible hybrid work model – work from anywhere while staying connected
  • Opportunity to take ownership and lead critical DevOps initiatives


Read more
Indian private sector bank

Indian private sector bank

Agency job
via Pluginlive by Harsha Saggi
Chennai
5 - 10 yrs
₹10L - ₹30L / yr
Terraform
DevOps
skill iconAmazon Web Services (AWS)
skill iconDocker
skill iconKubernetes

Roles and Responsibilities:

  • AWS Cloud Management: Design, deploy, and manage AWS cloud infrastructure. Optimize and maintain cloud resources for performance and cost efficiency. Monitor and ensure the security of cloud-based systems.
  • Automated Provisioning: Develop and implement automated provisioning processes for infrastructure deployment. Utilize tools like Terraform and Packer to automate and streamline the provisioning of resources.
  • Infrastructure as Code (IaC): Champion the use of Infrastructure as Code principles. Collaborate with development and operations teams to define and maintain IaC scripts for infrastructure deployment and configuration.
  • Collaboration and Communication: Work closely with cross-functional teams to understand project requirements and provide DevOps expertise. Communicate effectively with team members and stakeholders regarding infrastructure changes, updates, and improvements.
  • Continuous Integration/Continuous Deployment (CI/CD): Implement and maintain CI/CD pipelines to automate software delivery processes. Ensure reliable and efficient deployment of applications through the development lifecycle.
  • Performance Monitoring and Optimization: Implement monitoring solutions to track system performance, troubleshoot issues, and optimize resource utilization. Proactively identify opportunities for system and process improvements. 


Mandatory Skills:

  • Proven experience as a DevOps Engineer or similar role, with a focus on AWS.
  • Strong proficiency in automated provisioning and cloud management.
  • Experience with Infrastructure as Code tools, particularly Terraform and Packer.
  • Solid understanding of CI/CD pipelines and version control systems.
  • Strong scripting skills (e.g., Python, Bash) for automation tasks.
  • Excellent problem-solving and troubleshooting skills.
  • Good interpersonal and communication skills for effective collaboration.


Secondary Skills:

  • AWS certifications (e.g., AWS Certified DevOps Engineer, AWS Certified Solutions Architect).
  • Experience with containerization and orchestration tools (e.g., Docker, Kubernetes).
  • Knowledge of microservices architecture and serverless computing.
  • Familiarity with monitoring and logging tools (e.g., CloudWatch, ELK stack).


Read more
VyTCDC
Gobinath Sundaram
Posted by Gobinath Sundaram
Chennai, Coimbatore, Bengaluru (Bangalore)
5 - 8 yrs
₹4.5L - ₹22L / yr
DevOps
Ansible
skill iconKubernetes
skill iconJenkins
Bash
+3 more

Job Description

What does a successful Senior DevOps Engineer do at Fiserv?

This role’s focus will be on contributing and enhancing our DevOps environment within Issuer Solution group, where our cross functional Scrum teams are delivering solutions built on cutting-edge mobile technology and products. You will be expected to support across the wider business unit, leading DevOps practices and initiatives.

 

What will you do:

 Build, manage, and deploy CI/CD pipelines.

 DevOps Engineer - Helm Chart, Rundesk, Openshift

 Strive for continuous improvement and build continuous integration, continuous development, and constant deployment pipeline.

 Implementing various development, testing, automation tools, and IT infrastructure

 Optimize and automate release/development cycles and processes.

 Be part of and help promote our DevOps culture.

 Identify and implement continuous improvements to the development practice

 

What you must have:

 3+ years of experience in devops with hands-on experience in the following:

- Writing automation scripts for deployments and housekeeping using shell scripts (bash) and ansible playbooks

- Building docker images and running/managing docker instances

- Building Jenkins pipelines using groovy scripts

- Working knowledge on kubernetes including application deployments, managing application configurations and persistence volumes

 Has good understanding on infrastructure as code

 Ability to write and update documentation

 Demonstrate a logical, process orientated approach to problems and troubleshooting

 Ability to collaborate with multi development teams

 

What you are preferred to have:

 8+ years of development experience

 Jenkins administration experience

 Hands-on experience in building and deploying helm charts

Process Skills:

• Should have worked in Agile Project

 

Behavioral Skills :

• Good Communication skills

 

Skills

PRIMARY COMPETENCY : Cloud Infra PRIMARY SKILL : DevOps PRIMARY SKILL PERCENTAGE : 100

Read more
Leading Payment Solution Company

Leading Payment Solution Company

Agency job
via People First Consultants by Aishwarya KA
Chennai, Bengaluru (Bangalore), Pune, Hyderabad, Mumbai
9 - 16 yrs
Best in industry
skill iconDocker
skill iconKubernetes
DevOps
skill iconAmazon Web Services (AWS)
Microsoft Windows Azure
+9 more

About Company:

The company is a global leader in secure payments and trusted transactions. They are at the forefront of the digital revolution that is shaping new ways of paying, living, doing business and building relationships that pass on trust along the entire payments value chain, enabling sustainable economic growth. Their innovative solutions, rooted in a rock-solid technological base, are environmentally friendly, widely accessible and support social transformation.

  • Role Overview
    • Senior Engineer with a strong background and experience in cloud related technologies and architectures. Can design target cloud architectures to transform existing architectures together with the in-house team. Can actively hands-on configure and build cloud architectures and guide others.
  • Key Knowledge
    • 3-5+ years of experience in AWS/GCP or Azure technologies
    • Is likely certified on one or more of the major cloud platforms
    • Strong experience from hands-on work with technologies such as Terraform, K8S, Docker and orchestration of containers.
    • Ability to guide and lead internal agile teams on cloud technology
    • Background from the financial services industry or similar critical operational experience
Read more
Leading Payment Solution Company

Leading Payment Solution Company

Agency job
Remote, Bengaluru (Bangalore), Chennai, Pune, Hyderabad, Mumbai
3 - 10 yrs
₹8L - ₹28L / yr
skill iconDocker
skill iconKubernetes
DevOps
skill iconAmazon Web Services (AWS)
Windows Azure
+3 more

Experience: 3+ years of experience in Cloud Architecture

About Company:

The company is a global leader in secure payments and trusted transactions. They are at the forefront of the digital revolution that is shaping new ways of paying, living, doing business and building relationships that pass on trust along the entire payments value chain, enabling sustainable economic growth. Their innovative solutions, rooted in a rock-solid technological base, are environmentally friendly, widely accessible and support social transformation.



Cloud Architect / Lead

  • Role Overview
    • Senior Engineer with a strong background and experience in cloud related technologies and architectures. Can design target cloud architectures to transform existing architectures together with the in-house team. Can actively hands-on configure and build cloud architectures and guide others.
  • Key Knowledge
    • 3-5+ years of experience in AWS/GCP or Azure technologies
    • Is likely certified on one or more of the major cloud platforms
    • Strong experience from hands-on work with technologies such as Terraform, K8S, Docker and orchestration of containers.
    • Ability to guide and lead internal agile teams on cloud technology
    • Background from the financial services industry or similar critical operational experience
 
Read more
HappyFox

at HappyFox

1 video
6 products
Lindsey A
Posted by Lindsey A
Chennai, Bengaluru (Bangalore)
5 - 10 yrs
₹10L - ₹15L / yr
DevOps
skill iconKubernetes
skill iconDocker
skill iconAmazon Web Services (AWS)
Windows Azure
+12 more

About us:

HappyFox is a software-as-a-service (SaaS) support platform. We offer an enterprise-grade help desk ticketing system and intuitively designed live chat software.

 

We serve over 12,000 companies in 70+ countries. HappyFox is used by companies that span across education, media, e-commerce, retail, information technology, manufacturing, non-profit, government and many other verticals that have an internal or external support function.

 

To know more, Visit! - https://www.happyfox.com/

 

Responsibilities:

  • Build and scale production infrastructure in AWS for the HappyFox platform and its products.
  • Research, Build/Implement systems, services and tooling to improve uptime, reliability and maintainability of our backend infrastructure. And to meet our internal SLOs and customer-facing SLAs.
  • Proficient in managing/patching servers with Unix-based operating systems like Ubuntu Linux.
  • Proficient in writing automation scripts or building infrastructure tools using Python/Ruby/Bash/Golang
  • Implement consistent observability, deployment and IaC setups
  • Patch production systems to fix security/performance issues
  • Actively respond to escalations/incidents in the production environment from customers or the support team
  • Mentor other Infrastructure engineers, review their work and continuously ship improvements to production infrastructure.
  • Build and manage development infrastructure, and CI/CD pipelines for our teams to ship & test code faster.
  • Participate in infrastructure security audits

 

Requirements:

  • At least 5 years of experience in handling/building Production environments in AWS.
  • At least 2 years of programming experience in building API/backend services for customer-facing applications in production.
  • Demonstrable knowledge of TCP/IP, HTTP and DNS fundamentals.
  • Experience in deploying and managing production Python/NodeJS/Golang applications to AWS EC2, ECS or EKS.
  • Proficient in containerised environments such as Docker, Docker Compose, Kubernetes
  • Proficient in managing/patching servers with Unix-based operating systems like Ubuntu Linux.
  • Proficient in writing automation scripts using any scripting language such as Python, Ruby, Bash etc.,
  • Experience in setting up and managing test/staging environments, and CI/CD pipelines.
  • Experience in IaC tools such as Terraform or AWS CDK
  • Passion for making systems reliable, maintainable, scalable and secure.
  • Excellent verbal and written communication skills to address, escalate and express technical ideas clearly
  • Bonus points – if you have experience with Nginx, Postgres, Redis, and Mongo systems in production.

 

Read more
Reputed client of People First

Reputed client of People First

Agency job
Chennai, Bengaluru (Bangalore), Gurugram
2 - 9 yrs
₹10L - ₹30L / yr
DevOps
skill iconKubernetes
skill iconDocker
skill iconJenkins
CI/CD
+4 more
Job Description:
As a MLOps Engineer in QuantumBlack you will:

Develop and deploy technology that enables data scientists and data engineers to build, productionize and deploy machine learning models following best practices. Work to set the standards for SWE and
DevOps practices within multi-disciplinary delivery teams

Choose and use the right cloud services, DevOps tooling and ML tooling for the team to be able to produce high-quality code that allows your team to release to production
.
Build modern, scalable, and secure CI/CD pipelines to automate development and deployment
workflows used by data scientists (ML pipelines) and data engineers (Data pipelines)

Shape and support next generation technology that enables scaling ML products and platforms. Bring
expertise in cloud to enable ML use case development, including MLOps

Our Tech Stack-

We leverage AWS, Google Cloud, Azure, Databricks, Docker, Kubernetes, Argo, Airflow, Kedro, Python,
Terraform, GitHub actions, MLFlow, Node.JS, React, Typescript amongst others in our projects

Key Skills:

• Excellent hands-on expert knowledge of cloud platform infrastructure and administration
(Azure/AWS/GCP) with strong knowledge of cloud services integration, and cloud security

• Expertise setting up CI/CD processes, building and maintaining secure DevOps pipelines with at
least 2 major DevOps stacks (e.g., Azure DevOps, Gitlab, Argo)

• Experience with modern development methods and tooling: Containers (e.g., docker) and
container orchestration (K8s), CI/CD tools (e.g., Circle CI, Jenkins, GitHub actions, Azure
DevOps), version control (Git, GitHub, GitLab), orchestration/DAGs tools (e.g., Argo, Airflow,
Kubeflow)

• Hands-on coding skills Python 3 (e.g., API including automated testing frameworks and libraries
(e.g., pytest) and Infrastructure as Code (e.g., Terraform) and Kubernetes artifacts (e.g.,
deployments, operators, helm charts)

• Experience setting up at least one contemporary MLOps tooling (e.g., experiment tracking,
model governance, packaging, deployment, feature store)

• Practical knowledge delivering and maintaining production software such as APIs and cloud
infrastructure

• Knowledge of SQL (intermediate level or more preferred) and familiarity working with at least
one common RDBMS (MySQL, Postgres, SQL Server, Oracle)

Read more
Our client company is into Financial services. ( O1)

Our client company is into Financial services. ( O1)

Agency job
via Multi Recruit by Kavitha S
Chennai
1 - 4 yrs
₹12L - ₹15L / yr
DevOps
Terraform
skill iconDocker
skill iconKubernetes
Linux/Unix
+2 more
  • Experience working on Linux based infrastructure
  • Strong hands-on knowledge of setting up production, staging, and dev environments on AWS/GCP/Azure
  • Strong hands-on knowledge of technologies like Terraform, Docker, Kubernetes
  • Strong understanding of continuous testing environments such as Travis-CI, CircleCI, Jenkins, etc.
  • Configuration and managing databases such as MySQL, Mongo
  • Excellent troubleshooting
  • Working knowledge of various tools, open-source technologies, and cloud services
  • Awareness of critical concepts in DevOps and Agile principles
Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort