Cutshort logo

11+ SCA Jobs in India

Apply to 11+ SCA Jobs on CutShort.io. Find your next job, effortlessly. Browse SCA Jobs and apply today!

icon
AdTech Industry

AdTech Industry

Agency job
via Peak Hire Solutions by Dhara Thakkar
Noida
8 - 12 yrs
₹50L - ₹75L / yr
Ansible
Terraform
skill iconAmazon Web Services (AWS)
Platform as a Service (PaaS)
CI/CD
+30 more

ROLE & RESPONSIBILITIES:

We are hiring a Senior DevSecOps / Security Engineer with 8+ years of experience securing AWS cloud, on-prem infrastructure, DevOps platforms, MLOps environments, CI/CD pipelines, container orchestration, and data/ML platforms. This role is responsible for creating and maintaining a unified security posture across all systems used by DevOps and MLOps teams — including AWS, Kubernetes, EMR, MWAA, Spark, Docker, GitOps, observability tools, and network infrastructure.


KEY RESPONSIBILITIES:

1.     Cloud Security (AWS)-

  • Secure all AWS resources consumed by DevOps/MLOps/Data Science: EC2, EKS, ECS, EMR, MWAA, S3, RDS, Redshift, Lambda, CloudFront, Glue, Athena, Kinesis, Transit Gateway, VPC Peering.
  • Implement IAM least privilege, SCPs, KMS, Secrets Manager, SSO & identity governance.
  • Configure AWS-native security: WAF, Shield, GuardDuty, Inspector, Macie, CloudTrail, Config, Security Hub.
  • Harden VPC architecture, subnets, routing, SG/NACLs, multi-account environments.
  • Ensure encryption of data at rest/in transit across all cloud services.

 

2.     DevOps Security (IaC, CI/CD, Kubernetes, Linux)-

Infrastructure as Code & Automation Security:

  • Secure Terraform, CloudFormation, Ansible with policy-as-code (OPA, Checkov, tfsec).
  • Enforce misconfiguration scanning and automated remediation.

CI/CD Security:

  • Secure Jenkins, GitHub, GitLab pipelines with SAST, DAST, SCA, secrets scanning, image scanning.
  • Implement secure build, artifact signing, and deployment workflows.

Containers & Kubernetes:

  • Harden Docker images, private registries, runtime policies.
  • Enforce EKS security: RBAC, IRSA, PSP/PSS, network policies, runtime monitoring.
  • Apply CIS Benchmarks for Kubernetes and Linux.

Monitoring & Reliability:

  • Secure observability stack: Grafana, CloudWatch, logging, alerting, anomaly detection.
  • Ensure audit logging across cloud/platform layers.


3.     MLOps Security (Airflow, EMR, Spark, Data Platforms, ML Pipelines)-

Pipeline & Workflow Security:

  • Secure Airflow/MWAA connections, secrets, DAGs, execution environments.
  • Harden EMR, Spark jobs, Glue jobs, IAM roles, S3 buckets, encryption, and access policies.

ML Platform Security:

  • Secure Jupyter/JupyterHub environments, containerized ML workspaces, and experiment tracking systems.
  • Control model access, artifact protection, model registry security, and ML metadata integrity.

Data Security:

  • Secure ETL/ML data flows across S3, Redshift, RDS, Glue, Kinesis.
  • Enforce data versioning security, lineage tracking, PII protection, and access governance.

ML Observability:

  • Implement drift detection (data drift/model drift), feature monitoring, audit logging.
  • Integrate ML monitoring with Grafana/Prometheus/CloudWatch.


4.     Network & Endpoint Security-

  • Manage firewall policies, VPN, IDS/IPS, endpoint protection, secure LAN/WAN, Zero Trust principles.
  • Conduct vulnerability assessments, penetration test coordination, and network segmentation.
  • Secure remote workforce connectivity and internal office networks.


5.     Threat Detection, Incident Response & Compliance-

  • Centralize log management (CloudWatch, OpenSearch/ELK, SIEM).
  • Build security alerts, automated threat detection, and incident workflows.
  • Lead incident containment, forensics, RCA, and remediation.
  • Ensure compliance with ISO 27001, SOC 2, GDPR, HIPAA (as applicable).
  • Maintain security policies, procedures, RRPs (Runbooks), and audits.


IDEAL CANDIDATE:

  • 8+ years in DevSecOps, Cloud Security, Platform Security, or equivalent.
  • Proven ability securing AWS cloud ecosystems (IAM, EKS, EMR, MWAA, VPC, WAF, GuardDuty, KMS, Inspector, Macie).
  • Strong hands-on experience with Docker, Kubernetes (EKS), CI/CD tools, and Infrastructure-as-Code.
  • Experience securing ML platforms, data pipelines, and MLOps systems (Airflow/MWAA, Spark/EMR).
  • Strong Linux security (CIS hardening, auditing, intrusion detection).
  • Proficiency in Python, Bash, and automation/scripting.
  • Excellent knowledge of SIEM, observability, threat detection, monitoring systems.
  • Understanding of microservices, API security, serverless security.
  • Strong understanding of vulnerability management, penetration testing practices, and remediation plans.


EDUCATION:

  • Master’s degree in Cybersecurity, Computer Science, Information Technology, or related field.
  • Relevant certifications (AWS Security Specialty, CISSP, CEH, CKA/CKS) are a plus.


PERKS, BENEFITS AND WORK CULTURE:

  • Competitive Salary Package
  • Generous Leave Policy
  • Flexible Working Hours
  • Performance-Based Bonuses
  • Health Care Benefits
Read more
Spark Eighteen
Rishabh Jain
Posted by Rishabh Jain
Delhi
5 - 10 yrs
₹23L - ₹30L / yr
cicd
skill iconAmazon Web Services (AWS)
skill iconDocker
skill iconKubernetes
skill iconJenkins
+2 more

About the Job

This is a full-time role for a Lead DevOps Engineer at Spark Eighteen. We are seeking an experienced DevOps professional to lead our infrastructure strategy, design resilient systems, and drive continuous improvement in our deployment processes. In this role, you will architect scalable solutions, mentor junior engineers, and ensure the highest standards of reliability and security across our cloud infrastructure. The job location is flexible with preference for the Delhi NCR region.


Responsibilities

  • Lead and mentor the DevOps/SRE team
  • Define and drive DevOps strategy and roadmaps
  • Oversee infrastructure automation and CI/CD at scale
  • Collaborate with architects, developers, and QA teams to integrate DevOps practices
  • Ensure security, compliance, and high availability of platforms
  • Own incident response, postmortems, and root cause analysis
  • Budgeting, team hiring, and performance evaluation


Requirements

Technical Skills

  • Bachelor's or Master's degree in Computer Science, Engineering, or related field.
  • 7+ years of professional DevOps experience with demonstrated progression.
  • Strong architecture and leadership background
  • Deep hands-on knowledge of infrastructure as code, CI/CD, and cloud
  • Proven experience with monitoring, security, and governance
  • Effective stakeholder and project management
  • Experience with tools like Jenkins, ArgoCD, Terraform, Vault, ELK, etc.
  • Strong understanding of business continuity and disaster recovery


Soft Skills

  • Cross-functional communication excellence with ability to lead technical discussions.
  • Strong mentorship capabilities for junior and mid-level team members.
  • Advanced strategic thinking and ability to propose innovative solutions.
  • Excellent knowledge transfer skills through documentation and training.
  • Ability to understand and align technical solutions with broader business strategy.
  • Proactive problem-solving approach with focus on continuous improvement.
  • Strong leadership skills in guiding team performance and technical direction.
  • Effective collaboration across development, QA, and business teams.
  • Ability to make complex technical decisions with minimal supervision.
  • Strategic approach to risk management and mitigation.


What We Offer

  • Professional Growth: Continuous learning opportunities through diverse projects and mentorship from experienced leaders
  • Global Exposure: Work with clients from 20+ countries, gaining insights into different markets and business cultures
  • Impactful Work: Contribute to projects that make a real difference, with solutions generating over $1B in revenue
  • Work-Life Balance: Flexible arrangements that respect personal wellbeing while fostering productivity
  • Career Advancement: Clear progression pathways as you develop skills within our growing organization
  • Competitive Compensation: Attractive salary packages that recognize your contributions and expertise


Our Culture

At Spark Eighteen, our culture centers on innovation, excellence, and growth. We believe in:

  • Quality-First: Delivering excellence rather than just quick solutions
  • True Partnership: Building relationships based on trust and mutual respect
  • Communication: Prioritizing clear, effective communication across teams
  • Innovation: Encouraging curiosity and creative approaches to problem-solving
  • Continuous Learning: Supporting professional development at all levels
  • Collaboration: Combining diverse perspectives to achieve shared goals
  • Impact: Measuring success by the value we create for clients and users


Apply Here - https://tinyurl.com/t6x23p9b

Read more
Glan management Consultancy
Gurugram
7 - 15 yrs
₹20L - ₹40L / yr
DevOps
aws
Terraform
skill iconMongoDB

Job Title: AWS Devops Engineer – Manager Business solutions

Location: Gurgaon, India

Experience Required: 8-12 years

Industry: IT


 

We are looking for a seasoned AWS DevOps Engineer with robust experience in AWS middleware services and MongoDB Cloud Infrastructure Management. The role involves designing, deploying, and maintaining secure, scalable, and high-availability infrastructure, along with developing efficient CI/CD pipelines and automating operational processes.

 

Key Deliverables (Essential functions & Responsibilities of the Job):

 

·         Design, deploy, and manage AWS infrastructure, with a focus on middleware services such as API Gateway, Lambda, SQS, SNS, ECS, and EKS.

·         Administer and optimize MongoDB Atlas or equivalent cloud-based MongoDB solutions for high availability, security, and performance.

·         Develop, manage, and enhance CI/CD pipelines using tools like AWS CodePipeline, Jenkins, GitHub Actions, GitLab CI/CD, or Bitbucket Pipelines.

·         Automate infrastructure provisioning using Terraform, AWS CloudFormation, or AWS CDK.

·         Implement monitoring and logging solutions using CloudWatch, Prometheus, Grafana, or the ELK Stack.

·         Enforce cloud security best practices — IAM, VPC setups, encryption, certificate management, and compliance controls.

·         Work closely with development teams to improve application reliability, scalability, and performance.

·         Manage containerized environments using Docker, Kubernetes (EKS), or AWS ECS.

·         Perform MongoDB administration tasks such as backups, performance tuning, indexing, and sharding.

·         Participate in on-call rotations to ensure 24/7 infrastructure availability and quick incident resolution.

 

Knowledge Skills and Abilities:

·         7+ years of hands-on AWS DevOps experience, especially with middleware services.

·         Strong expertise in MongoDB Atlas or other cloud MongoDB services.

·         Proficiency in Infrastructure as Code (IaC) tools like Terraform, CloudFormation, or AWS CDK.

·         Solid experience with CI/CD tools: Jenkins, CodePipeline, GitHub Actions, GitLab, Bitbucket, etc.

·         Excellent scripting skills in Python, Bash, or PowerShell.

·         Experience in containerization and orchestration: Docker, EKS, ECS.

·         Familiarity with monitoring tools like CloudWatch, ELK, Prometheus, Grafana.

·         Strong understanding of AWS networking and security: IAM, VPC, KMS, Security Groups.

·         Ability to solve complex problems and thrive in a fast-paced environment.

 

Preferred Qualifications

·         AWS Certified DevOps Engineer – Professional or AWS Solutions Architect – Associate/Professional.

·         MongoDB Certified DBA or Developer.

·         Experience with serverless services like AWS Lambda, Step Functions.

·         Exposure to multi-cloud or hybrid cloud environments.




Mail updated resume with current salary-

Email: etalenthire[at]gmail[dot]com

Satish; 88O 27 49 743

Read more
Product Base Company into Logistic

Product Base Company into Logistic

Agency job
via Qrata by Rayal Rajan
Mumbai, Navi Mumbai
8 - 10 yrs
₹20L - ₹30L / yr
DevOps
skill iconKubernetes
skill iconDocker
skill iconAmazon Web Services (AWS)
Linux/Unix
+14 more

Role : Principal Devops Engineer


About the Client


It is a Product base company that has to build a platform using AI and ML technology for their transportation and logiticsThey also have a presence in the global market


Responsibilities and Requirements


• Experience in designing and maintaining high volume and scalable micro-services architecture on cloud infrastructure

• Knowledge in Linux/Unix Administration and Python/Shell Scripting

• Experience working with cloud platforms like AWS (EC2, ELB, S3, Auto-scaling, VPC, Lambda), GCP, Azure

• Knowledge in deployment automation, Continuous Integration and Continuous Deployment (Jenkins, Maven, Puppet, Chef, GitLab) and monitoring tools like Zabbix, Cloud Watch Monitoring, Nagios

• Knowledge of Java Virtual Machines, Apache Tomcat, Nginx, Apache Kafka, Microservices architecture, Caching mechanisms

• Experience in enterprise application development, maintenance and operations

• Knowledge of best practices and IT operations in an always-up, always-available service

• Excellent written and oral communication skills, judgment and decision-making skill

Read more
Classplus

at Classplus

1 video
4 recruiters
Peoples Office
Posted by Peoples Office
Noida
5 - 8 yrs
Best in industry
skill iconDocker
skill iconKubernetes
DevOps
skill iconAmazon Web Services (AWS)
Google Cloud Platform (GCP)
+11 more
About us

Classplus is India's largest B2B ed-tech start-up, enabling 1 Lac+ educators and content creators to create their digital identity with their own branded apps. Starting in 2018, we have grown more than 10x in the last year, into India's fastest-growing video learning platform.
Over the years, marquee investors like Tiger Global, Surge, GSV Ventures, Blume, Falcon, Capital, RTP Global, and Chimera Ventures have supported our vision. Thanks to our awesome and dedicated team, we achieved a major milestone in March this year when we secured a “Series-D” funding.

Now as we go global, we are super excited to have new folks on board who can take the rocketship higher🚀. Do you think you have what it takes to help us achieve this? Find Out Below!

What will you do?

Define the overall process, which includes building a team for DevOps activities and ensuring that infrastructure changes are reviewed from an architecture and security perspective

Create standardized tooling and templates for development teams to create CI/CD pipelines

Ensure infrastructure is created and maintained using terraform

Work with various stakeholders to design and implement infrastructure changes to support new feature sets in various product lines.

Maintain transparency and clear visibility of costs associated with various product verticals, environments and work with stakeholders to plan for optimization and implementation

Spearhead continuous experimenting and innovating initiatives to optimize the infrastructure in terms of uptime, availability, latency and costs

You should apply, if you

1. Are a seasoned Veteran: Have managed infrastructure at scale running web apps, microservices, and data pipelines using tools and languages like JavaScript(NodeJS), Go, Python, Java, Erlang, Elixir, C++ or Ruby (experience in any one of them is enough)

2. Are a Mr. Perfectionist: You have a strong bias for automation and taking the time to think about the right way to solve a problem versus quick fixes or band-aids.

3. Bring your A-Game: Have hands-on experience and ability to design/implement infrastructure with GCP services like Compute, Database, Storage, Load Balancers, API Gateway, Service Mesh, Firewalls, Message Brokers, Monitoring, Logging and experience in setting up backups, patching and DR planning

4. Are up with the times: Have expertise in one or more cloud platforms (Amazon WebServices or Google Cloud Platform or Microsoft Azure), and have experience in creating and managing infrastructure completely through Terraform kind of tool

5. Have it all on your fingertips: Have experience building CI/CD pipeline using Jenkins, Docker for applications majorly running on Kubernetes. Hands-on experience in managing and troubleshooting applications running on K8s

6. Have nailed the data storage game: Good knowledge of Relational and NoSQL databases (MySQL,Mongo, BigQuery, Cassandra…)

7. Bring that extra zing: Have the ability to program/script is and strong fundamentals in Linux and Networking.

8. Know your toys: Have a good understanding of Microservices architecture, Big Data technologies and experience with highly available distributed systems, scaling data store technologies, and creating multi-tenant and self hosted environments, that’s a plus

Being Part of the Clan

At Classplus, you’re not an “employee” but a part of our “Clan”. So, you can forget about being bound by the clock as long as you’re crushing it workwise😎. Add to that some passionate people working with and around you, and what you get is the perfect work vibe you’ve been looking for!

It doesn’t matter how long your journey has been or your position in the hierarchy (we don’t do Sirs and Ma’ams); you’ll be heard, appreciated, and rewarded. One can say, we have a special place in our hearts for the Doers! ✊🏼❤️

Are you a go-getter with the chops to nail what you do? Then this is the place for you.
Read more
wwwthehiveai
Ashish Kapoor
Posted by Ashish Kapoor
Gurugram
5 - 20 yrs
₹10L - ₹40L / yr
DevOps
skill iconKubernetes
skill iconDocker
RabbitMQ
skill iconPostgreSQL
DevOps Engineer, Gurgaon

About Hive
Hive is the leading provider of cloud-based AI solutions for content understanding,
trusted by the world’s largest, fastest growing, and most innovative organizations. The
company empowers developers with a portfolio of best-in-class, pre-trained AI models, serving billions of customer API requests every month. Hive also offers turnkey software applications powered by proprietary AI models and datasets, enabling breakthrough use cases across industries. Together, Hive’s solutions are transforming content moderation, brand protection, sponsorship measurement, context-based ad targeting, and more.
Hive has raised over $120M in capital from leading investors, including General Catalyst, 8VC, Glynn Capital, Bain & Company, Visa Ventures, and others. We have over 250 employees globally in our San Francisco, Seattle, and Delhi offices. Please reach out if you are interested in joining the future of AI!

About Role
Our unique machine learning needs led us to open our own data centers, with an
emphasis on distributed high performance computing integrating GPUs. Even with these data centers, we maintain a hybrid infrastructure with public clouds when the right fit. As we continue to commercialize our machine learning models, we also need to grow our DevOps and Site Reliability team to maintain the reliability of our enterprise SaaS offering for our customers. Our ideal candidate is someone who is
able to thrive in an unstructured environment and takes automation seriously. You believe there is no task that can’t be automated and no server scale too large. You take pride in optimizing performance at scale in every part of the stack and never manually performing the same task twice.

Responsibilities
● Create tools and processes for deploying and managing hardware for Private Cloud Infrastructure.
● Improve workflows of developer, data, and machine learning teams
● Manage integration and deployment tooling
● Create and maintain monitoring and alerting tools and dashboards for various services, and audit infrastructure
● Manage a diverse array of technology platforms, following best practices and
procedures
● Participate in on-call rotation and root cause analysis
Requirements
● Minimum 5 - 10 years of previous experience working directly with Software
Engineering teams as a developer, DevOps Engineer, or Site Reliability
Engineer.
● Experience with infrastructure as a service, distributed systems, and software design at a high-level.
● Comfortable working on Linux infrastructures (Debian) via the CLIAble to learn quickly in a fast-paced environment.
● Able to debug, optimize, and automate routine tasks
● Able to multitask, prioritize, and manage time efficiently independently
● Can communicate effectively across teams and management levels
● Degree in computer science, or similar, is an added plus!
Technology Stack
● Operating Systems - Linux/Debian Family/Ubuntu
● Configuration Management - Chef
● Containerization - Docker
● Container Orchestrators - Mesosphere/Kubernetes
● Scripting Languages - Python/Ruby/Node/Bash
● CI/CD Tools - Jenkins
● Network hardware - Arista/Cisco/Fortinet
● Hardware - HP/SuperMicro
● Storage - Ceph, S3
● Database - Scylla, Postgres, Pivotal GreenPlum
● Message Brokers: RabbitMQ
● Logging/Search - ELK Stack
● AWS: VPC/EC2/IAM/S3
● Networking: TCP / IP, ICMP, SSH, DNS, HTTP, SSL / TLS, Storage systems,
RAID, distributed file systems, NFS / iSCSI / CIFS
Who we are
We are a group of ambitious individuals who are passionate about creating a revolutionary AI company. At Hive, you will have a steep learning curve and an opportunity to contribute to one of the fastest growing AI start-ups in San Francisco. The work you do here will have a noticeable and direct impact on the
development of the company.
Thank you for your interest in Hive and we hope to meet you soon
Read more
MyNextDeveloper
Neha Gandhi
Posted by Neha Gandhi
Remote only
2 - 5 yrs
₹15L - ₹30L / yr
skill iconDocker
skill iconKubernetes
DevOps
skill iconAmazon Web Services (AWS)
Windows Azure
+4 more

Company Introduction :


My Next Developer is a global network of top talent in business, design, and technology that enables companies to scale their teams, on-demand.


We take the best elements of virtual teams and combine them with a support structure that encourages innovation, social interaction, and fun. We see no borders, move at a fast pace, and are never afraid to break the mold.


Job Responsibilities


- Ensure security in the development activities.


- Implement risk management techniques and threat modeling.


- Implement infrastructure automation, monitoring and alerts as part of ISO 27001 and SOC 2 certifications.


- Collaborate with internal teams to produce the best security solutions


Minimum requirements :


- Bachelor's/Master's degree in degree in computer science, cybersecurity, engineering, or equivalent degree.


- 3+ years of experience as a DevSecOps engineer.


- Proficiency in back-end technologies such as NodeJs or Python.


- Expertise in using DevOps tools like GitHub, dependency management, and CI/CD.


- Profound knowledge of AWS cloud, DevOps culture and automation tools.


- Fluent in both spoken and written English communication.


- Ability to work full-time (40 hours/week) with a 2-3 hour overlap with European time zone.


- Stay up-to-date on cybersecurity threats and follow the best practices.


- Previous project experience related to ISO 27001 and SOC 2 certifications.

Read more
Jar

at Jar

1 video
2 recruiters
Nischal Hebbar
Posted by Nischal Hebbar
Remote, Bengaluru (Bangalore)
3 - 5 yrs
₹10L - ₹30L / yr
skill iconJenkins
skill iconDocker
skill iconKubernetes
DevOps
skill iconAmazon Web Services (AWS)
+2 more
  • 3 - 6 years of software development, and operations experience deploying and maintaining multi-tiered infrastructure and applications at scale.
  • Design cloud infrastructure that is secure, scalable, and highly available on AWS
  • Experience managing any distributed NoSQL system (Kafka/Cassandra/etc.)
  • Experience with Containers, Microservices, deployment and service orchestration using Kubernetes, EKS (preferred), AKS or GKE.
  • Strong scripting language knowledge, such as Python, Shell
  • Experience and a deep understanding of Kubernetes.
  • Experience in Continuous Integration and Delivery.
  • Work collaboratively with software engineers to define infrastructure and deployment requirements
  • Provision, configure and maintain AWS cloud infrastructure
  • Ensure configuration and compliance with configuration management tools
  • Administer and troubleshoot Linux-based systems
  • Troubleshoot problems across a wide array of services and functional areas
  • Build and maintain operational tools for deployment, monitoring, and analysis of AWS infrastructure and systems
  • Perform infrastructure cost analysis and optimization
 
Tech used:
  • AWS
  • Docker
  • Kubernetes
  • Envoy
  • Istio
  • Jenkins
  • Cloud Security & SIEM stacks
  • Terraform
Read more
Consulting and Product Engineering Company

Consulting and Product Engineering Company

Agency job
via Exploro Solutions by Sapna Prabhudesai
Hyderabad, Bengaluru (Bangalore), Chennai, Pune
8 - 12 yrs
₹6L - ₹30L / yr
DevOps
skill iconKubernetes
Architecture
Microservices

 

Job Dsecription: (8-12 years)

 

○ Develop best practices for team and also responsible for the architecture

○ solutions and documentation operations in order to meet the engineering departments quality and standards

○ Participate in production outage and handle complex issues and works towards Resolution

○ Develop custom tools and integration with existing tools to increase engineering Productivity

 

Required Experience and Expertise

 

○ Deep understanding of Kernel, Networking and OS fundamentals

○ Strong experience in writing helm charts.

○ Deep understanding of K8s.

○ Good knowledge in service mesh.

○ Good Database understanding

 

 

Location:
Bangalore / Chennai / Hyderabad / Pune

Notice Period: 30 day max
Read more
Directi

at Directi

13 recruiters
Richa Pancholy
Posted by Richa Pancholy
Bengaluru (Bangalore)
2 - 8 yrs
₹10L - ₹40L / yr
skill iconAmazon Web Services (AWS)
skill iconPython
Linux/Unix
DevOps
skill iconMongoDB
+8 more
What is the Job like?We are looking for a talented individual to join our DevOps and Platforms Engineering team. You will play an important role in helping build and run our globally distributed infrastructure stack and platforms. Technologies you can expect to work on every day include Linux, AWS, MySQL/PostgreSQL, MongoDB, Hadoop/HBase, ElasticSearch, FreeSwitch, Jenkins, Nagios, and CFEngine amongst others.Responsibilities:- * Troubleshoot and fix production outages and performance issues in our AWS/Linux infrastructure stack* Build automation tools for provisioning and managing our cloud infrastructure by leveraging the AWS API for EC2, S3, CloudFront, RDS and Route53 amongst others* Contribute to enhancing and managing our continuous delivery pipeline* Proactively seek out opportunities to improve monitoring and alerting of our hosts and services, and implement them in a timely fashion* Code scripts and tools to collect and visualize metrics from linux hosts and JVM applications* Enhance and maintain our logs collection, processing and visualization infrastructure* Automate systems configuration by writing policies and modules for configuration management tools* Write both frontend (html/css/js) and backend code (Python, Ruby, Perl)* Participate in periodic oncall rotations for devopsSkills:- * DevOps/System Admin experience ranging between 3-4 years* In depth Linux/Unix knowledge, good understanding the various linux kernel subsystems (memory, storage, network etc)* DNS, TCP/IP, Routing, HA & Load Balancing* Configuration management using tools like CFEngine, Puppet or Chef* SQL and NoSQL databases like MySQL, PostgreSQL, MongoDB and HBase* Build and packaging tools like Jenkins and RPM/Yum* HA and Load balancing using tools like the Elastic Load Balancer and HAProxy* Monitoring tools like Nagios, Pingdom or similar* Log management tools like logstash, fluentd, syslog, elasticsearch or similar* Metrics collection tools like Ganglia, Graphite, OpenTSDB or similar* Programming in a high level language like Python or Ruby
Read more
smartShift Technologies

at smartShift Technologies

2 recruiters
Ramesh Kumar
Posted by Ramesh Kumar
Bengaluru (Bangalore), Bengaluru (Bangalore)
7 - 11 yrs
₹8L - ₹16L / yr
Apache
Tomcat
skill iconAmazon Web Services (AWS)
JBoss
Linux/Unix
+1 more
Modern Systems empowers IT efficiency, agility and competitive advantage through core application modernization. Our solutions facilitate the transition from legacy systems to modern environments that leverage cloud, in-memory, data warehousing, business intelligence and agile development methodologies.
Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort