
Senior Engineer – Cloud & AI" to drive the development through deployment of cutting-edge cloud-native AI solutions. This role requires a combination of deep technical expertise in Microsoft Azure, strong DevOps and CI/CD practices, and an engineering mindset focused on delivering impactful and value-centric solutions. The ideal candidate is a collaborative team player who thrives in both research-oriented problem solving and timeline-driven execution. The individual comes with a proven background on deploying Generative AI workloads (LLM’s, LVM’s, TTS, Speech to Text etc.) cloud native, with “Security by design” using Commercial (for ex: Azure OpenAI), or Open-source (for ex: Llama, Mistral etc.) models. Key Responsibilities: • Drive the design, development, and deployment of cloud-native Generative AI solutions on Microsoft Azure. • Implement and maintain DevOps best practices, including CI/CD pipelines, Infrastructure as Code (IaC), and version control using GitHub or Azure DevOps. • Collaborate with data scientists, AI/ML teams, and business stakeholders to deploy and optimize AI models within cloud-native architectures, ensuring seamless integration and performance at scale. • Drive cloud architecture decisions in partnership with Enterprise Architecture, TechOps, and Cybersecurity functions, ensuring scalability, security, and cost efficiency • Build and manage CI/CD pipelines to streamline the deployment process and ensure rapid delivery of high-quality applications. • Advocate and implement Infrastructure as Code (IaC) to automate infrastructure provisioning, ensuring consistency across environments. • Collaborate closely with cross-functional teams, promoting a culture of teamwork and innovation. • Provide technical leadership, mentoring team members, and fostering a continuous learning environment. • Stay up to date with the latest trends in Azure services, AI technologies, and DevOps practices to recommend innovative solutions.
REQUIRED SKILLS AND EXPERIENCE
Bachelor's degree (Masters Degree preferred) in Computer Science, Information Technology, or related fields 5+ years of experience in Cloud & AI 2+ years on Generative AI, Experience in integrating AI/ML models into cloud solutions Deep understanding of Azure services, including but not limited to Azure Virtual Machines, Azure Container Apps Environment, Azure Machine Learning, Azure Functions, and Azure Cognitive Services Experience working with Generative AI Frameworks such as LangChain, LlamaIndex, etc Hands-on experience with IaC tools such as Terraform, ARM templates, or Bicep

Similar jobs
Job Description
Experience: 5 - 9 years
Location: Bangalore/Pune/Hyderabad
Work Mode: Hybrid(3 Days WFO)
Senior Cloud Infrastructure Engineer for Data Platform
The ideal candidate will play a critical role in designing, implementing, and maintaining cloud infrastructure and CI/CD pipelines to support scalable, secure, and efficient data and analytics solutions. This role requires a strong understanding of cloud-native technologies, DevOps best practices, and hands-on experience with Azure and Databricks.
Key Responsibilities:
Cloud Infrastructure Design & Management
Architect, deploy, and manage scalable and secure cloud infrastructure on Microsoft Azure.
Implement best practices for Azure Resource Management, including resource groups, virtual networks, and storage accounts.
Optimize cloud costs and ensure high availability and disaster recovery for critical systems
Databricks Platform Management
Set up, configure, and maintain Databricks workspaces for data engineering, machine learning, and analytics workloads.
Automate cluster management, job scheduling, and monitoring within Databricks.
Collaborate with data teams to optimize Databricks performance and ensure seamless integration with Azure services.
CI/CD Pipeline Development
Design and implement CI/CD pipelines for deploying infrastructure, applications, and data workflows using tools like Azure DevOps, GitHub Actions, or similar.
Automate testing, deployment, and monitoring processes to ensure rapid and reliable delivery of updates.
Monitoring & Incident Management
Implement monitoring and alerting solutions using tools like Dynatrace, Azure Monitor, Log Analytics, and Databricks metrics.
Troubleshoot and resolve infrastructure and application issues, ensuring minimal downtime.
Security & Compliance
Enforce security best practices, including identity and access management (IAM), encryption, and network security.
Ensure compliance with organizational and regulatory standards for data protection and cloud operations.
Collaboration & Documentation
Work closely with cross-functional teams, including data engineers, software developers, and business stakeholders, to align infrastructure with business needs.
Maintain comprehensive documentation for infrastructure, processes, and configurations.
Required Qualifications
Education: Bachelor’s degree in Computer Science, Engineering, or a related field.
Must Have Experience:
6+ years of experience in DevOps or Cloud Engineering roles.
Proven expertise in Microsoft Azure services, including Azure Data Lake, Azure Databricks, Azure Data Factory (ADF), Azure Functions, Azure Kubernetes Service (AKS), and Azure Active Directory.
Hands-on experience with Databricks for data engineering and analytics.
Technical Skills:
Proficiency in Infrastructure as Code (IaC) tools like Terraform, ARM templates, or Bicep.
Strong scripting skills in Python, or Bash.
Experience with containerization and orchestration tools like Docker and Kubernetes.
Familiarity with version control systems (e.g., Git) and CI/CD tools (e.g., Azure DevOps, GitHub Actions).
Soft Skills:
Strong problem-solving and analytical skills.
Excellent communication and collaboration abilities.
Profile: Devops Lead
Location: Gurugram
Experience: 08+ Years
Notice Period: can join Immediate to 1 week
Company: Watsoo
Required Skills & Qualifications
- Bachelor’s degree in Computer Science, Engineering, or related field.
- 5+ years of proven hands-on DevOps experience.
- Strong experience with CI/CD tools (Jenkins, GitLab CI, GitHub Actions, etc.).
- Expertise in containerization & orchestration (Docker, Kubernetes, Helm).
- Hands-on experience with cloud platforms (AWS, Azure, or GCP).
- Proficiency in Infrastructure as Code (IaC) tools (Terraform, Ansible, Pulumi, or CloudFormation).
- Experience with monitoring and logging solutions (Prometheus, Grafana, ELK, CloudWatch, etc.).
- Proficiency in scripting languages (Python, Bash, or Shell).
- Knowledge of networking, security, and system administration.
- Strong problem-solving skills and ability to work in fast-paced environments.
- Troubleshoot production issues, perform root cause analysis, and implement preventive measures.
Advocate DevOps best practices, automation, and continuous improvement
What You will do
We're looking for bright Senior DevOps engineers that are deeply experienced individual contributors and dynamic team leaders, to join our dedicated DevOps Practice Group. As a Senior DevOps practitioner at Zemoso you’ll use your expertise to provide DevSecOps services to our customers and to our inhouse Dev teams.
Basic Qualifications
- Bachelor's degree in computer science, engineering or equivalent experience.
- 15+ years of overall IT experience
- 10+ years of DevOps experience with 5+ years of it on Azure platform
- Relevant Azure Certification(s)
Must have experience
- IaC tools ARM/bicep template, Azure DevOps
- GITHub repo and GIT tools
- scripting languages Python, PowerShell, bash development experience
- CI/CD design, build management.
- Container tools Docker and Kubernetes
- Azure Policy and permissions provisioning.
- IT experience to design, development (Java Spring Boot), automated testing, REST API’s, Microservices, Pub/Sub architecture
- Costing, budgeting, forecasting, monitoring cloud expenses.
- Architecture drawings and proposals, presentations.
- Stakeholder management.
- Prototyping.
Nice to have experience
- IaC tools CloudFormation, Terraform
- Build tools Apache Maven, Jenkins
- Networking: DNS, TCP/IP, HTTP, CDN and VPN
- Security: NSG, firewall and Azure policies, HTTPS, SSL certificates, KMS key
- Authentication (IAM): SAML 2.0, OpenID Connect, OAuth 2.0, Keycloak-based IdAM, LDAP and AD, and/or 3rd-party SSO.
- Access: RBAC and ABAC
- Load Balancers: Apache, Nginx, HAProxy
- Encryption: PKI infrastructure, Certificate authorities, OCSP, CRL, CSRs, x.509 certificate structures, pkcs12 certificate containers
Additionally, we’re looking for individuals who:
- Will assist in the rollout and deployment of new product features, and deploy applications via automation with configuration management tools
- Will build applications and tools to reduce barriers, decrease friction and speed up delivery of products
- Can effectively articulate technical challenges and solutions.
- Have strong collaborative instincts and interpersonal skills; must enjoy and work well with a wide variety of stakeholders in a complex organization
- Will manage individual project priorities with guidance, meeting deadlines and deliverables.
Benefits
- Competitive salary.
- Work from anywhere.
- Learning and gaining experience rapidly.
- Reimbursement for basic working set up at home.
- Insurance (including a top up insurance for COVID).
Location
Remote - work from anywhere.
About us:
Zemoso Technologies is a Software Product Market Fit Studio that brings silicon valley style rapid prototyping and rapid application builds to Entrepreneurs and Corporate innovation. We offer Innovation as a service and work on ideas from scratch and take it to the Product Market Fit stage using Design Thinking->Lean Execution->Agile Methodology.
We were featured as one of Deloitte Fastest 50 growing tech companies from India thrice (2016, 2018 and 2019). We were also featured in Deloitte Technology Fast 500 Asia Pacific both in 2016 and 2018.
We are located in Hyderabad, India, and Dallas, US. We have recently incorporated another office in Waterloo, Canada.
Our founders have had past successes - founded a decision management company acquired by SAP AG (now part of Hana Big data stack & NetWeaver BPM), early engineering team of Zoho (leading billion $ SaaS player) & some Private Equity experience.
Marquee customers along with some exciting start-ups are part of our clientele.
environments: AWS / Azure / GCP
• Must have strong work experience (2 + years) developing IaC (i.e. Terraform)
• Must have strong work experience in Ansible development and deployment.
• Bachelor’s degree with a background in math will be a PLUS.
• Must have 8+ years experience with a mix of Linux and Window systems in a medium to large business
environment.
• Must have command level fluency and shell scripting experience in a mix of Linux and Windows
environments.
•
• Must enjoy the experience of working in small, fast-paced teams
• Identify opportunities for improvement in existing process and automate the process using Ansible Flows.
• Fine tune performance and operation issues that arise with Automation flows.
• Experience administering container management systems like Kubernetes would be plus.
• Certification with Red Hat or any other Linux variant will be a BIG PLUS.
• Fluent in the use of Microsoft Office Applications (Outlook / Word / Excel).
• Possess a strong aptitude towards automating and timely completion of standard/routine tasks.
• Experience with automation and configuration control systems like Puppet or Chef is a plus.
• Experience with Docker, Kubernetes (or container orchestration equivalent) is nice to have
Devops Engineer Position - 3+ years
Kubernetes, Helm - 3+ years (dev & administration)
Monitoring platform setup experience - Prometheus, Grafana
Azure/ AWS/ GCP Cloud experience - 1+ years.
Ansible/Terraform/Puppet - 1+ years
CI/CD - 3+ years
- 7-10 years experience with secure SDLC/DevSecOps practices such as automating security processes within CI/CD pipeline.
- At least 4 yrs. experience designing, and securing Data Lake & Web applications deployed to AWS, Azure, Scripting/Automation skills on Python, Shell, YAML, JSON
- At least 4 years of hands-on experience with software development lifecycle, Agile project management (e.g. Jira, Confluence), source code management (e.g. Git), build automation (e.g. Jenkins), code linting and code quality (e.g. SonarQube), test automation (e.g. Selenium)
- Hand-on & Solid understanding of Amazon Web Services & Azure-based Infra & applications
- Experience writing cloud formation templates, Jenkins, Kubernetes, Docker, and microservice application architecture and deployment.
- Strong know-how on VA/PT integration in CI/CD pipeline.
- Experience in handling financial solutions & customer-facing applications
Roles
- Accelerate enterprise cloud adoption while enabling rapid and stable delivery of capabilities using continuous integration and continuous deployment principles, methodologies, and technologies
- Manage & deliver diverse cloud [AWS, Azure, GCP] DevSecOps journeys
- Identify, prototype, engineer, and deploy emerging software engineering methodologies and tools
- Maximize automation and enhance DevSecOps pipelines and other tasks
- Define and promote enterprise software engineering and DevSecOps standards, practices, and behaviors
- Operate and support a suite of enterprise DevSecOps services
- Implement security automation to decrease the loop between the development and deployment processes.
- Support project teams to adopt & integrate the DevSecOps environment
- Managing application vulnerabilities, Data security, encryption, tokenization, access management, Secure SDLC, SAST/DAST
- Coordinate with development and operations teams for practical automation solutions and custom flows.
- Own DevSecOps initiatives by providing objective, practical and relevant ideas, insights, and advice.
- Act as Release gatekeeper with an understanding of OWASP top 10 lists of vulnerabilities, NIST SP-800-xx, NVD, CVSS scoring, etc concepts
- Build workflows to ensure a successful DevSecOps journey for various enterprise applications.
- Understand the strategic direction to reach business goals across multiple projects & teams
- Collaborate with development teams to understand project deliverables and promote DevSecOps culture
- Formulate & deploy cloud automation strategies and tools
Skills
- Knowledge of the DevSecOps culture and principles.
- An understanding of cloud technologies & components
- A flair for programming languages such as Shell, Python, Java Scripts,
- Strong teamwork and communication skills.
- Knowledge of threat modeling and risk assessment techniques.
- Up-to-date knowledge of cybersecurity threats, current best practices, and the latest software.
- An understanding of programs such as Puppet, Chef, ThreatModeler, Checkmarx, Immunio, and Aqua.
- Strong know-how of Kubernetes, Docker, AWS, Azure-based deployments
- On the job learning for new programming languages, automation tools, deployment architectures
MTX Group Inc. is seeking a motivated Lead DevOps Engineer to join our team. MTX Group Inc. is a global implementation partner enabling organizations to become fit enterprises. MTX provides expertise across various platforms and technologies, including Google Cloud, Salesforce, artificial intelligence/machine learning, data integration, data governance, data quality, analytics, visualization and mobile technology. MTX’s very own Artificial Intelligence platform Maverick, enables clients to accelerate processes and critical decisions by leveraging a Cognitive Decision Engine, a collection of purpose-built Artificial Neural Networks designed to leverage the power of Machine Learning. The Maverick Platform includes Smart Asset Detection and Monitoring, Chatbot Services, Document Verification, to name a few.
Responsibilities:
- Be responsible for software releases, configuration, monitoring and support of production system components and infrastructure.
- Troubleshoot technical or functional issues in a complex environment to provide timely resolution, with various applications and platforms that are global.
- Bring experience on Google Cloud Platform.
- Write scripts and automation tools in languages such as Bash/Python/Ruby/Golang.
- Configure and manage data sources like PostgreSQL, MySQL, Mongo, Elasticsearch, Redis, Cassandra, Hadoop, etc
- Build automation and tooling around Google Cloud Platform using technologies such as Anthos, Kubernetes, Terraform, Google Deployment Manager, Helm, Cloud Build etc.
- Bring a passion to stay on top of DevOps trends, experiment with and learn new CI/CD technologies.
- Work with users to understand and gather their needs in our catalogue. Then participate in the required developments
- Manage several streams of work concurrently
- Understand how various systems work
- Understand how IT operations are managed
What you will bring:
- 5 years of work experience as a DevOps Engineer.
- Must possess ample knowledge and experience in system automation, deployment, and implementation.
- Must possess experience in using Linux, Jenkins, and ample experience in configuring and automating the monitoring tools.
- Experience in the software development process and tools and languages like SaaS, Python, Java, MongoDB, Shell scripting, Python, MySQL, and Git.
- Knowledge in handling distributed data systems. Examples: Elasticsearch, Cassandra, Hadoop, and others.
What we offer:
- Group Medical Insurance (Family Floater Plan - Self + Spouse + 2 Dependent Children)
- Sum Insured: INR 5,00,000/-
- Maternity cover upto two children
- Inclusive of COVID-19 Coverage
- Cashless & Reimbursement facility
- Access to free online doctor consultation
- Personal Accident Policy (Disability Insurance) -
- Sum Insured: INR. 25,00,000/- Per Employee
- Accidental Death and Permanent Total Disability is covered up to 100% of Sum Insured
- Permanent Partial Disability is covered as per the scale of benefits decided by the Insurer
- Temporary Total Disability is covered
- An option of Paytm Food Wallet (up to Rs. 2500) as a tax saver benefit
- Monthly Internet Reimbursement of upto Rs. 1,000
- Opportunity to pursue Executive Programs/ courses at top universities globally
- Professional Development opportunities through various MTX sponsored certifications on multiple technology stacks including Salesforce, Google Cloud, Amazon & others
*******************
- Experience working on Linux based infrastructure
- Strong hands-on knowledge of setting up production, staging, and dev environments on AWS/GCP/Azure
- Strong hands-on knowledge of technologies like Terraform, Docker, Kubernetes
- Strong understanding of continuous testing environments such as Travis-CI, CircleCI, Jenkins, etc.
- Configuration and managing databases such as MySQL, Mongo
- Excellent troubleshooting
- Working knowledge of various tools, open-source technologies, and cloud services
- Awareness of critical concepts in DevOps and Agile principles
- Azure Devops - Working experience in Azure yaml pipelines. (Note – Some say they worked in yaml but it’s for Jenkins and not Azure devops)
- Azure – Infrastructure automation using Terraform/ARM templates. (Note – Some say that they worked in terraform but it’s for AWS and not Azure). Please confirm Terraform for Azure infrastructure automation.
- Powershell scripting to automate and deploy .Net applications.
A strong background in Azure OR Amazon Web Services (AWS) or a similar cloud platform is a must-have, certification is a plus.
Excellent technical skills and knowledge include but not limited to: cloud methodologies like PaaS and SaaS; programming languages such as Python, Java, .Net; orchestration systems such as Chef, Ansible, Terraform; Azure IaaS servers; PowerShell scripting.
Fully aware of the DevOps cycle with hands-on on deployment models to the cloud.
Experience working with Docker and related containerization technologies.
Experience working on orchestration platforms such as AKS, RedHat OpenShift etc.
Extensive knowledge working with logging and monitoring tools such as EFK, visualization tools such as Grafana and Prometheus.
Exposure to security alert monitoring tools.
Experience in building both microservices and public facing API's.- Experience in working with any of the API gateways.






