50+ Remote Windows Azure Jobs in India
Apply to 50+ Remote Windows Azure Jobs on CutShort.io. Find your next job, effortlessly. Browse Windows Azure Jobs and apply today!
Key Responsibilities
AI Architecture & Solution Design
- Design end-to-end AI solution architectures, including:
- Generative AI and LLM-based systems
- Retrieval-Augmented Generation (RAG) pipelines
- Agentic and multi-agent workflows
- Define reference architectures and best practices for AI-enabled features within enterprise products.
- Ensure AI solutions integrate seamlessly with existing applications, data, and cloud architectures.
AI Integration & MCP Servers
- Design and implement Model Context Protocol (MCP) servers to securely expose tools, APIs, and data to AI agents.
- Define standards for tool interfaces, access control, auditing, and safety guardrails.
- Enable product teams to onboard AI tools and capabilities using reusable, scalable integration patterns.
Agentic AI & Workflow Enablement
- Architect AI-driven workflows that support collaboration between humans and AI agents.
- Design AI-to-AI (A2A) and AI-to-system interaction patterns.
- Ensure agent behaviors are deterministic, explainable, and aligned with enterprise requirements.
Hands-On Development & Prototyping
- Build proofs-of-concept and production-ready implementations using Python and/or TypeScript.
- Rapidly validate ideas from ideation to deployment.
- Establish reusable frameworks, libraries, and CI/CD pipelines for AI development.
AI Governance, Quality & Safety
- Implement guardrails to minimize hallucinations, unsafe actions, and data leakage.
- Define evaluation and monitoring strategies for AI systems, including prompt regression and RAG accuracy checks.
- Ensure AI solutions comply with enterprise security, privacy, and governance standards.
Developer Enablement & Collaboration
- Partner with Product, Engineering, QE, Performance, and Security teams to deliver AI capabilities.
- Mentor teams on AI design patterns, tooling, and best practices.
- Contribute to internal AI communities through demos, documentation, and knowledge sharing.
Qualifications :
Required Qualifications
- Bachelor’s degree in computer science, Engineering, or a related technical field, or equivalent practical experience.
- Demonstrated expertise in cloud‑native system design, distributed architectures, and enterprise‑scale integrations.
- Proven ability to architect and implement AI-enabled systems, including integrating Large Language Models (LLMs) into production-grade software.
- Strong ownership of architectural decisions, technical direction, and solution delivery across complex, cross-functional initiatives.
- Hands-on experience applying security, observability, and automation best practices within enterprise environments.
- 6–10 years of experience in software architecture and distributed systems.
- 5+ years of experience building Generative AI or LLM-based solutions.
- Practical experience designing and implementing:
- Retrieval-Augmented Generation (RAG) architectures
- Agentic AI systems
- Tool-calling frameworks and AI integration layers
- Proficiency in Python and/or .Net/TypeScript/Node.js.
- Experience working with major cloud platforms such as Azure, AWS, or Google Cloud Platform (GCP).
Preferred Qualifications
- Experience with OpenAI, Azure OpenAI, Anthropic, or similar LLM platforms.
- Familiarity with Model Context Protocol (MCP) or equivalent AI tool-integration frameworks.
- Experience applying AI engineering practices beyond prototyping, including evaluation, reliability, and scalability considerations.
- Ability to translate ambiguous business problems into clear technical architecture and execution plans.
- History of influencing technical standards and mentoring senior engineers or architects.
- Experience with vector databases, embeddings, and retrieval optimisation.
- Experience building AI-enabled developer tooling and CI/CD pipelines.
- Prior experience in enterprise SaaS environments.
Description
As a Power Apps Developer, you will be at the forefront of crafting innovative, low‑code solutions that streamline business processes and empower end‑users across the organization. You will collaborate closely with functional analysts, business stakeholders, and fellow developers to translate complex requirements into intuitive, scalable applications on the Microsoft Power Platform. The role offers a dynamic environment where continuous learning is encouraged, providing access to the latest Power Apps features, Azure services, and integration techniques. You will contribute to a culture of knowledge sharing, participate in code reviews, and mentor junior team members, ensuring high‑quality deliverables that drive operational efficiency and measurable business impact.
Requirements:
- 5–15 years of experience developing enterprise‑grade solutions using Microsoft Power Apps, Power Automate, and Power BI.
- Strong proficiency in Canvas and Model‑Driven apps, Common Data Service (Dataverse), and integration with Azure services (e.g., Azure Functions, Logic Apps).
- Solid understanding of relational databases, SQL, and data modeling concepts.
- Experience with JavaScript, TypeScript, and RESTful APIs for extending Power Apps functionality.
- Excellent problem‑solving abilities, strong communication skills, and a collaborative mindset.
- Relevant certifications such as Microsoft Power Platform Developer Associate (PL‑400) are a plus.
Roles and Responsibilities:
- Design, develop, and deploy custom Power Apps solutions that meet business requirements and adhere to best practices.
- Create and maintain automated workflows using Power Automate to streamline repetitive tasks and improve efficiency.
- Integrate Power Apps with external systems via connectors, APIs, and Azure services to ensure seamless data flow.
- Perform performance tuning, debugging, and troubleshooting of applications to ensure optimal user experience.
- Collaborate with business analysts and stakeholders to gather requirements, provide technical guidance, and deliver prototypes.
- Conduct code reviews, enforce governance standards, and contribute to the development of a reusable component library.
- Stay updated with the latest Power Platform releases, evaluate new features, and recommend adoption strategies.
- Provide training and mentorship to junior developers and end‑users to foster platform adoption.
Must have skills
Power apps - 5 years
Microsoft Power Automate - 1 years
Nice to have skills
Canvas App Development and Scripting - 4 years
Canvas Apps Development - 4 years
SQL - 2 years
SharePoint APIs - 1 years
Power Fx - 2 years
C Sharp - 3 years
RESTful API - 2 years
🚀 Hiring: AI/M and Gen AI Engineer
⭐ Experience: 5+ Years
⭐ Work Mode:- Remote
⏱️ Notice Period: Immediate Joiners
(Only immediate joiners & candidates serving notice period)
🌟 About the Role
We are looking for a highly skilled AI/ML Software Engineer to design, build, and productionize enterprise-grade AI solutions. This role focuses on Generative AI, RAG systems, and AI agent–driven automation, with deployment on Microsoft Azure.
You will collaborate with cross-functional teams including architects, engineers, and business stakeholders to deliver scalable and secure AI solutions that create real business impact.
🔑 Mandatory Skills (Must Have)
- ✅ Azure AI Ecosystem (Azure Machine Learning, Azure OpenAI, Cognitive Services)
- ✅ Generative AI & RAG Systems (vector embeddings, retrieval pipelines)
- ✅ Strong Software Engineering + MLOps (CI/CD, containerization, scalable deployments)
💼 Key Responsibilities
- Design, develop, and deploy AI/ML models in production environments
- Build and optimize RAG-based applications and AI agent workflows
- Develop scalable data pipelines and integrate with enterprise systems
- Implement MLOps practices for continuous deployment and monitoring
- Work with big data tools to process large-scale datasets
- Ensure security, scalability, and performance of AI systems
- Collaborate with stakeholders to translate business problems into AI solutions
🧠 Required Experience & Skills
- 5–8 years of hands-on experience in AI/ML development
- Strong programming and software engineering expertise
- Experience with Azure services (ML, Data Lake, OpenAI, Cognitive Services)
- Knowledge of vector databases and embedding models
- Experience with Databricks, Azure Data Factory, or Kafka
- Familiarity with multi-agent systems / agentic AI frameworks
- Proficiency in TensorFlow, PyTorch, Keras, or Scikit-learn
- Background in NLP, Computer Vision, or Deep Learning
- Experience with SQL/NoSQL databases and ETL pipelines
- Strong analytical and problem-solving skills
Job Overview
As a software Engineer, you will play a crucial role in leading our development efforts, ensuring best practices, and supporting the team on a day-to-day basis. This role requires deep technical knowledge, a proactive mindset, and a commitment to guiding the team in tackling challenging issues. You will work primarily with .NET Core on the backend while also keeping a strategic focus on product security, DevOps, quality assurance, and cloud infrastructure.
Responsibilities
• Forward-Looking Product Development:
o Collaborate with product and engineering teams to align on the technical
direction, scalability, and maintainability of the product.
o Proactively consider and address security, performance, and scalability
requirements during development.
- Cloud and Infrastructure: Leverage Microsoft Azure for cloud infrastructure,
- ensuring efficient and secure use of cloud services. Work closely with DevOps to
- improve deployment processes.
- DevOps & CI/CD: Support the setup and maintenance of CI/CD pipelines, enabling
- smooth and frequent deployments. Collaborate with the DevOps team to automate and
- optimize the development process.
- Technical Mentorship: Provide technical guidance and support to team members,
- helping them solve day-to-day challenges, enhance code quality, and adopt best
- practices.
- Quality Assurance: Collaborate with QA to ensure thorough testing, automated testing
- coverage, and overall product quality.
- Product Security: Actively implement and promote security best practices to protect
- data and ensure compliance with industry standards.
- Documentation & Code Reviews: Promote good coding practices, conduct code
- reviews, and maintain clear documentation.
- Qualifications
• Technical Skills:
o Strong experience with .NET Core for backend development and RESTful API
design.
o Hands-on experience with Microsoft Azure services, including but not limited
to VMs, databases, application gateways, and user management.
o Familiarity with DevOps practices and tools, particularly CI/CD pipeline
configuration and deployment automation.
o Strong knowledge of product security best practices and experience implementing secure coding practices.
o Familiarity with QA processes and automated testing tools is a plus.
o Ability to support team members in solving technical challenges and sharing
knowledge effectively.
Preferred Qualifications
- 4+ years of experience in software development, with a strong focus on .NET Core
- Previous experience as a Staff SE, tech lead, or in a similar hands-on tech role.
- Strong problem-solving skills and ability to work in a fast-paced, startup environment.
- What We Offer
- Opportunity to lead and grow within a dynamic and ambitious team.
- Challenging projects that focus on innovation and cutting-edge technology.
- Collaborative work environment with a focus on learning, mentorship, and growth.
- Competitive compensation, benefits, and stock options.
- If you’re a proactive, forward-thinking technology leader with a passion for .NET Core and you’re ready to make an impact, we’d love to meet you!
Description
Join company as a Backend Developer and become a pivotal force in building the robust, scalable services that power our innovative platforms. In this role, you will design, develop, and maintain server‑side applications, ensuring high performance and reliability for millions of users. You’ll collaborate closely with cross‑functional product, front‑end, and DevOps teams to translate business requirements into clean, efficient code, while participating in code reviews and architectural discussions. Our dynamic environment encourages continuous learning, offering opportunities to work with cutting‑edge technologies, cloud infrastructures, and modern development practices. As a key contributor, your work will directly impact product quality, user satisfaction, and the overall success of company's mission to streamline hiring solutions.
Requirements:
- 1–15 years of professional experience in backend development, with a strong focus on building APIs and microservices.
- Proficiency in server‑side languages such as Python, Java, Node.js, or Go, and solid understanding of object‑oriented and functional programming paradigms.
- Extensive experience with relational (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., MongoDB, Redis), including schema design and query optimization.
- Familiarity with cloud platforms (AWS, GCP, Azure) and containerization technologies like Docker and Kubernetes.
- Hands‑on experience with version control (Git), CI/CD pipelines, and automated testing frameworks.
- Strong problem‑solving abilities, effective communication skills, and a collaborative mindset for working within multidisciplinary teams.
Roles and Responsibilities:
- Design, develop, and maintain high‑throughput backend services and RESTful APIs that support core product features.
- Implement data models and storage solutions, ensuring data integrity, security, and optimal performance.
- Collaborate with front‑end engineers, product managers, and designers to define technical requirements and deliver end‑to‑end solutions.
- Participate in code reviews, provide constructive feedback, and uphold coding standards and best practices.
- Monitor, troubleshoot, and optimize production systems, implementing robust logging, alerting, and performance tuning.
- Contribute to the continuous improvement of development workflows, including CI/CD automation, testing strategies, and deployment processes.
- Stay current with emerging technologies and industry trends, proposing innovative approaches to enhance system architecture.
Budget:
- Job Type: payroll
- Experience Range: 1–15 years
Hiring: IT Operations & Helpdesk Engineer (3–5 Years)
📍 Location: [Bangalore / Hybrid]
We are looking for a hands-on IT Operations Engineer who will anchor our internal IT helpdesk while managing servers, backups, DR drills, and cloud infrastructure. This role is responsible for day-to-day IT stability across endpoints, servers, and Azure environments.
Key Activities
Internal IT Helpdesk (Primary Anchor)
· Act as the single point of contact for internal IT support.
· Resolve L1/L2 issues (laptops, OS, network, access, software installs).
· Manage onboarding/offboarding IT setup.
· Track tickets, SLAs, and recurring issues.
Infrastructure & Servers
· Install and maintain Windows & Linux servers.
· Maintain the centralized IT asset inventory.
· Support manual and automated application deployments
· Handle patching, upgrades, performance monitoring.
Cloud Administration (Azure)
· Manage VMs, storage, networking.
· Maintain access controls and security configurations.
Backup & DR Readiness
· Manage and test backup processes.
· Conduct periodic DR drills to support organizational continuity standards.
· Maintain recovery runbooks and documentation.
What We’re Looking For
· Strong Windows Server & Linux hands-on experience.
· Experience managing Azure Cloud infrastructure.
· Practical backup & restore execution experience.
· Strong troubleshooting mindset.
· Process-driven and documentation disciplined.
· Comfortable working with DevOps & Cyber Security teams.
Impact of This Role
· Stable internal IT operations.
· DR-tested infrastructure.
· Reduced downtime and faster issue resolution.
· Strong operational hygiene in a growing environment.
Way2DreamJobs is building a premium cloud mentorship ecosystem focused on real-world Microsoft Azure and Modern Workplace skills.
We are inviting experienced Azure professionals to collaborate as founding weekend mentors for a remote mentorship program.
This is not a traditional full-time job. It is a flexible mentor collaboration model designed for working IT professionals who want to share real enterprise experience.
Responsibilities:
• Conduct weekend mentorship sessions
• Guide learners through practical Azure scenarios
• Support hands-on lab oriented learning
Ideal Profile:
• 4+ years Azure infrastructure experience
• Exposure to Microsoft Intune or M365 device management preferred
• Comfortable guiding professionals in live sessions
Benefits:
• Remote weekend engagement
• Build industry mentor brand authority
• Paid mentorship collaboration (structure discussed during call)
- Strong experience in Azure – mainly Azure ML Studio, AKS, Blob Storage, ADF, ADO Pipelines.
- Ability and experience to register and deploy ML/AI/GenAI models via Azure ML Studio.
- Working knowledge of deploying models in AKS clusters.
- Design and implement data processing, training, inference, and monitoring pipelines using Azure ML.
- Excellent Python skills – environment setup and dependency management, coding as per best practices, and knowledge of automatic code review tools like linting and Black.
- Experience with MLflow for model experiments, logging artifacts and models, and monitoring.
- Experience in orchestrating machine learning pipelines using MLOps best practices.
- Experience in DevOps with CI/CD knowledge (Git in Azure DevOps).
- Experience in model monitoring (drift detection and performance monitoring).
- Fundamentals of data engineering.
- Docker-based deployment is good to have.
Job Description -
Profile: Senior ML Lead
Experience Required: 10+ Years
Work Mode: Remote
Key Responsibilities:
- Design end-to-end AI/ML architectures including data ingestion, model development, training, deployment, and monitoring
- Evaluate and select appropriate ML algorithms, frameworks, and cloud platforms (Azure, Snowflake)
- Guide teams in model operationalization (MLOps), versioning, and retraining pipelines
- Ensure AI/ML solutions align with business goals, performance, and compliance requirements
- Collaborate with cross-functional teams on data strategy, governance, and AI adoption roadmap
Required Skills:
- Strong expertise in ML algorithms, Linear Regression, and modeling fundamentals
- Proficiency in Python with ML libraries and frameworks
- MLOps: CI/CD/CT pipelines for ML deployment with Azure
- Experience with OpenAI/Generative AI solutions
- Cloud-native services: Azure ML, Snowflake
- 8+ years in data science with at least 2 years in solution architecture role
- Experience with large-scale model deployment and performance tuning
Good-to-Have:
- Strong background in Computer Science or Data Science
- Azure certifications
- Experience in data governance and compliance
About This Opportunity
We're seeking a Solution Architect who thrives on autonomy and impact. If you're a tech-savvy innovator ready to architect cloud solutions at scale while working from anywhere in the world, this is your role.
What You'll Lead
Core Responsibilities
- Design scalable cloud architectures that serve as the technical foundation for enterprise applications
- Lead major projects end-to-end, from concept through deployment, owning architectural decisions and outcomes
- Architect high-performance APIs and microservices using modern tech stacks (Node.js, TypeScript, Java, GoLang)
- Integrate cutting-edge AI capabilities into solutions—leveraging Azure AI, OpenAI, and similar platforms to drive competitive advantage
- Own cloud strategy & optimization on Azure, ensuring security, scalability, and cost-efficiency
- Mentor technical teams and guide cross-functional stakeholders toward shared architectural vision
What We're Looking For
- 8+ years in solution architecture, cloud engineering, or equivalent leadership roles
- Technical depth across Node.js, TypeScript, Java, GoLang, and Generative AI frameworks
- Cloud mastery with Azure, Kubernetes, containerization, and CI/CD pipelines
- Leadership mindset: You drive decisions, mentor peers, and own project outcomes
- Communication excellence: You translate complex technical concepts for both engineers and business stakeholders
- Entrepreneurial spirit: You work best with autonomy and take ownership of major initiatives
Why Join Us
✅ 100% Remote – Work from home, a coffee shop, or anywhere globally
✅ Lead Significant Projects – Take ownership of architectural decisions that impact our platform
✅ Tech-Forward Culture – We invest in latest cloud, AI, and DevOps technologies
✅ Founded by Architects – Leadership team with 25+ years of cloud expertise—mentorship built in
✅ Rapid Innovation – We ship fast. You'll see your designs live in production within weeks
✅ Continuous Learning – Access to tools, courses, and conferences to sharpen your craft
Ready to Shape the Future?
If you're energized by architectural challenges and want the freedom to work from anywhere, we'd love to connect.
🌐 Learn more: prismcloudinc.com
𝐇𝐢 𝐂𝐨𝐧𝐧𝐞𝐜𝐭𝐢𝐨𝐧𝐬! 👋 𝐖𝐞𝐥𝐜𝐨𝐦𝐞 𝐭𝐨 2026! 🎉
Starting the new year with an exciting opportunity!
Deqode 𝐈𝐒 𝐇𝐈𝐑𝐈𝐍𝐆! 💻
Hiring: .Net Developer
⭐ Experience: 4+ Years
⭐ Work Mode: Remote
⏱️ Notice Period: Immediate Joiners
(Only immediate joiners & candidates serving notice period)
🔧 Role Overview
We are looking for passionate .NET Developers to design, develop, and maintain scalable microservices for enterprise-grade applications. You’ll work closely with cross-functional teams and clients on high-performance, cloud-native solutions.
🛠️ Key Responsibilities
✅Build and maintain scalable .NET microservices
✅Develop secure, high-quality RESTful Web APIs
✅Write unit and integration tests to ensure code quality
✅Optimize performance and implement caching strategies
💫 Must-Have Skills
✅ 4+ years of experience with .NET Core / .NET 5+ & C#
✅Strong hands-on experience with ASP.NET Core Web API & EF Core
✅REST API development & middleware implementation
✅Solid understanding of SOLID principles & design patterns
✅Unit testing experience (xUnit, NUnit, MSTest, Moq)
Hiring :.NET Full Stack Developer with ReactJS
Designation: Team Lead
Location: Bidadi, Bengaluru (Karnataka) ,Hybrid mode
Relevant Experience: 7-10 years
Preferred Qualifications
• Bachelors in CSE with minimum 7-10 years of relevant experience
• Exposure to cloud platforms (Azure) and API Gateway.
• Knowledge of microservices architecture.
• Experience with unit testing frameworks (xUnit, NUnit).
Required Skills & Qualifications
• Strong hands-on experience in .NET Core, C#, .Net framework (.NET Core, .NET 5+) and API development.
• Experience with RESTful API design and development.
• Strong experience on ReactJS for front-end development.
• Expertise in SQL Server (queries, stored procedures, performance tuning).
• Experience in system integration, especially with SAP.
• Ability to manage and mentor a team effectively.
• Strong requirement gathering and client communication skills.
• Familiarity with Git, CI/CD pipelines, and Agile methodologies.
Role Overview
• Design, develop, and maintain scalable backend services using .NET technologies.
• Work on ReactJS components as well as UI integration and ensure seamless communication between front-end and back-end.
• Write clean, efficient, and well-documented code.
• Lead and mentor a team of developers/Testing, ensuring adherence to best practices and timely delivery.
• Good exposure to Agile and scrum methodology
• Design and implement secure RESTful APIs using .NET Core.
• Apply best practices for authentication, authorization, and data security.
• Develop and maintain integrations with multiple systems, including SAP.
• Design and optimize SQL Server queries, stored procedures, and schemas.
• Gather requirements from clients and translate them into technical specifications.
• Implement Excel file uploaders and data processing workflows.
• Coordinate with stakeholders, manage timelines, and ensure quality deliverables.
• Troubleshoot and debug issues, ensuring smooth operation of backend systems.
Role: Senior Data Engineer (Azure)
Experience: 5+ Years
Location: Anywhere in india
Work Mode: Remote
Notice Period - Immediate joiners or Serving notice period
𝐊𝐞𝐲 𝐑𝐞𝐬𝐩𝐨𝐧𝐬𝐢𝐛𝐢𝐥𝐢𝐭𝐢𝐞𝐬:
- Data processing on Azure using ADF, Streaming Analytics, Event Hubs, Azure Databricks, Data Migration Services, and Data Pipelines
- Provisioning, configuring, and developing Azure solutions (ADB, ADF, ADW, etc.)
- Designing and implementing scalable data models and migration strategies
- Working on distributed big data batch or streaming pipelines (Kafka or similar)
- Developing data integration & transformation solutions for structured and unstructured data
- Collaborating with cross-functional teams for performance tuning and optimization
- Monitoring data workflows and ensuring compliance with governance and quality standards
- Driving continuous improvement through automation and DevOps practices
𝐌𝐚𝐧𝐝𝐚𝐭𝐨𝐫𝐲 𝐒𝐤𝐢𝐥𝐥𝐬 & 𝐄𝐱𝐩𝐞𝐫𝐢𝐞𝐧𝐜𝐞:
- 5–10 years of experience as a Data Engineer
- Strong proficiency in Azure Databricks, PySpark, Python, SQL, and Azure Data Factory
- Experience in Data Modelling, Data Migration, and Data Warehousing
- Good understanding of database structure principles and schema design
- Hands-on experience with MS SQL Server, Oracle, or similar RDBMS platforms
- Experience with DevOps tools (Azure DevOps, Jenkins, Airflow, Azure Monitor) — good to have
- Knowledge of distributed data processing and real-time streaming (Kafka/Event Hub)
- Familiarity with visualization tools like Power BI or Tableau
- Strong analytical, problem-solving, and debugging skills
- Self-motivated, detail-oriented, and capable of managing priorities effectively
Job Summary:
We are looking for a highly skilled and experienced DevOps Engineer who will be responsible for the deployment, configuration, and troubleshooting of various infrastructure and application environments. The candidate must have a proficient understanding of CI/CD pipelines, container orchestration, and cloud services, with experience in AWS services like EKS, EC2, ECS, EBS, ELB, S3, Route 53, RDS, ALB, etc., in a highly available and scalable production environment. The DevOps Engineer will be responsible for monitoring, automation, troubleshooting, security, user management, reporting, migrations, upgrades, disaster recovery, and infrastructure restoration, among other tasks. They will also work with application teams on infrastructure design and issues, and architect solutions to optimally meet business needs.
Responsibilities:
- Deploy, configure, and troubleshoot various infrastructure and application environments
- Work with AWS services like EC2, ECS, EBS, ELB, S3, Route 53, RDS, ALB, etc., in a highly available and scalable production environment
- Monitor, automate, troubleshoot, secure, maintain users, and report on infrastructure and applications
- Collaborate with application teams on infrastructure design and issues
- Architect solutions that optimally meet business needs
- Implement CI/CD pipelines and automate deployment processes
- Disaster recovery and infrastructure restoration
- Restore/Recovery operations from backups
- Automate routine tasks
- Execute company initiatives in the infrastructure space
- Expertise with observability tools like ELK, Prometheus, Grafana , Loki
Qualifications:
- Proficient understanding of CI/CD pipelines, container orchestration, and various cloud services
- Experience with AWS services like EC2, ECS, EBS, ELB, S3, Route 53, RDS, ALB, etc.
- Experience in monitoring, automation, troubleshooting, security, user management, reporting, migrations, upgrades, disaster recovery, and infrastructure restoration
- Experience in architecting solutions that optimally meet business needs
- Experience with scripting languages (e.g., Shell, Python) and infrastructure as code (IaC) tools (e.g., Terraform, CloudFormation)
- Strong understanding of system concepts like high availability, scalability, and redundancy
- Ability to work with application teams on infrastructure design and issues
- Excellent problem-solving and troubleshooting skills
- Experience with automation of routine tasks
- Good communication and interpersonal skills
Education and Experience:
- Bachelor's degree in Computer Science or a related field
- 5 to 10 years of experience as a DevOps Engineer or in a related role
- Experience with observability tools like ELK, Prometheus, Grafana
Working Conditions:
The DevOps Engineer will work in a fast-paced environment, collaborating with various application teams, stakeholders, and management. They will work both independently and in teams, and they may need to work extended hours or be on call to handle infrastructure emergencies.
Note: This is a remote role. The team member is expected to be in the Bangalore office for one week each quarter.
Role: Azure AI Tech Lead
Exp-3.5-7 Years
Location: Remote / Noida (NCR)
Notice Period: Immediate to 15 days
Mandatory Skills: Python, Azure AI/ML, PyTorch, TensorFlow, JAX, HuggingFace, LangChain, Kubeflow, MLflow, LLMs, RAG, MLOps, Docker, Kubernetes, Generative AI, Model Deployment, Prometheus, Grafana
JOB DESCRIPTION
As the Azure AI Tech Lead, you will serve as the principal technical expert leading the design, development, and deployment of advanced AI and ML solutions on the Microsoft Azure platform. You will guide a team of engineers, establish robust architectures, and drive end-to-end implementation of AI projects—transforming proof-of-concepts into scalable, production-ready systems.
Key Responsibilities:
- Lead architectural design and development of AI/ML solutions using Azure AI, Azure OpenAI, and Cognitive Services.
- Develop and deploy scalable AI systems with best practices in MLOps across the full model lifecycle.
- Mentor and upskill AI/ML engineers through technical reviews, training, and guidance.
- Implement advanced generative AI techniques including LLM fine-tuning, RAG systems, and diffusion models.
- Collaborate cross-functionally to translate business goals into innovative AI solutions.
- Enforce governance, responsible AI practices, and performance optimization standards.
- Stay ahead of trends in LLMs, agentic AI, and applied research to shape next-gen solutions.
Qualifications:
- Bachelor’s or Master’s in Computer Science or related field.
- 3.5–7 years of experience delivering end-to-end AI/ML solutions.
- Strong expertise in Azure AI ecosystem and production-grade model deployment.
- Deep technical understanding of ML, DL, Generative AI, and MLOps pipelines.
- Excellent analytical and problem-solving abilities; applied research or open-source contributions preferred.
Infrastructure Engineer – Database & Storage
Responsibilities
- Design and maintain PostgreSQL, OpenSearch, and Azure Blob/S3 clusters.
- Implement schema registry, metadata catalog, and time-versioned storage.
- Configure read replicas, backups, encryption-at-rest, and WORM (Write Once Read Many) compliance.
- Optimize query execution, indexing, and replication latency.
- Partner with DevOps on infrastructure as code and cross-region replication.
Requirements
- 6 + years database / data-infrastructure administration.
- Mastery of indexing, partitioning, query tuning, sharding.
- Proven experience deploying cloud-native DB stacks with Terraform or Helm.
What You Will be Doing:
● Develop and maintain software that is scalable, secure, and efficient
● Collaborate with Technical Architects & Business Analysts
● Architect and design software solutions that meet project requirements
● Mentor and train junior developers to improve their skills and knowledge
● Conduct code reviews ensuring the code is maintainable, readable, and efficient
● Research and evaluate new technologies to improve the processes
● Effective communication skills, particularly in documenting and explaining code and technical concepts.
Skills We Are Looking For:
● 5+ years extensive hands-on experience with NodeJS and Typescript
● Strong understanding of RESTful API design and implementation.
● Comfortable with debugging, performance tuning, and optimizing Node.js applications.
● Strong problem-solving abilities and attention to detail.
● Experience with authentication and authorization protocols, such as OAuth, JWT and session management.
● Understanding of security best practices in backend development, including data encryption and vulnerability mitigation.
Bonus Skills
● Experience with server-side frameworks such as Express.js or NestJS.
● Familiarity with cloud platforms (e.g., AWS, Azure, (preferred) Google Cloud) and their services for backend deployment.
● Familiarity with NoSQL databases (Mongo preferred), and the ability to design and optimize database queries.
Why You’ll Love It Here
● Innovative Culture - We believe in pushing boundaries
● Impactful Work - You won’t just write code, you will help build the future
● Collaborative Environment - We believe that everyone has a voice that matters
● Work Life Balance - Our flexible work environment encourages you to have space to
recharge
Position Name:- Azure Infrastructure Engineer
Year of Exp:- 8+ years
Mode :- Remote
Key Important Skill Set:-
Cloud Networking
Azure Networking
Hub-and-Spoke Architecture
Azure Kubernetes Service (AKS)
Terraform Automation
State File Management (Terraform)
Disaster Recovery (DR) Setup
High Availability Configuration
Network Troubleshooting (Azure)
AKS Networking
Scaling and Load Balancing (AKS)
VM Deployment (Terraform)
Service Mesh
Helm for AKS
Production-Level Infrastructure Management
Network Security Groups (NSG)
Container Management (AKS)
Detailed Job Description: -
Key Responsibilities:
Operate and troubleshoot production AKS clusters
Build and deploy workloads using Docker and Helm
Automate infra provisioning with Terraform
Configure autoscaling using KEDA and HPA
Manage Istio or equivalent service mesh (ingress, routing, mTLS)
Maintain robust CI/CD pipelines (Azure DevOps/GitHub Actions)
Handle complex Azure networking (VNet, NSG, DNS, LB, peering)
Support and execute disaster recovery procedures
Required Skills:
- 4+ years of experience designing, developing, and implementing enterprise-level, n-tier, software solutions.
- Proficiency with Microsoft C# is a must.
- In-depth experience with .NET framework and .NET Core.
- Knowledge of OOP, server technologies, and SOA is a must. 3+ Years Micro-service experience .
- Relevant experience with database design and SQL (Postgres is preferred).
- Experience with ORM tooling.
- Experience delivering software that is correct, stable, and security compliant.
- Basic understanding of common cloud platform. (Good to have)
- Financial services experience is strongly preferred.
- Thorough understanding of XML/JSON and related technologies.
- Thorough understanding of unit, integration, and performance testing for APIs.
- Entrepreneurial spirit. You are self-directed, innovative, and biased towards action. You love to build new things and thrive in fast-paced environments.
- Excellent communication and interpersonal skills, with an emphasis on strong writing and analytical problem-solving.
Full Stack Engineer (Frontend Strong, Backend Proficient)
5-10 Years Experience
Contract: 6months+extendable
Location: Remote
Technical Requirements Frontend Expertise (Strong)
*Need at least 4 Yrs in React web developement, Node & AI.*
● Deep proficiency in React, Next.js, TypeScript
● Experience with state management (Redux, Context API)
● Frontend testing expertise (Jest, Cypress)
● Proven track record of achieving high Lighthouse performance scores Backend Proficiency
● Solid experience with Node.js, NestJS (preferred), or ExpressJS
● Database management (SQL, NoSQL)
● Cloud technologies experience (AWS, Azure)
● Understanding of OpenAI and AI integration capabilities (bonus) Full Stack Integration
● Excellent ability to manage and troubleshoot integration issues between frontend and backend systems
● Experience designing cohesive systems with proper separation of concerns
Position: Lead Backend Engineer
Location: Remote
Experience: 10+ Years
Budget: 1.7 LPM
Employment Type: [Contract]
Required Skills & Qualifications:
- 10+ years of proven backend engineering experience.
- Strong proficiency in Python.
- Expertise in SQL (Postgres) and database optimization.
- Hands-on experience with OpenAI APIs.
- Strong command of FastAPI and microservices architecture.
- Solid knowledge of debugging, troubleshooting, and performance tuning.
Nice to Have:
- Experience with Agentic Systems or ability to quickly adopt them.
- Exposure to modern CI/CD pipelines, containerization (Docker/Kubernetes), and cloud platforms (AWS, Azure, or GCP).
Role: Cleo EDI Solution Architect / Sr EDI Developer
Location : Remote
Start Date – asap
This is a niche technology (Cleo EDI), which enables the integration of ERP with Transp. Mgt/Extended Supply Chain etc
Expertise in designing and developing end-to-end integration solutions, especially B2B integrations involving EDI (Electronic Data Interchange) and APIs.
Familiarity with Cleo Integration Cloud or similar EDI platforms.
Strong experience with Azure Integration Services, particularly:
- Azure Data Factory – for orchestrating data movement and transformation
- Azure Functions – for serverless compute tasks in integration pipelines
- Azure Logic Apps or Service Bus – for message handling and triggering workflows
Understanding of ETL/ELT processes and data mapping.
Solid grasp of EDI standards (e.g., X12, EDIFACT) and workflows.
Experience working with EDI developers and analysts to align business requirements with technical implementation.
Familiarity with Cleo EDI tools or similar platforms.
Develop and maintain EDI integrations using Cleo Integration Cloud (CIC), Cleo Clarify, or similar Cleo solutions.
Create, test, and deploy EDI maps for transactions such as 850, 810, 856, etc., and other EDI/X12/EDIFACT documents.
Configure trading partner setups, including communication protocols (AS2, SFTP, FTP, HTTPS).
Monitor EDI transaction flows, identify errors, troubleshoot, and implement fixes.
Collaborate with business analysts, ERP teams, and external partners to gather and analyze EDI requirements.
Document EDI processes, mappings, and configurations for ongoing support and knowledge sharing.
Provide timely support for EDI-related incidents, ensuring minimal disruption to business operations.
Participate in EDI onboarding projects for new trading partners and customers.
At TechBiz Global, we are providing recruitment service to our TOP clients from our portfolio. We are currently seeking 4 DevOps Support Engineer to join one of our clients' teams in India who can start until 20th of July. If you're looking for an exciting opportunity to grow in a innovative environment, this could be the perfect fit for you.
Job requirements
Key Responsibilities:
- Monitor and troubleshoot AWS and/or Azure environments to ensure optimal performance and availability.
- Respond promptly to incidents and alerts, investigating and resolving issues efficiently.
- Perform basic scripting and automation tasks to streamline cloud operations (e.g., Bash, Python).
- Communicate clearly and fluently in English with customers and internal teams.
- Collaborate closely with the Team Lead, following Standard Operating Procedures (SOPs) and escalation workflows.
- Work in a rotating shift schedule, including weekends and nights, ensuring continuous support coverage.
Shift Details:
- Engineers rotate shifts, typically working 4–5 shifts per week.
- Each engineer works about 4 to 5 shifts per week, rotating through morning, evening, and night shifts—including weekends—to cover 24/7 support evenly among the team
- Rotation ensures no single engineer is always working nights or weekends; the load is shared fairly among the team.
Qualifications:
- 2–5 years of experience in DevOps or cloud support roles.
- Strong familiarity with AWS and/or Azure cloud environments.
- Experience with CI/CD tools such as GitHub Actions or Jenkins.
- Proficiency with monitoring tools like Datadog, CloudWatch, or similar.
- Basic scripting skills in Bash, Python, or comparable languages.
- Excellent communication skills in English.
- Comfortable and willing to work in a shift-based support role, including night and weekend shifts.
- Prior experience in a shift-based support environment is preferred.
What We Offer:
- Remote work opportunity — work from anywhere in India with a stable internet connection.
- Comprehensive training program including:
- Shadowing existing processes to gain hands-on experience.
- Learning internal tools, Standard Operating Procedures (SOPs), ticketing systems, and escalation paths to ensure smooth onboarding and ongoing success.
We’re looking for a .NET Core Developer who’s confident in their skills and ready to take on meaningful, real-world projects. If you're someone who enjoys solving problems, writing clean code, and working with Azure cloud services, we’d love to hear from you.
What you'll be doing:
- Building and maintaining scalable web applications using .NET Core
- Working with Azure services (like Functions, App Services, Logic Apps, etc.)
- Collaborating with cross-functional teams to design, develop, and deliver high-quality solutions
- Writing clean, testable, and efficient code
- Participating in code reviews and team discussions to improve practices
We’re looking for someone who has:
- 4.5 to 7 years of hands-on experience in .NET Core and C#
- Practical experience with Microsoft Azure – especially services like Azure Functions, App Services, Logic Apps, or Azure DevOps
- Good understanding of RESTful APIs, microservices, and SQL
- Strong problem-solving skills and the ability to work independently or as part of a team
Why this role?
- Flexible hybrid work model – work from anywhere, come together when it counts
- Opportunity to work on innovative projects with a supportive team
- A space to grow, learn, and actually enjoy the work you do
Job Title : Azure DevOps Engineer
Experience Required : 7+ Years
Work Mode : Remote / Hybrid
Location : Remote
Notice Period : Immediate Joiners / Serving Candidates (within 20 days only)
Interview Mode : Face-to-Face or Virtual
Open Positions : 2
Job Description :
We are seeking an experienced Azure DevOps Engineer with 7+ years of relevant experience in DevOps practices, especially around Azure infrastructure, deployment automation, and CI/CD pipeline management. The ideal candidate should have hands-on expertise with Azure DevOps, GitHub, YAML, and Azure services, along with solid communication and coordination capabilities.
Mandatory Skills : Azure DevOps, GitHub Actions, YAML, Bicep, Azure services (App Gateway, WAF, NSG, CosmosDB, Storage Accounts), Unix scripting, and Azure Fundamentals certification.
Key Responsibilities :
- Manage deployments for Dynamics 365 and proxy applications
- Run and maintain ADO pipelines and GitHub Actions
- Ensure proper status updates on ADO Boards and deployed work items
- Coordinate with QA teams to execute smoke testing post-deployment
- Communicate deployment progress across team channels effectively
- Monitor deployment cycles, approval gates, logs, and alerts
- Ensure smooth integration of infrastructure and DevOps practices
Mandatory Skills :
- Minimum 7+ years in DevOps, with strong experience in Azure DevOps (ADO).
- Proven expertise in building pipelines using Azure DevOps and GitHub.
- Proficiency in Bicep, YAML scripting, and Azure Infrastructure-as-Code (IaC).
- Hands-on with Azure services like :
- App Gateway, WAF, NSG, CosmosDB, Storage Accounts.
- vNet, Managed Identity, KeyVault, AppConfig, App Insights.
- Basic Azure Fundamentals Certification (AZ-900).
- Excellent communication skills in English.
Nice to Have :
- Experience in managing large enterprise-scale deployments.
- Familiarity with branching strategies and monitoring tools.
- Exposure to Approval Gates and Deployment Governance.
Job Role : Azure DevSecOps Engineer (Security-Focused)
Experience : 12 to 18 Years
Location : Preferably Delhi NCR (Hybrid); Remote possible with 1–2 office visits per quarter (Gurgaon)
Joining Timeline : Max 45 days (Buyout option available)
Work Mode : Full-time | 5 Days Working
About the Role :
We are looking for a highly experienced Azure DevSecOps Engineer with a strong focus on cloud security practices.
This role is 60–70% security-driven, involving threat modeling, secure cloud architecture, and infrastructure security on Azure using Terraform.
Key Responsibilities :
- Architect and maintain secure, scalable Azure cloud infrastructure using Terraform.
- Implement security best practices : IAM, threat modeling, network security, data protection, and compliance (e.g., GDPR).
- Build CI/CD pipelines and automate deployments using Azure DevOps, Jenkins, Prometheus.
- Monitor, analyze, and proactively improve security posture.
- Collaborate with global teams to ensure secure design, development, and operations.
- Stay updated on cloud security trends and lead mitigation efforts.
Mandatory Skills :
Azure, Terraform, DevSecOps, Cloud Security, Threat Modelling, IAM, CI/CD (Azure DevOps), Docker, Kubernetes, Prometheus, Infrastructure as Code (IaC), Compliance Frameworks (GDPR)
Preferred Certifications :
Certified DevSecOps Professional (CDP), Microsoft Azure Certifications
Company name: PulseData labs Pvt Ltd (captive Unit for URUS, USA)
About URUS
We are the URUS family (US), a global leader in products and services for Agritech.
SENIOR DATA ENGINEER
This role is responsible for the design, development, and maintenance of data integration and reporting solutions. The ideal candidate will possess expertise in Databricks and strong skills in SQL Server, SSIS and SSRS, and experience with other modern data engineering tools such as Azure Data Factory. This position requires a proactive and results-oriented individual with a passion for data and a strong understanding of data warehousing principles.
Responsibilities
Data Integration
- Design, develop, and maintain robust and efficient ETL pipelines and processes on Databricks.
- Troubleshoot and resolve Databricks pipeline errors and performance issues.
- Maintain legacy SSIS packages for ETL processes.
- Troubleshoot and resolve SSIS package errors and performance issues.
- Optimize data flow performance and minimize data latency.
- Implement data quality checks and validations within ETL processes.
Databricks Development
- Develop and maintain Databricks pipelines and datasets using Python, Spark and SQL.
- Migrate legacy SSIS packages to Databricks pipelines.
- Optimize Databricks jobs for performance and cost-effectiveness.
- Integrate Databricks with other data sources and systems.
- Participate in the design and implementation of data lake architectures.
Data Warehousing
- Participate in the design and implementation of data warehousing solutions.
- Support data quality initiatives and implement data cleansing procedures.
Reporting and Analytics
- Collaborate with business users to understand data requirements for department driven reporting needs.
- Maintain existing library of complex SSRS reports, dashboards, and visualizations.
- Troubleshoot and resolve SSRS report issues, including performance bottlenecks and data inconsistencies.
Collaboration and Communication
- Comfortable in entrepreneurial, self-starting, and fast-paced environment, working both independently and with our highly skilled teams.
- Collaborate effectively with business users, data analysts, and other IT teams.
- Communicate technical information clearly and concisely, both verbally and in writing.
- Document all development work and procedures thoroughly.
Continuous Growth
- Keep abreast of the latest advancements in data integration, reporting, and data engineering technologies.
- Continuously improve skills and knowledge through training and self-learning.
This job description reflects managements assignment of essential functions; it does not prescribe or restrict the tasks that may be assigned.
Requirements
- Bachelor's degree in computer science, Information Systems, or a related field.
- 7+ years of experience in data integration and reporting.
- Extensive experience with Databricks, including Python, Spark, and Delta Lake.
- Strong proficiency in SQL Server, including T-SQL, stored procedures, and functions.
- Experience with SSIS (SQL Server Integration Services) development and maintenance.
- Experience with SSRS (SQL Server Reporting Services) report design and development.
- Experience with data warehousing concepts and best practices.
- Experience with Microsoft Azure cloud platform and Microsoft Fabric desirable.
- Strong analytical and problem-solving skills.
- Excellent communication and interpersonal skills.
- Ability to work independently and as part of a team.
- Experience with Agile methodologies.
Title: Senior Software Engineer – Python (Remote: Africa, India, Portugal)
Experience: 9 to 12 Years
INR : 40 LPA - 50 LPA
Location Requirement: Candidates must be based in Africa, India, or Portugal. Applicants outside these regions will not be considered.
Must-Have Qualifications:
- 8+ years in software development with expertise in Python
- kubernetes is important
- Strong understanding of async frameworks (e.g., asyncio)
- Experience with FastAPI, Flask, or Django for microservices
- Proficiency with Docker and Kubernetes/AWS ECS
- Familiarity with AWS, Azure, or GCP and IaC tools (CDK, Terraform)
- Knowledge of SQL and NoSQL databases (PostgreSQL, Cassandra, DynamoDB)
- Exposure to GenAI tools and LLM APIs (e.g., LangChain)
- CI/CD and DevOps best practices
- Strong communication and mentorship skills
Overview
As an engineer in the Service Operations division, you will be responsible for the day-to-day management of the systems and services that power client products. Working with your team, you will ensure daily tasks and activities are successfully completed and where necessary, use standard operating procedures and knowledge to resolve any faults/errors encountered.
Job Description
Key Tasks and Responsibilities:
Ensure daily tasks and activities have successfully completed. Where this is not the case, recovery and remediation steps will be undertaken.
Undertake patching and upgrade activities in support of ParentPay compliance programs. These being PCI DSS, ISO27001 and Cyber Essentials+.
Action requests from the ServiceNow work queue that have been allocated to your relevant resolver group. These include incidents, problems, changes and service requests.
Investigate alerts and events detected from the monitoring systems that indicate a change in component health.
Create and maintain support documentation in the form of departmental wiki and ServiceNow knowledge articles that allow for continual improvement of fault detection and recovery times.
Work with colleagues to identify and champion the automation of all manual interventions undertaken within the team.
Attend and complete all mandatory training courses.
Engage and own the transition of new services into Service Operations.
Participate in the out of hours on call support rota.
Qualifications and Experience:
Experience working in an IT service delivery or support function OR
MBA or Degree in Information Technology or Information Security.
Experience working with Microsoft technologies.
Excellent communication skills developed working in a service centric organisation.
Ability to interpret fault descriptions provided by customers or internal escalations and translate these into resolutions.
Ability to manage and prioritise own workload.
Experience working within Education Technology would be an advantage.
Technical knowledge:
Advanced automation scripting using Terraform and Powershell.
Knowledge of bicep and ansible advantageous.
Advanced Microsoft Active Directory configuration and support.
Microsoft Azure and AWS cloud hosting platform administration.
Advanced Microsoft SQL server experience.
Windows Server and desktop management and configuration.
Microsoft IIS web services administration and configuration.
Advanced management of data and SQL backup solutions.
Advanced scripting and automation capabilities.
Advanced knowledge of Azure analytics and KQL.
Skills & Requirements
IT Service Delivery, Information Technology, Information Security, Microsoft Technologies, Communication Skills, Fault Interpretation, Workload Prioritization, Automation Scripting, Terraform, PowerShell, Microsoft Active Directory, Microsoft Azure, AWS, Microsoft SQL Server, Windows Server, Windows Desktop Configuration, Microsoft IIS, Data Backup Management, SQL Backup Solutions, Scripting, Azure Analytics, KQL.
Immediate Joiners Preferred. Notice Period - Immediate to 30 Days
Interested candidates are requested to email their resumes with the subject line "Application for [Job Title]".
Only applications received via email will be reviewed. Applications through other channels will not be considered.
Overview
Adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job Description
As part of our dynamic international cross-functional team you will be responsible for the design, development and deployment of modern high quality software solutions and applications as an experienced and skilled Full-stack PHP Developer.
Responsibilities:
Design, develop, and maintain the application.
Write clean, efficient, and reusable code.
Implement new features and functionality based on business requirements.
Participate in system and application architecture discussions.
Create technical designs and specifications for new features or enhancements.
Write and execute unit tests to ensure code quality.
Debug and resolve technical issues and software defects.
Conduct code reviews to ensure adherence to best practices.
Identify and fix vulnerabilities to ensure application integrity.
Working with other developers to ensure seamless integration backend and frontend elements.
Collaborating with DevOps teams for deployment and scaling.
Requirements:
Bachelor’s degree in computer science or information technology, or a related field.
Proven experience as a skilled Full-stack developer with focus on PHP and web technologies.
Strong experience with PHP 8 and 7, Symfony Framework, AWS / Azure or GCP, GitLab and Angular and / or React. Additional technologies like Java, Python, Go, Kotlin, Rust or similar are welcome.
Experienced with test driven development and QA Tools like f.e. PHPstan, Deptrac.
Excellent problem-solving and debugging skills.
Strong communication and collaboration abilities to work effectively in a team environment.
Skills & Requirements
Full-stack development, PHP 8, PHP 7, Symfony Framework, AWS, Azure, GCP, GitLab, Angular, React, Java, Python, Go, Kotlin, Rust, Test-driven development, QA tools, PHPStan, Deptrac, Problem-solving, Debugging, Communication, Collaboration, System architecture, Technical design, Unit testing, Code review, Deployment, Scalability.
Overview
adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job Description
We are seeking a skilled Cloud Data Engineer who has experience with cloud data platforms like AWS or Azure and especially Snowflake and dbt to join our dynamic team. As a consultant, you will be responsible for developing new data platforms and create the data processes. You will collaborate with cross-functional teams to design, develop, and deploy high-quality frontend solutions.
Responsibilities:
Customer consulting: You develop data-driven products in the Snowflake Cloud and connect data & analytics with specialist departments. You develop ELT processes using dbt (data build tool)
Specifying requirements: You develop concrete requirements for future-proof cloud data architectures.
Develop data routes: You design scalable and powerful data management processes.
Analyze data: You derive sound findings from data sets and present them in an understandable way.
Requirements:
Requirements management and project experience: You successfully implement cloud-based data & analytics projects.
Data architectures: You are proficient in DWH/data lake concepts and modeling with Data Vault 2.0.
Cloud expertise: You have extensive knowledge of Snowflake, dbt and other cloud technologies (e.g. MS Azure, AWS, GCP).
SQL know-how: You have a sound and solid knowledge of SQL.
Data management: You are familiar with topics such as master data management and data quality.
Bachelor's degree in computer science, or a related field.
Strong communication and collaboration abilities to work effectively in a team environment.
Skills & Requirements
Cloud Data Engineering, AWS, Azure, Snowflake, dbt, ELT processes, Data-driven consulting, Cloud data architectures, Scalable data management, Data analysis, Requirements management, Data warehousing, Data lake, Data Vault 2.0, SQL, Master data management, Data quality, GCP, Strong communication, Collaboration.
Overview
Adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job description
As part of our dynamic international cross-functional team you will be responsible for the design, development and deployment of modern high quality software solutions and applications as an experienced and skilled Full-stack developer.
Responsibilities:
Design, develop, and maintain the application
Write clean, efficient, and reusable code
Implement new features and functionality based on business requirements
Participate in system and application architecture discussions
Create technical designs and specifications for new features or enhancements
Write and execute unit tests to ensure code quality
Debug and resolve technical issues and software defects
Conduct code reviews to ensure adherence to best practices
Identify and fix vulnerabilities to ensure application integrity
Working with other developers to ensure seamless integration backend and frontend elements
Collaborating with DevOps teams for deployment and scaling
Requirements:
Bachelor’s degree in computer science or information technology, or a related field.
Proven experience as a skilled Full-stack developer. Experience in Utilities / Energy domain is appreciated.
Strong experience with Java (Springboot), AWS / Azure or GCP, GitLab and Angular and / or React. Additional technologies like Python, Go, Kotlin, Rust or similar are welcome
Excellent problem-solving and debugging skills.
Strong communication and collaboration abilities to work effectively in a team environment.
Skills & Requirements
Java, Spring Boot, AWS, Azure, GCP, GitLab, Angular, React, Python, Go, Kotlin, Rust, Full-stack development, Software architecture, Unit testing, Debugging, Code reviews, DevOps collaboration, Microservices, Cloud computing, RESTful APIs, Frontend-backend integration, Problem-solving, Communication, Team collaboration, Software deployment, Application security, Technical documentation.
Overview
adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job Description
We are looking for an experienced Technical Team Lead to guide a local IT Services Management Team and also acting as a software developer. In this role, you will be responsible for the application management of a B2C application to meet the agreed Service Level Agreements (SLAs) and fulfil customer expectations.
Your Team will act as a on-call-duty team in the time between 6 pm to 8 am, 365 days a year. You will work together with the responsible Senior Project Manager in Germany.
We are seeking a hands-on leader who thrives in both team management and operational development. Whether you have experience in DevOps and Backend or Frontend, your expertise in both leadership and technical skills will be key to success in this position.
Responsibilities:
Problem Management & Incident Management activities: Identifying and resolving technical issues and errors that arise during application usage.
Release and Update Coordination: Planning and executing software updates, new versions, or system upgrades to keep applications up to date.
Change Management: Responsible for implementing and coordinating changes to the application, considering the impact on ongoing operations.
Requirements:
Education und Experience: A Bachelor’s or Master’s degree in a relevant field, with a minimum of 5 years of professional experience or equivalent work experience.
Skills & Expertise:
Proficient in ITIL service management frameworks.
Strong analytical and problem-solving abilities.
Experienced in project management methodologies (Agile, Kanban).
Leadership: Very good leadership skills with a customer orientated, proactive and results driven approach.
Communication: Excellent communication, presentation, and interpersonal skills, with the ability to engage and collaborate with stakeholders.
Language: English on a C2 Level.
Skills & Requirements
kubeAPI high Kustomize high docker/container high Debug Tools openSSL high Curl high Azure Devops, Pipeline, Repository, Deployment, ArgoCD, Certificates: Certificate Management / SSL, LetsEncrypt, Linux Shell, Keycloak.
Overview
Adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job description
The client’s department DPS, Digital People Solutions, offers a sophisticated portfolio of IT applications, providing a strong foundation for professional and efficient People & Organization (P&O) and Business Management, both globally and locally, for a well-known German company listed on the DAX-40 index, which includes the 40 largest and most liquid companies on the Frankfurt Stock Exchange.
We are seeking talented Java Application Developers to join our dynamic DPS team. In this role, you will design and implement change requests for existing applications or develop new projects using Jakarta EE (Java Enterprise Technologies) and Angular for the frontend. Your responsibilities will include end-to-end process mapping within the HR application landscape, analyzing developed functionalities, and addressing potential issues.
As part of our dynamic international cross-functional team you will be responsible for the design, development and deployment of modern high quality software solutions and applications as an experienced and skilled Full-stack developer.
Responsibilities:
Design, develop, and maintain the application.
Write clean, efficient, and reusable code.
Implement new features and functionality based on business requirements.
Participate in system and application architecture discussions.
Create technical designs and specifications for new features or enhancements.
Write and execute unit tests to ensure code quality.
Debug and resolve technical issues and software defects.
Conduct code reviews to ensure adherence to best practices.
Identify and fix vulnerabilities to ensure application integrity.
Working with other developers to ensure seamless integration backend and frontend elements.
Collaborating with DevOps teams for deployment and scaling.
Requirements:
Bachelor’s degree in computer science or information technology, or a related field.
Proven experience as a skilled Full-stack developer. Experience in Utilities / Energy domain is appreciated.
Strong experience with Java (Springboot), AWS / Azure or GCP, GitLab and Angular and / or React. Additional technologies like Python, Go, Kotlin, Rust or similar are welcome
Excellent problem-solving and debugging skills.
Strong communication and collaboration abilities to work effectively in a team environment.
Skills & Requirements
Java, Spring Boot, Jakarta EE, Angular, React, AWS, Azure, GCP, GitLab, Python, Go, Kotlin, Rust, Full-stack Development, Unit Testing, Debugging, Code Review, DevOps, Software Architecture, Microservices, HR Applications, Cloud Computing, Frontend Development, Backend Development, System Integration, Technical Design, Deployment, Problem-Solving, Communication, Collaboration.
Interested candidates are requested to email their resumes with the subject line "Application for [Job Title]".
Only applications received via email will be reviewed. Applications through other channels will not be considered.
Overview
Adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job description
The client’s department DPS, Digital People Solutions, offers a sophisticated portfolio of IT applications, providing a strong foundation for professional and efficient People & Organization (P&O) and Business Management, both globally and locally, for a well-known German company listed on the DAX-40 index, which includes the 40 largest and most liquid companies on the Frankfurt Stock Exchange.
We are seeking talented Java Application Developers to join our dynamic DPS team. In this role, you will design and implement change requests for existing applications or develop new projects using Jakarta EE (Java Enterprise Technologies) and Angular for the frontend. Your responsibilities will include end-to-end process mapping within the HR application landscape, analyzing developed functionalities, and addressing potential issues.
As part of our dynamic international cross-functional team you will be responsible for the design, development and deployment of modern high quality software solutions and applications as an experienced and skilled Full-stack developer.
Responsibilities:
Design, develop, and maintain the application.
Write clean, efficient, and reusable code.
Implement new features and functionality based on business requirements.
Participate in system and application architecture discussions.
Create technical designs and specifications for new features or enhancements.
Write and execute unit tests to ensure code quality.
Debug and resolve technical issues and software defects.
Conduct code reviews to ensure adherence to best practices.
Identify and fix vulnerabilities to ensure application integrity.
Working with other developers to ensure seamless integration backend and frontend elements.
Collaborating with DevOps teams for deployment and scaling.
Requirements:
Bachelor’s degree in computer science or information technology, or a related field.
Proven experience as a skilled Full-stack developer. Experience in Utilities / Energy domain is appreciated.
Strong experience with Java (Springboot), AWS / Azure or GCP, GitLab and Angular and / or React. Additional technologies like Python, Go, Kotlin, Rust or similar are welcome
Excellent problem-solving and debugging skills.
Strong communication and collaboration abilities to work effectively in a team environment.
Skills & Requirements
Java, Spring Boot, Jakarta EE, Angular, React, AWS, Azure, GCP, GitLab, Python, Go, Kotlin, Rust, Full-stack Development, Unit Testing, Debugging, Code Review, DevOps, Software Architecture, Microservices, HR Applications, Cloud Computing, Frontend Development, Backend Development, System Integration, Technical Design, Deployment, Problem-Solving, Communication, Collaboration.
Interested candidates are requested to email their resumes with the subject line "Application for [Job Title]".
Only applications received via email will be reviewed. Applications through other channels will not be considered.
Overview
Adesso India specialises in optimization of core business processes for organizations. Our focus is on providing state-of-the-art solutions that streamline operations and elevate productivity to new heights.
Comprised of a team of industry experts and experienced technology professionals, we ensure that our software development and implementations are reliable, robust, and seamlessly integrated with the latest technologies. By leveraging our extensive knowledge and skills, we empower businesses to achieve their objectives efficiently and effectively.
Job description
As part of our dynamic international cross-functional team you will be responsible for the design, development and deployment of modern high quality software solutions and applications as an experienced and skilled Full-stack developer.
Responsibilities:
Design, develop, and maintain the application
Write clean, efficient, and reusable code
Implement new features and functionality based on business requirements
Participate in system and application architecture discussions
Create technical designs and specifications for new features or enhancements
Write and execute unit tests to ensure code quality
Debug and resolve technical issues and software defects
Conduct code reviews to ensure adherence to best practices
Identify and fix vulnerabilities to ensure application integrity
Working with other developers to ensure seamless integration backend and frontend elements
Collaborating with DevOps teams for deployment and scaling
Requirements:
Bachelor’s degree in computer science or information technology, or a related field.
Proven experience as a skilled Full-stack developer. Experience in Utilities / Energy domain is appreciated.
Strong experience with Java (Springboot), AWS / Azure or GCP, GitLab and Angular and / or React. Additional technologies like Python, Go, Kotlin, Rust or similar are welcome
Excellent problem-solving and debugging skills.
Strong communication and collaboration abilities to work effectively in a team environment.
Skills & Requirements
Java, Spring Boot, AWS, Azure, GCP, GitLab, Angular, React, Python, Go, Kotlin, Rust, Full-stack development, Software architecture, Unit testing, Debugging, Code reviews, DevOps collaboration, Microservices, Cloud computing, RESTful APIs, Frontend-backend integration, Problem-solving, Communication, Team collaboration, Software deployment, Application security, Technical documentation.
Job Purpose
We are looking for an energetic and self-starter software developer to join our product development practice as a Senior Software Engineer (SSE).
You will get to work with some of the best and knowledgeable tech talent in the financial world and you will build next generation digital services and platforms that will lead the transformation goals for our customers. You will work closely with the engineering, UX, product and test automation communities, as part of the agile team, to lead product design and development and to help the Digital Service Product Owner to deliver and maximize value.
You will drive engineering and architecture best practices for writing and encouraging others to write secure code and dev-ops process while getting opportunities for learning new business domain and topics, to work with industry SMEs and to learn new technology and behavioral skills.
Key Responsibilities
As a Full-stack Developer
- 8+ year’s professional experience in enterprise software design and development in an N-tier architecture environment;
- Understanding of 12-factor app framework is highly desirable
- Must have experience building web applications using .NET core 3.x (.NET 5.0 is better), Web API, HTML5, React OR other JS-based frameworks like Angular
- Must have experience with tools such as Jira, Github, Confluence (or other wiki), SonarQube (or similar), OWASP ZAP (or similar) and Snyk (or similar)
- Experience with data visualization libraries /framework like D3js, Plotly, HighCharts etc. will be an advantage
- Must have experience with SOA and Web Service standards (REST & JSON/SOAP & WSDL/WS-I Basic Profile), and IIS
- Understand the business requirements from the product owner(s)
- Design and implement the system from scratch & build enhancements, features request using modern application frameworks using C# and React with .NET Core, Web API, AWS services etc.
- Participate in both development & maintenance tasks
- Independently troubleshoot difficult and complex issues on production and other environments
As a Technical Lead in the pod
- Must have experience of working in an automated CI/CD environment and with fast moving teams using Scrum/Agile; Experience with AWS and other cloud providers is highly desirable
- Must have extensive experience with object oriented design principles. Ability to articulate the pros and cons of design/implementation options
- Participate in design review and peer code review
- Work collaboratively in a global setting, should be eager to learn new technologies
- Responsible for extending and maintaining existing codebase with focus on quality, re-usability, maintainability and consistency
- Coach teams on best practices and architecture design
As member of the Engineering community
- Must have extensive experience with object oriented design principles.
- Ability to articulate the pros and cons of design/implementation options
- Good understanding and knowledge of areas including but not limited to requirement gathering, designing, development, testing, maintenance, quality control etc.
- Stay up-to-date on latest developments in technology
- Learn and share learnings with the community
Behavioral Competencies
- A self-starter, excellent planner and executor and above all, a good team player
- Excellent communication skills and inter-personal skills are a must
- Must have organizational skills, including multi-task capability, priority setting and meeting deadlines
- Ability to build collaborative relationships and effectively leverage networks to mobilize resources
- Liking and initiative to learn business domain is highly desirable
- Likes dynamic and constantly evolving environment and requirements
Job Title : AI Engineer / Gen AI Engineer (LLM/NLP)
Location : Remote
Duration : Long-term Contract
Openings : 2
Job Description :
We are looking for two AI Engineers with expertise in Generative AI, LLMs, and NLP. One of them must have a strong Azure background and be well-versed in Azure AI services.
Key Requirements :
- Experience in Generative AI, LLMs, and NLP
- Azure AI Expertise (For one of the roles):
- Azure AI Search
- Azure AI Document Intelligence
- Azure AI Foundry
- Development & deployment of RAG-based chatbots and AI agents using Azure AI
- Programming : Python (Intermediate to Expert Level)
Job Title: DevOps Engineer
Location: Remote
Type: Full-time
About Us:
At Tese, we are committed to advancing sustainability through innovative technology solutions. Our platform empowers SMEs, financial institutions, and enterprises to achieve their Environmental, Social, and Governance (ESG) goals. We are looking for a skilled and passionate DevOps Engineer to join our team and help us build and maintain scalable, reliable, and efficient infrastructure.
Role Overview:
As a DevOps Engineer, you will be responsible for designing, implementing, and managing the infrastructure that supports our applications and services. You will work closely with our development, QA, and data science teams to ensure smooth deployment, continuous integration, and continuous delivery of our products. Your role will be critical in automating processes, enhancing system performance, and maintaining high availability.
Key Responsibilities:
- Infrastructure Management:
- Design, implement, and maintain scalable cloud infrastructure on platforms such as AWS, Google Cloud, or Azure.
- Manage server environments, including provisioning, monitoring, and maintenance.
- CI/CD Pipeline Development:
- Develop and maintain continuous integration and continuous deployment pipelines using tools like Jenkins, GitLab CI/CD, or CircleCI.
- Automate deployment processes to ensure quick and reliable releases.
- Configuration Management and Automation:
- Implement infrastructure as code (IaC) using tools like Terraform, Ansible, or CloudFormation.
- Automate system configurations and deployments to improve efficiency and reduce manual errors.
- Monitoring and Logging:
- Set up and manage monitoring tools (e.g., Prometheus, Grafana, ELK Stack) to track system performance and troubleshoot issues.
- Implement logging solutions to ensure effective incident response and system analysis.
- Security and Compliance:
- Ensure systems are secure and compliant with industry standards and regulations.
- Implement security best practices, including identity and access management, network security, and vulnerability assessments.
- Collaboration and Support:
- Work closely with development and QA teams to support application deployments and troubleshoot issues.
- Provide support for infrastructure-related inquiries and incidents.
Qualifications:
- Education:
- Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent practical experience.
- Experience:
- 3-5 years of experience in DevOps, system administration, or related roles.
- Hands-on experience with cloud platforms such as AWS, Google Cloud Platform, or Azure.
- Technical Skills:
- Proficiency in scripting languages like Bash, Python, or Ruby.
- Strong experience with containerization technologies like Docker and orchestration tools like Kubernetes.
- Knowledge of configuration management tools (Ansible, Puppet, Chef).
- Experience with CI/CD tools (Jenkins, GitLab CI/CD, CircleCI).
- Familiarity with monitoring and logging tools (Prometheus, Grafana, ELK Stack).
- Understanding of networking concepts and security best practices.
- Soft Skills:
- Strong problem-solving skills and attention to detail.
- Excellent communication and collaboration abilities.
- Ability to work in a fast-paced environment and manage multiple tasks.
Preferred Qualifications:
- Experience with infrastructure as code (IaC) tools like Terraform or CloudFormation.
- Knowledge of microservices architecture and serverless computing.
- Familiarity with database administration (SQL and NoSQL databases).
- Experience with Agile methodologies and working in a Scrum or Kanban environment.
- Passion for sustainability and interest in ESG initiatives.
Benefits:
- Competitive salary and benefits package,and performance bonuses.
- Flexible working hours and remote work options.
- Opportunity to work on impactful projects that promote sustainability.
- Professional development opportunities, including access to training and conferences.
About The Role:
The products/services of Eclat Engineering Pvt. Ltd. are being used by some of the leading institutions in India and abroad. Our services/Products are rapidly growing in demand. We are looking for a capable and dynamic Senior DevOps engineer to help setup, maintain and scale the infrastructure operations. This Individual will have the challenging responsibility of channelling our IT infrastructure and offering customer services with stringent international standard levels of service quality. This individual will leverage the latest IT tools to automate and streamline the delivery of our services while implementing industry-standard processes and knowledge management.
Roles & Responsibilities:
- Infrastructure and Deployment Automation: Design, implement, and maintain automation for infrastructure
provisioning and application deployment. Own the CI/CD pipelines and ensure they are efficient, reliable, and
scalable.
- System Monitoring and Performance: -Take ownership of monitoring systems and ensure the health and
performance of the infrastructure. Proactively identify and address performance bottlenecks and system issues.
- Cloud Infrastructure Management: Manage cloud infrastructure (e.g., AWS, Azure, GCP) and optimize resource
usage. Implement cost-saving measures while maintaining scalability and reliability.
- Configuration Management: Manage configuration management tools (e.g., Ansible, Puppet, Chef) to ensure
consistency across environments. Automate configuration changes and updates.
- Security and Compliance: Own security policies, implement best practices, and ensure compliance with industry
standards. Lead efforts to secure infrastructure and applications, including patch management and access controls.
- Collaboration with Development and Operations Teams: Foster collaboration between development and
operations teams, promoting a DevOps culture. Be the go-to person for resolving cross-functional infrastructure
issues and improving the development process.
- Disaster Recovery and Business Continuity: Develop and maintain disaster recovery plans and procedures. Ensure
business continuity in the event of system failures or other disruptions.
- Documentation and Knowledge Sharing: Create and maintain comprehensive documentation for configurations,
processes, and best practices. Share knowledge and mentor junior team members.
- Technical Leadership and Innovation: Stay up-to-date with industry trends and emerging technologies. Lead efforts
to introduce new tools and technologies that enhance DevOps practices.
- Problem Resolution and Troubleshooting: Be responsible for diagnosing and resolving complex issues related to
infrastructure and deployments. Implement preventive measures to reduce recurring problems.
Requirements:
● B.E / B.Tech / M.E / M.Tech / MCA / M.Sc.IT (if not should be able to demonstrate required skills)
● Overall 3+ years of experience in DevOps and Cloud operations specifically in AWS.
● Experience with Linux Administrator
● Experience with microservice architecture, containers, Kubernetes, and Helm is a must
● Experience in Configuration Management preferably Ansible
● Experience in Shell Scripting is a must
● Experience in developing and maintaining CI/CD processes using tools like Gitlab, Jenkins
● Experience in logging, monitoring and analytics
● An Understanding of writing Infrastructure as a Code using tools like Terraform
● Preferences - AWS, Kubernetes, Ansible
Must Have:
● Knowledge of AWS Cloud Platform.
● Good experience with microservice architecture, Kubernetes, helm and container-based technologies
● Hands-on experience with Ansible.
● Should have experience in working and maintaining CI/CD Processes.
● Hands-on experience in version control tools like GIT.
● Experience with monitoring tools such as Cloudwatch/Sysdig etc.
● Sound experience in administering Linux servers and Shell Scripting.
● Should have a good understanding of IT security and have the knowledge to secure production environments (OS and server software).
TVARIT GmbH develops and delivers solutions in the field of artificial intelligence (AI) for the Manufacturing, automotive, and process industries. With its software products, TVARIT makes it possible for its customers to make intelligent and well-founded decisions, e.g., in forward-looking Maintenance, increasing the OEE and predictive quality. We have renowned reference customers, competent technology, a good research team from renowned Universities, and the award of a renowned AI prize (e.g., EU Horizon 2020) which makes Tvarit one of the most innovative AI companies in Germany and Europe.
We are looking for a self-motivated person with a positive "can-do" attitude and excellent oral and written communication skills in English.
We are seeking a skilled and motivated Data Engineer from the manufacturing Industry with over two years of experience to join our team. As a data engineer, you will be responsible for designing, building, and maintaining the infrastructure required for the collection, storage, processing, and analysis of large and complex data sets. The ideal candidate will have a strong foundation in ETL pipelines and Python, with additional experience in Azure and Terraform being a plus. This role requires a proactive individual who can contribute to our data infrastructure and support our analytics and data science initiatives.
Skills Required
- Experience in the manufacturing industry (metal industry is a plus)
- 2+ years of experience as a Data Engineer
- Experience in data cleaning & structuring and data manipulation
- ETL Pipelines: Proven experience in designing, building, and maintaining ETL pipelines.
- Python: Strong proficiency in Python programming for data manipulation, transformation, and automation.
- Experience in SQL and data structures
- Knowledge in big data technologies such as Spark, Flink, Hadoop, Apache and NoSQL databases.
- Knowledge of cloud technologies (at least one) such as AWS, Azure, and Google Cloud Platform.
- Proficient in data management and data governance
- Strong analytical and problem-solving skills.
- Excellent communication and teamwork abilities.
Nice To Have
- Azure: Experience with Azure data services (e.g., Azure Data Factory, Azure Databricks, Azure SQL Database).
- Terraform: Knowledge of Terraform for infrastructure as code (IaC) to manage cloud.
· Support and fix issues in the IT infrastructure, end-user devices and applications.
· Manage tickets and provide helpdesk support.
· Perform desktop troubleshooting for end users.
· Aid with remote computer deployments.
· Install and maintain network infrastructure.
· Provide support for Azure/AWS/Google Cloud resources.
· Handle support escalations.
Develop and document SOPs to streamline support processes
TVARIT GmbH develops and delivers solutions in the field of artificial intelligence (AI) for the Manufacturing, automotive, and process industries. With its software products, TVARIT makes it possible for its customers to make intelligent and well-founded decisions, e.g., in forward-looking Maintenance, increasing the OEE and predictive quality. We have renowned reference customers, competent technology, a good research team from renowned Universities, and the award of a renowned AI prize (e.g., EU Horizon 2020) which makes TVARIT one of the most innovative AI companies in Germany and Europe.
We are looking for a self-motivated person with a positive "can-do" attitude and excellent oral and written communication skills in English.
We are seeking a skilled and motivated senior Data Engineer from the manufacturing Industry with over four years of experience to join our team. The Senior Data Engineer will oversee the department’s data infrastructure, including developing a data model, integrating large amounts of data from different systems, building & enhancing a data lake-house & subsequent analytics environment, and writing scripts to facilitate data analysis. The ideal candidate will have a strong foundation in ETL pipelines and Python, with additional experience in Azure and Terraform being a plus. This role requires a proactive individual who can contribute to our data infrastructure and support our analytics and data science initiatives.
Skills Required:
- Experience in the manufacturing industry (metal industry is a plus)
- 4+ years of experience as a Data Engineer
- Experience in data cleaning & structuring and data manipulation
- Architect and optimize complex data pipelines, leading the design and implementation of scalable data infrastructure, and ensuring data quality and reliability at scale
- ETL Pipelines: Proven experience in designing, building, and maintaining ETL pipelines.
- Python: Strong proficiency in Python programming for data manipulation, transformation, and automation.
- Experience in SQL and data structures
- Knowledge in big data technologies such as Spark, Flink, Hadoop, Apache, and NoSQL databases.
- Knowledge of cloud technologies (at least one) such as AWS, Azure, and Google Cloud Platform.
- Proficient in data management and data governance
- Strong analytical experience & skills that can extract actionable insights from raw data to help improve the business.
- Strong analytical and problem-solving skills.
- Excellent communication and teamwork abilities.
Nice To Have:
- Azure: Experience with Azure data services (e.g., Azure Data Factory, Azure Databricks, Azure SQL Database).
- Terraform: Knowledge of Terraform for infrastructure as code (IaC) to manage cloud.
- Bachelor’s degree in computer science, Information Technology, Engineering, or a related field from top-tier Indian Institutes of Information Technology (IIITs).
- Benefits And Perks
- A culture that fosters innovation, creativity, continuous learning, and resilience
- Progressive leave policy promoting work-life balance
- Mentorship opportunities with highly qualified internal resources and industry-driven programs
- Multicultural peer groups and supportive workplace policies
- Annual workcation program allowing you to work from various scenic locations
- Experience the unique environment of a dynamic start-up
Why should you join TVARIT ?
Working at TVARIT, a deep-tech German IT startup, offers a unique blend of innovation, collaboration, and growth opportunities. We seek individuals eager to adapt and thrive in a rapidly evolving environment.
If this opportunity excites you and aligns with your career aspirations, we encourage you to apply today!
Staff DevOps Engineer with Azure
EGNYTE YOUR CAREER. SPARK YOUR PASSION.
Egnyte is a place where we spark opportunities for amazing people. We believe that every role has meaning, and every Egnyter should be respected. With 22,000+ customers worldwide and growing, you can make an impact by protecting their valuable data. When joining Egnyte, you’re not just landing a new career, you become part of a team of Egnyters that are doers, thinkers, and collaborators who embrace and live by our values:
Invested Relationships
Fiscal Prudence
Candid Conversations
ABOUT EGNYTE
Egnyte is the secure multi-cloud platform for content security and governance that enables organizations to better protect and collaborate on their most valuable content. Established in 2008, Egnyte has democratized cloud content security for more than 22,000 organizations, helping customers improve data security, maintain compliance, prevent and detect ransomware threats, and boost employee productivity on any app, any cloud, anywhere. For more information, visit www.egnyte.com.
Our Production Engineering team enables Egnyte to provide customers access to their data 24/7 by providing best in class infrastructure.
ABOUT THE ROLE
We store multibillion files and multiple petabytes of data. We observe more than 11K API requests per second on average. To make that possible and to provide the best possible experience, we rely on great engineers. For us, people who own their work, from start to finish, are integral. Our engineers are part of the process from design to code, to test, to deployment and back again for further iterations. You can, and will, touch every level of the infrastructure depending on the day and what project you are working on. The ideal candidate should be able to take a complex problem and execute end to end. Mentor and set higher standards for the rest of the team and for the new hires.
WHAT YOU’LL DO:
• Design, build and maintain self-hosted and cloud environments to serve our own applications and services.
• Collaborate with software developers to build stable, scalable and high-performance solutions.
• Taking part in big projects like migrating solutions from self-hosted environments to the cloud, from virtual machines to Kubernetes, from monolith to microservices.
- Proactively make our organization and technology better!
- Advising others as to how DevOps can make a positive impact on their work.
• Share knowledge, mentor more junior team members while also still learning and gaining new skills.
- Maintain consistently high standards of communication, productivity, and teamwork across all teams.
YOUR QUALIFICATIONS:
• 5+ years of proven experience in a DevOps Engineer, System Administrator or Developer role, working on infrastructure or build processes.
• Expert knowledge of Microsoft Azure.
• Programming prowess (Python, Golang).
• Knowledge and experience about deployment and maintenance of Java and Python apps using application and web servers (Tomcat, Nginx, etc.).
• Ability to solve complex problems with simple, elegant and clean code.
• Practical knowledge of CI/CD solutions, GitLab CI or similar.
• Practical knowledge of Docker as a tool for testing and building an environment.
• Knowledge of Kubernetes and related technologies.
• Experience with metric-based monitoring solutions.
• Solid English skills to effectively communicate with other team members.
• Good understanding of the Linux Operating System on the administration level.
• Drive to grow as a DevOps Engineer (we value open-mindedness and a can-do attitude).
• Strong sense of ownership and ability to drive big projects.
BONUS SKILLS:
• Work experience as a Microsoft Azure architect.
• Experience in Cloud migrations projects.
• Leadership skills and experience.
COMMITMENT TO DIVERSITY, EQUITY, AND INCLUSION:
At Egnyte, we celebrate our differences and thrive on our diversity for our employees, our products, our customers, our investors, and our communities. Egnyters are encouraged to bring their whole selves to work and to appreciate the many differences that collectively make Egnyte a higher-performing company and a great place to be.
Responsibilities:
Develop and maintain high-quality, scalable, and efficient Java codebase for our ad-serving platform.
Collaborate with cross-functional teams including product managers, designers, and other developers to
understand requirements and translate them into technical solutions.
Design and implement new features and functionalities in the ad-serving system, focusing on performance
optimization and reliability.
Troubleshoot and debug complex issues in the ad server environment, providing timely resolutions to ensure
uninterrupted service.
Conduct code reviews, provide constructive feedback, and enforce coding best practices to maintain code quality
and consistency across the platform.
Stay updated with emerging technologies and industry trends in ad serving and digital advertising, and integrate
relevant innovations into our platform.
Work closely with DevOps and infrastructure teams to deploy and maintain the ad-serving platform in a cloud- based environment.
Collaborate with stakeholders to gather requirements, define technical specifications, and estimate development
efforts for new projects and features.
Mentor junior developers, sharing knowledge and best practices to foster a culture of continuous learning and
improvement within the development team.
Participate in on-call rotations and provide support for production issues as needed, ensuring maximum uptime
and reliability of the ad-serving platform.
• 5 to 10 years of work experience as a .net developer
• Proficiency in building web-based applications
• In-depth knowledge of the .net languages and database - asp.net web API, c# - asp.net MVC - SQL
• In-depth knowledge angularjs, service stack
• Familiarity with identity (3rd party)
• Proficiency using git version control.
• Familiarity with agile environments
• Troubleshooting abilities
• Good communication skills
• "Azure web app, azure storage "analyze system requirements and prioritize tasks
• Write clean, testable code using .net programming languages
• Develop technical specifications and architecture
• Test and debug various .net applications
• Review and refactor code
• Deploy fully functional applications
• Upgrade existing programs
• Document development and operational procedures
Job Description: DevOps Engineer
About Hyno:
Hyno Technologies is a unique blend of top-notch designers and world-class developers for new-age product development. Within the last 2 years we have collaborated with 32 young startups from India, US and EU to to find the optimum solution to their complex business problems. We have helped them to address the issues of scalability and optimisation through the use of technology and minimal cost. To us any new challenge is an opportunity.
As part of Hyno’s expansion plans,Hyno, in partnership with Sparity, is seeking an experienced DevOps Engineer to join our dynamic team. As a DevOps Engineer, you will play a crucial role in enhancing our software development processes, optimising system infrastructure, and ensuring the seamless deployment of applications. If you are passionate about leveraging cutting-edge technologies to drive efficiency, reliability, and scalability in software development, this is the perfect opportunity for you.
Position: DevOps Engineer
Experience: 5-7 years
Responsibilities:
- Collaborate with cross-functional teams to design, develop, and implement CI/CD pipelines for automated application deployment, testing, and monitoring.
- Manage and maintain cloud infrastructure using tools like AWS, Azure, or GCP, ensuring scalability, security, and high availability.
- Develop and implement infrastructure as code (IaC) using tools like Terraform or CloudFormation to automate the provisioning and management of resources.
- Constantly evaluate continuous integration and continuous deployment solutions as the industry evolves, and develop standardised best practices.
- Work closely with development teams to provide support and guidance in building applications with a focus on scalability, reliability, and security.
- Perform regular security assessments and implement best practices for securing the entire development and deployment pipeline.
- Troubleshoot and resolve issues related to infrastructure, deployment, and application performance in a timely manner.
- Follow regulatory and ISO 13485 requirements.
- Stay updated with industry trends and emerging technologies in the DevOps and cloud space, and proactively suggest improvements to current processes.
Requirements:
- Bachelor's degree in Computer Science, Engineering, or related field (or equivalent work experience).
- Minimum of 5 years of hands-on experience in DevOps, system administration, or related roles.
- Solid understanding of containerization technologies (Docker, Kubernetes) and orchestration tools
- Strong experience with cloud platforms such as AWS, Azure, or GCP, including services like ECS, S3, RDS, and more.
- Proficiency in at least one programming/scripting language such as Python, Bash, or PowerShell..
- Demonstrated experience in building and maintaining CI/CD pipelines using tools like Jenkins, GitLab CI/CD, or CircleCI.
- Familiarity with configuration management tools like Ansible, Puppet, or Chef.
- Experience with container (Docker, ECS, EKS), serverless (Lambda), and Virtual Machine (VMware, KVM) architectures.
- Experience with infrastructure as code (IaC) tools like Terraform, CloudFormation, or Pulumi.
- Strong knowledge of monitoring and logging tools such as Prometheus, ELK stack, or Splunk.
- Excellent problem-solving skills and the ability to work effectively in a fast-paced, collaborative environment.
- Strong communication skills and the ability to work independently as well as in a team.
Nice to Have:
- Relevant certifications such as AWS Certified DevOps Engineer, Azure DevOps Engineer, Certified Kubernetes Administrator (CKA), etc.
- Experience with microservices architecture and serverless computing.
Soft Skills:
- Excellent written and verbal communication skills.
- Ability to manage conflict effectively.
- Ability to adapt and be productive in a dynamic environment.
- Strong communication and collaboration skills supporting multiple stakeholders and business operations.
- Self-starter, self-managed, and a team player.
Join us in shaping the future of DevOps at Hyno in collaboration with Sparity. If you are a highly motivated and skilled DevOps Engineer, eager to make an impact in a remote setting, we'd love to hear from you.
Now, more than ever, the Toast team is committed to our customers. We’re taking steps to help restaurants navigate these unprecedented times with technology, resources, and community. Our focus is on building a restaurant platform that helps restaurants adapt, take control, and get back to what they do best: building the businesses they love. And because our technology is purpose-built for restaurants by restaurant people, restaurants can trust that we’ll deliver on their needs for today while investing in experiences that will power their restaurant of the future.
At Toast, our Site Reliability Engineers (SREs) are responsible for keeping all customer-facing services and other Toast production systems running smoothly. SREs are a blend of pragmatic operators and software craftspeople who apply sound software engineering principles, operational discipline, and mature automation to our environments and our codebase. Our decisions are based on instrumentation and continuous observability, as well as predictions and capacity planning.
About this roll* (Responsibilities)
- Gather and analyze metrics from both operating systems and applications to assist in performance tuning and fault finding
- Partner with development teams to improve services through rigorous testing and release procedures
- Participate in system design consulting, platform management, and capacity planning
- Create sustainable systems and services through automation and uplift
- Balance feature development speed and reliability with well-defined service level objectives
Troubleshooting and Supporting Escalations:
- Gather and analyze metrics from both operating systems and applications to assist in performance tuning and fault finding
- Diagnose performance bottlenecks and implement optimizations across infrastructure, databases, web, and mobile applications
- Implement strategies to increase system reliability and performance through on-call rotation and process optimization
- Perform and run blameless RCAs on incidents and outages aggressively, looking for answers that will prevent the incident from ever happening again
Do you have the right ingredients? (Requirements)
- Extensive industry experience with at least 7+ years in SRE and/or DevOps roles
- Polyglot technologist/generalist with a thirst for learning
- Deep understanding of cloud and microservice architecture and the JVM
- Experience with tools such as APM, Terraform, Ansible, GitHub, Jenkins, and Docker
- Experience developing software or software projects in at least four languages, ideally including two of Go, Python, and Java
- Experience with cloud computing technologies ( AWS cloud provider preferred)
Bread puns are encouraged but not required
Hiring for Azure Data Engineers.
Location: Bangalore
Employment type: Full-time, permanent
website: www.amazech.com
Qualifications:
B.E./B.Tech/M.E./M.Tech in Computer Science, Information Technology, Electrical or Electronic with good academic background.
Experience and Required Skill Sets:
• Minimum 5 years of hands-on experience with Azure Data Lake, Azure Data Factory, SQL Data Warehouse, Azure Blob, Azure Storage Explorer
• Experience in Data warehouse/analytical systems using Azure Synapse.
Proficient in creating Azure Data Factory pipelines for ETL processing; copy activity, custom Azure development, Synapse, etc.
• Knowledge of Azure Data Catalog, Event Grid, Service Bus, SQL, and Purview.
• Good technical knowledge in Microsoft SQL Server BI Suite (ETL, Reporting, Analytics, Dashboards) using SSIS, SSAS, SSRS, Power BI
• Design and develop batch and real-time streaming of data loads to data warehouse systems
Other Requirements:
A Bachelor's or Master's degree (Engineering or computer-related degree preferred)
Strong understanding of Software Development Life Cycles including Agile/Scrum
Responsibilities:
• Ability to create complex, enterprise-transforming applications that meet and exceed client expectations.
• Responsible for the bottom line. Strong project management abilities. Ability to encourage the team to stick to timelines.
- Candidate should be able to write the sample programs using the Tools (Bash, PowerShell, Python or Shell scripting)
- Analytical/logical reasoning
- GitHub Actions
- Should have good working experience with GitHub Actions
- Repository/Workflow Dispatch, writing reusable workflows, etc
- AZ CLI commands
- Hands-on experience with AZ CLI commands






















