50+ Google Cloud Platform (GCP) Jobs in India
Apply to 50+ Google Cloud Platform (GCP) Jobs on CutShort.io. Find your next job, effortlessly. Browse Google Cloud Platform (GCP) Jobs and apply today!


Salary (Lacs): Up to 22 LPA
Required Qualifications
• 4–7 years of total experience, with a minimum of 4 years in a full-time DevOps role
• Hands-on experience with major cloud platforms (GCP, AWS, Azure, OCI), more than one will be a plus
• Proficient in Kubernetes administration and container technologies (Docker, containerd)
• Strong Linux fundamentals
• Scripting skills in Python and shell scripting
• Knowledge of infrastructure as code with hands-on experience in Terraform and/or Pulumi (mandatory)
• Experience in maintaining and troubleshooting production environments
• Solid understanding of CI/CD concepts with hands-on experience in tools like Jenkins, GitLab CI, GitHub Actions, ArgoCD, Devtron, GCP Cloud Build, or Bitbucket Pipelines

Job Title : Backend / API Developer - Python (FastAPI) / Node.js (Express)
Location : Remote
Experience : 4+ Years
Job Description :
We are looking for a skilled Backend / API Developer - Python (FastAPI) / Node.js (Express) with strong expertise in building secure, scalable, and reliable backend systems. The ideal candidate should be proficient in Python (FastAPI preferred) or Node.js (Express) and have hands-on experience deploying applications to serverless environments.
Key Responsibilities :
- Design, develop, and maintain RESTful APIs and backend services.
- Deploy and manage serverless applications on Cloudflare Workers, Firebase Functions, and Google Cloud Functions.
- Work with Google Cloud services including Cloud Run, Cloud Functions, Secret Manager, and IAM roles.
- Implement secure API development practices (HTTPS, input validation, and secrets management).
- Ensure performance optimization, scalability, and reliability of backend systems.
- Collaborate with front-end developers, DevOps, and product teams to deliver high-quality solutions.
Mandatory Skills :
Python (FastAPI) / Node.js (Express), Serverless Deployment (Cloudflare Workers, Firebase, GCP Functions), Google Cloud Services (Cloud Run, IAM, Secret Manager), API Security (HTTPS, Input Validation, Secrets Management).
Required Skills :
- Proficiency in Python (FastAPI preferred) or Node.js (Express).
- Hands-on experience with serverless platforms (Cloudflare Workers, Firebase Functions, GCP Functions).
- Familiarity with Google Cloud services (Cloud Run, IAM, Secret Manager, Cloud Functions).
- Strong understanding of secure API development (HTTPS, input validation, API keys & secret management).
- Knowledge of API design principles and best practices.
- Ability to work with CI/CD pipelines and modern development workflows.
Preferred Qualifications :
- Strong knowledge of microservices architecture.
- Experience with CI/CD pipelines.
- Knowledge of containerization (Docker, Kubernetes).
- Familiarity with monitoring and logging tools.
escription
Job Summary:
Join Springer Capital as a Cybersecurity & Cloud Intern to help architect, secure, and automate our cloud-based backend systems powering next-generation investment platforms.
Job Description:
Founded in 2015, Springer Capital is a technology-forward asset management and investment firm. We leverage cutting-edge digital solutions to uncover high-potential opportunities, transforming traditional finance through innovation, agility, and a relentless commitment to security and scalability.
Job Highlights
Work hands-on with AWS, Azure, or GCP to design and deploy secure, scalable backend infrastructure.
Collaborate with DevOps and engineering teams to embed security best practices in CI/CD pipelines.
Gain experience in real-world incident response, vulnerability assessment, and automated monitoring.
Drive meaningful impact on our security posture and cloud strategy from Day 1.
Enjoy a fully remote, flexible internship with global teammates.
Responsibilities
Assist in architecting and provisioning cloud resources (VMs, containers, serverless functions) with strict security controls.
Implement identity and access management, network segmentation, encryption, and logging best practices.
Develop and maintain automation scripts for security monitoring, patch management, and incident alerts.
Support vulnerability scanning, penetration testing, and remediation tracking.
Document cloud architectures, security configurations, and incident response procedures.
Partner with backend developers to ensure secure API gateways, databases, and storage services.
What We Offer
Mentorship: Learn directly from senior security engineers and cloud architects.
Training & Certifications: Access to online courses and support for AWS/Azure security certifications.
Impactful Projects: Contribute to critical security and cloud initiatives that safeguard our digital assets.
Remote-First Culture: Flexible hours and the freedom to collaborate from anywhere.
Career Growth: Build a strong foundation for a future in cybersecurity, cloud engineering, or DevSecOps.
Requirements
Pursuing or recently graduated in Computer Science, Cybersecurity, Information Technology, or a related discipline.
Familiarity with at least one major cloud platform (AWS, Azure, or GCP).
Understanding of core security principles: IAM, network security, encryption, and logging.
Scripting experience in Python, PowerShell, or Bash for automation tasks.
Strong analytical, problem-solving, and communication skills.
A proactive learner mindset and passion for securing cloud environments.
About Springer Capital
Springer Capital blends financial expertise with digital innovation to redefine asset management. Our mission is to drive exceptional value by implementing robust, technology-driven strategies that transform investment landscapes. We champion a culture of creativity, collaboration, and continuous improvement.
Location: Global (Remote)
Job Type: Internship
Pay: $50 USD per month
Work Location: Remote
Job Title: Cloud Engineer (Bench Role – Staff Augmentation)
Location: Remote
Engagement Type: Contract / Project-based / Full-time (depending on client)
Experience Range: 2–7 years
About the Role
We are building a strong remote bench of Cloud Engineers to design, deploy, and maintain cloud-native and hybrid infrastructure for SaaS, FinTech, and product-based startups across US, Canada, and Middle East.
This role focuses on scalable infrastructure, security, and performance with flexibility, clear deliverables, and competitive monthly retainers.
Core Skills Required
Cloud Platforms
- AWS (EC2, S3, RDS, Lambda, API Gateway, ECS/EKS)
- Azure (VMs, Blob Storage, Azure SQL, Functions, App Service)
- GCP (Compute Engine, Cloud Storage, BigQuery, Cloud Functions)
- Multi-cloud experience preferred
Infrastructure as Code (IaC)
- Terraform, AWS CloudFormation, Azure ARM Templates, or Pulumi
- Automated provisioning & environment management
Networking & Security
- VPC, subnets, routing, VPN setup, load balancers, firewalls
- CDN & DNS management (CloudFront, Azure CDN, Cloudflare)
- IAM policies, identity federation, MFA enforcement
- Encryption & compliance (SOC2, HIPAA, ISO 27001)
Storage & Databases
- Relational: RDS, Azure SQL, Cloud SQL
- NoSQL: DynamoDB, Cosmos DB, Firestore
- Data warehousing: BigQuery, Redshift
Monitoring & Optimization
- CloudWatch, Azure Monitor, GCP Operations Suite
- Datadog, Prometheus, Splunk (preferred)
- Cost monitoring & optimization (rightsizing, Reserved Instances)
Automation & Deployment
- Cloud CLI/SDK scripting (Bash, Python, PowerShell)
- Auto-scaling configuration, backup & disaster recovery strategies
- Blue-green & canary deployment strategies
Ideal Candidate Profile
- Experienced in designing secure, cost-effective, and scalable cloud solutions
- Strong problem-solving skills with focus on high availability
- Comfortable working with cross-functional teams in distributed environments
Perks for Bench Developers
- Access to multi-cloud, enterprise-level projects across US, Canada, and Middle East
- Competitive retainers (based on skills & experience)
- Flexible, remote-first work with clear deliverables
Apply If You Have
- Delivered production-ready cloud infrastructure
- Migrated workloads from on-premise to cloud or between cloud providers
- Worked with DevOps teams to integrate CI/CD into cloud environments
About HelloRamp.ai
HelloRamp is on a mission to revolutionize media creation for automotive and retail using AI. Our platform powers 3D/AR experiences for leading brands like Cars24, Spinny, and Samsung. We’re now building the next generation of Computer Vision + AI products, including cutting-edge NeRF pipelines and AI-driven video generation.
What You’ll Work On
- Develop and optimize Computer Vision pipelines for large-scale media creation.
- Implement NeRF-based systems for high-quality 3D reconstruction.
- Build and fine-tune AI video generation models using state-of-the-art techniques.
- Optimize AI inference for production (CUDA, TensorRT, ONNX).
- Collaborate with the engineering team to integrate AI features into scalable cloud systems.
- Research latest AI/CV advancements and bring them into production.
Skills & Experience
- Strong Python programming skills.
- Deep expertise in Computer Vision and Machine Learning.
- Hands-on with PyTorch/TensorFlow.
- Experience with NeRF frameworks (Instant-NGP, Nerfstudio, Plenoxels) and/or video synthesis models.
- Familiarity with 3D graphics concepts (meshes, point clouds, depth maps).
- GPU programming and optimization skills.
Nice to Have
- Knowledge of Three.js or WebGL for rendering AI outputs on the web.
- Familiarity with FFmpeg and video processing pipelines.
- Experience in cloud-based GPU environments (AWS/GCP).
Why Join Us?
- Work on cutting-edge AI and Computer Vision projects with global impact.
- Join a small, high-ownership team where your work matters.
- Opportunity to experiment, publish, and contribute to open-source.
- Competitive pay and flexible work setup.
Advanced SQL, data modeling skills - designing Dimensional Layer, 3NF, denormalized views & semantic layer, Expertise in GCP services
Role & Responsibilities:
● Design and implement robust semantic layers for data systems on Google Cloud Platform (GCP)
● Develop and maintain complex data models, including dimensional models, 3NF structures, and denormalized views
● Write and optimize advanced SQL queries for data extraction, transformation, and analysis
● Utilize GCP services to create scalable and efficient data architectures
● Collaborate with cross-functional teams to translate business requirements(specified in mapping sheets or Legacy
Datastage jobs) into effective data models
● Implement and maintain data warehouses and data lakes on GCP
● Design and optimize ETL/ELT processes for large-scale data integration
● Ensure data quality, consistency, and integrity across all data models and semantic layers
● Develop and maintain documentation for data models, semantic layers, and data flows
● Participate in code reviews and implement best practices for data modeling and database design
● Optimize database performance and query execution on GCP
● Provide technical guidance and mentorship to junior team members
● Stay updated with the latest trends and advancements in data modeling, GCP services, and big data technologies
● Collaborate with data scientists and analysts to enable efficient data access and analysis
● Implement data governance and security measures within the semantic layer and data model

Job Type : Contract
Location : Bangalore
Experience : 5+yrs
The role focuses on cloud security engineering with a strong emphasis on GCP, while also covering AWS and Azure.
Required Skills:
- 5+ years of experience in software and/or cloud platform engineering, particularly focused on GCP environment.
- Knowledge of the Shared Responsibility Model; keen understanding of the security risks inherent in hosting cloud-based applications and data.
- Experience developing across the security assurance lifecycle (including prevent, detect, respond, and remediate controls)?Experience in configuring Public Cloud native security tooling and capabilities with a focus on Google Cloud Organizational policies/constraints, VPC SC, IAM policies and GCP APIs.
- Experience with Cloud Security Posture Management (CSPM) 3rd Party tools such as Wiz, Prisma, Check Point CloudGuard, etc.
- Experience in Policy-as-code (Rego) and OPA platform.
- Experience solutioning and configuring event-driven serverless-based security controls in Azure, including but not limited to technologies such as Azure Function, Automation Runbook, AWS Lambda and Google Cloud Functions.
- Deep understanding of DevOps processes and workflows.
- Working knowledge of the Secure SDLC process
- Experience with Infrastructure as Code (IaC) tooling, preferably Terraform.
- Familiarity with Logging and data pipeline concepts and architectures in cloud.
- Strong in scripting languages such as PowerShell or Python or Bash or Go.
- Knowledge of Agile best practices and methodologies
- Experience creating technical architecture documentation.? Excellent communication, written and interpersonal skills.
- Practical experience in designing and configuring CICD pipelines. Practical experience in GitHub Actions and Jenkins.
- Experience in ITSM.
- Ability to articulate complex technical concepts to non-technical stakeholders.
- Experience with risk control frameworks and engagements with risk and regulatory functions
- Experience in the financial industry would be a plus.

About Us
CLOUDSUFI, a Google Cloud Premier Partner, a Data Science and Product Engineering organization building Products and Solutions for Technology and Enterprise industries. We firmly believe in the power of data to transform businesses and make better decisions. We combine unmatched experience in business processes with cutting edge infrastructure and cloud services. We partner with our customers to monetize their data and make enterprise data dance.
Our Values
We are a passionate and empathetic team that prioritizes human values. Our purpose is to elevate the quality of lives for our family, customers, partners and the community.
Equal Opportunity Statement
CLOUDSUFI is an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. All qualified candidates receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, and national origin status. We provide equal opportunities in employment, advancement, and all other areas of our workplace. Please explore more at https://www.cloudsufi.com/.
Role Overview:
As a Senior Data Scientist / AI Engineer, you will be a key player in our technical leadership. You will be responsible for designing, developing, and deploying sophisticated AI and Machine Learning solutions, with a strong emphasis on Generative AI and Large Language Models (LLMs). You will architect and manage scalable AI microservices, drive research into state-of-the-art techniques, and translate complex business requirements into tangible, high-impact products. This role requires a blend of deep technical expertise, strategic thinking, and leadership.
Key Responsibilities:
- Architect & Develop AI Solutions: Design, build, and deploy robust and scalable machine learning models, with a primary focus on Natural Language Processing (NLP), Generative AI, and LLM-based Agents.
- Build AI Infrastructure: Create and manage AI-driven microservices using frameworks like Python FastAPI, ensuring high performance and reliability.
- Lead AI Research & Innovation: Stay abreast of the latest advancements in AI/ML. Lead research initiatives to evaluate and implement state-of-the-art models and techniques for performance and cost optimization.
- Solve Business Problems: Collaborate with product and business teams to understand challenges and develop data-driven solutions that create significant business value, such as building business rule engines or predictive classification systems.
- End-to-End Project Ownership: Take ownership of the entire lifecycle of AI projects—from ideation, data processing, and model development to deployment, monitoring, and iteration on cloud platforms.
- Team Leadership & Mentorship: Lead learning initiatives within the engineering team, mentor junior data scientists and engineers, and establish best practices for AI development.
- Cross-Functional Collaboration: Work closely with software engineers to integrate AI models into production systems and contribute to the overall system architecture.
Required Skills and Qualifications
- Master’s (M.Tech.) or Bachelor's (B.Tech.) degree in Computer Science, Artificial Intelligence, Information Technology, or a related field.
- 6+ years of professional experience in a Data Scientist, AI Engineer, or related role.
- Expert-level proficiency in Python and its core data science libraries (e.g., PyTorch, Huggingface Transformers, Pandas, Scikit-learn).
- Demonstrable, hands-on experience building and fine-tuning Large Language Models (LLMs) and implementing Generative AI solutions.
- Proven experience in developing and deploying scalable systems on cloud platforms, particularly AWS. Experience with GCS is a plus.
- Strong background in Natural Language Processing (NLP), including experience with multilingual models and transcription.
- Experience with containerization technologies, specifically Docker.
- Solid understanding of software engineering principles and experience building APIs and microservices.
Preferred Qualifications
- A strong portfolio of projects. A track record of publications in reputable AI/ML conferences is a plus.
- Experience with full-stack development (Node.js, Next.js) and various database technologies (SQL, MongoDB, Elasticsearch).
- Familiarity with setting up and managing CI/CD pipelines (e.g., Jenkins).
- Proven ability to lead technical teams and mentor other engineers.
- Experience developing custom tools or packages for data science workflows.


Job Title: Software Engineer Consultant/Expert 34192
Location: Chennai
Work Type: Onsite
Notice Period: Immediate Joiners only or serving candidates upto 30 days.
Position Description:
- Candidate with strong Python experience.
- Full Stack Development in GCP End to End Deployment/ ML Ops Software Engineer with hands-on n both front end, back end and ML Ops
- This is a Tech Anchor role.
Experience Required:
- 7 Plus Years
Springer Capital is a cross-border asset management firm specializing in real estate investment banking between China and the USA. We are offering a remote internship for aspiring data engineers interested in data pipeline development, data integration, and business intelligence. The internship offers flexible start and end dates. A short quiz or technical task may be required as part of the selection process.
Responsibilities:
▪ Design, build, and maintain scalable data pipelines for structured and unstructured data sources
▪ Develop ETL processes to collect, clean, and transform data from internal and external systems
▪ Support integration of data into dashboards, analytics tools, and reporting systems
▪ Collaborate with data analysts and software developers to improve data accessibility and performance
▪ Document workflows and maintain data infrastructure best practices
▪ Assist in identifying opportunities to automate repetitive data tasks
About Us
CLOUDSUFI, a Google Cloud Premier Partner, is a global leading provider of data-driven digital transformation across cloud-based enterprises. With a global presence and focus on Software & Platforms, Life sciences and Healthcare, Retail, CPG, financial services, and supply chain, CLOUDSUFI is positioned to meet customers where they are in their data monetization journey.
Our Values
We are a passionate and empathetic team that prioritizes human values. Our purpose is to elevate the quality of lives for our family, customers, partners and the community.
Equal Opportunity Statement
CLOUDSUFI is an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. All qualified candidates receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, and national origin status. We provide equal opportunities in employment, advancement, and all other areas of our workplace. Please explore more at https://www.cloudsufi.com/
What are we looking for
We are seeking a highly skilled and experienced Senior DevOps Engineer to join our team. The ideal candidate will have extensive expertise in modern DevOps tools and practices, particularly in managing CI/CD pipelines, infrastructure as code, and cloud-native environments. This role involves designing, implementing, and maintaining robust, scalable, and efficient infrastructure and deployment pipelines to support our development and operations teams.
Required Skills and Experience:
- 7+ years of experience in DevOps, infrastructure automation, or related fields.
- Advanced expertise in Terraform for infrastructure as code.
- Solid experience with Helm for managing Kubernetes applications.
- Proficient with GitHub for version control, repository management, and workflows.
- Extensive experience with Kubernetes for container orchestration and management.
- In-depth understanding of Google Cloud Platform (GCP) services and architecture.
- Strong scripting and automation skills (e.g., Python, Bash, or equivalent).
- Excellent problem-solving skills and attention to detail. - Strong communication and collaboration abilities in agile development environments.
Preferred Qualifications:
- Experience with other CI/CD tools (e.g., Jenkins, GitLab CI/CD).
- Knowledge of additional cloud platforms (e.g., AWS, Azure).
- Certification in Kubernetes (CKA/CKAD) or Google Cloud (GCP Professional DevOps Engineer).
Behavioral Competencies
• Must have worked with US/Europe based clients in onsite/offshore delivery models.
• Should have very good verbal and written communication, technical articulation, listening and presentation skills.
• Should have proven analytical and problem solving skills.
• Should have collaborative mindset for cross-functional team work
• Passion for solving complex search problems
• Should have demonstrated effective task prioritization, time management and internal/external stakeholder management skills.
• Should be a quick learner, self starter, go-getter and team player.
• Should have experience of working under stringent deadlines in a Matrix organization structure.

Location: Hybrid/ Remote
Openings: 2
Experience: 5–12 Years
Qualification: Bachelor’s or Master’s in Computer Science or a related technical field
Key Responsibilities
Architect & Design:
- Provide technical and architectural direction for complex frontend solutions,ensuring alignment with enterprise standards and best practices.
- Conduct design and code reviews to maintain high-quality, reusable, and scalable frontend interfaces for enterprise applications.
- Collaborate with cross-functional teams to define and enforce UI/UX design guidelines, accessibility standards, and performance benchmarks.
- Identify and address potential security vulnerabilities in frontend implementations, ensuring compliance with security and data privacy requirements.
Development & Debugging:
- Write clean, maintainable, and efficient frontend code.
- Debug and troubleshoot code to ensure robust, high-performing applications.
- Develop reusable frontend libraries that can be leveraged across multiple projects.
AI Awareness (Preferred):
- Understand AI/ML fundamentals and how they can enhance frontend applications.
- Collaborate with teams integrating AI-based features into chat applications.
Collaboration & Reporting:
- Work closely with cross-functional teams to align on architecture and deliverables.
- Regularly report progress, identify risks, and propose mitigation strategies.
Quality Assurance:
- Implement unit tests and end-to-end tests to ensure code quality.
- Participate in code reviews and enforce best practices.
Required Skills
- 5-10 years of experience architecting and developing cloud-based global applications in a public cloud environment (AWS, Azure, or GCP).
- Strong hands-on expertise in frontend technologies: JavaScript, HTML5, CSS3
- Proficiency with Modern frameworks like React, Angular, or Node.js
- Backend familiarity with Java, Spring Boot (or similar technologies).
- Experience developing real-world, at-scale products.
- General knowledge of cloud platforms (AWS, Azure, or GCP) and their structure, use, and capabilities.
- Strong problem-solving, debugging, and performance optimization skills.

Location: Hybrid/ Remote
Openings: 2
Experience: 5+ Years
Qualification: Bachelor’s or Master’s in Computer Science or related field
Job Responsibilities
Problem Solving & Optimization:
- Analyze and resolve complex technical and application issues.
- Optimize application performance, scalability, and reliability.
Design & Develop:
- Build, test, and deploy scalable full-stack applications with high performance and security.
- Develop clean, reusable, and maintainable code for both frontend and backend.
AI Integration (Preferred):
- Collaborate with the team to integrate AI/ML models into applications where applicable.
- Explore Generative AI, NLP, or machine learning solutions that enhance product capabilities.
Technical Leadership & Mentorship:
- Provide guidance, mentorship, and code reviews for junior developers.
- Foster a culture of technical excellence and knowledge sharing.
Agile & Delivery Management:
- Participate in Agile ceremonies (sprint planning, stand-ups, retrospectives).
- Define and scope backlog items, track progress, and ensure timely delivery.
Collaboration:
- Work closely with cross-functional teams (product managers, designers, QA) to deliver high-quality solutions.
- Coordinate with geographically distributed teams.
Quality Assurance & Security:
- Conduct peer reviews of designs and code to ensure best practices.
- Implement security measures and ensure compliance with industry standards.
Innovation & Continuous Improvement:
- Identify areas for improvement in the software development lifecycle.
- Stay updated with the latest tech trends, especially in AI and cloud technologies, and recommend new tools or frameworks.
Required Skills
- Strong proficiency in JavaScript, HTML5, CSS3
- Hands-on expertise with frontend frameworks like React, Angular, or Vue.js
- Backend development experience with Java, Spring Boot (Node.js is a plus)
- Knowledge of REST APIs, microservices, and scalable architectures
- Familiarity with cloud platforms (AWS, Azure, or GCP)
- Experience with Agile/Scrum methodologies and JIRA for project tracking
- Proficiency in Git and version control best practices
- Strong debugging, performance optimization, and problem-solving skills
- Ability to analyze customer requirements and translate them into technical specifications

Location: Hybrid/ Remote
Openings: 5
Experience: 0 - 2Years
Qualification: Bachelor’s or Master’s in Computer Science or a related technical field
Key Responsibilities:
Backend Development & APIs
- Build microservices that provide REST APIs to power web frontends.
- Design clean, reusable, and scalable backend code meeting enterprise security standards.
- Conceptualize and implement optimized data storage solutions for high-performance systems.
Deployment & Cloud
- Deploy microservices using a common deployment framework on AWS and GCP.
- Inspect and optimize server code for speed, security, and scalability.
Frontend Integration
- Work on modern front-end frameworks to ensure seamless integration with back-end services.
- Develop reusable libraries for both frontend and backend codebases.
AI Awareness (Preferred)
- Understand how AI/ML or Generative AI can enhance enterprise software workflows.
- Collaborate with AI specialists to integrate AI-driven features where applicable.
Quality & Collaboration
- Participate in code reviews to maintain high code quality.
- Collaborate with teams using Agile/Scrum methodologies for rapid and structured delivery.
Required Skills:
- Proficiency in JavaScript (ES6+), Webpack, Mocha, Jest
- Experience with recent frontend frameworks – React.js, Redux.js, Node.js (or similar)
- Deep understanding of HTML5, CSS3, SASS/LESS, and Content Management Systems
- Ability to design and implement RESTful APIs and understand their impact on client-side applications
- Familiarity with cloud platforms (AWS, Azure, or GCP) – deployment, storage, and scalability
- Experience working with Agile and Scrum methodologies
- Strong backend expertise in Java, J2EE, Spring Boot is a plus but not mandatory

Ontrac Solutions is a leading technology consulting firm specializing in cutting-edge solutions that drive business transformation. We partner with organizations to modernize their infrastructure, streamline processes, and deliver tangible results.
Our client is actively seeking a Conversational AI Engineer with deep hands-on experience in Google Contact Center AI (CCAI) to join a high-impact digital transformation project via a GCP Premier Partner. As part of a staff augmentation model, you will be embedded within the client’s technology or contact center innovation team, delivering scalable virtual agent solutions that improve customer experience, agent productivity, and call deflection.
Key Responsibilities:
- Lead the design and buildout of Dialogflow CX/ES agents across chat and voice channels.
- Integrate virtual agents with client systems and platforms (e.g., Genesys, Twilio, NICE CXone, Salesforce).
- Develop fulfillment logic using Google Cloud Functions, Cloud Run, and backend integrations (via REST APIs and webhooks).
- Collaborate with stakeholders to define intent models, entity schemas, and user flows.
- Implement Agent Assist and CCAI Insights to augment live agent productivity.
- Leverage Google Cloud tools including Pub/Sub, Logging, and BigQuery to support and monitor the solution.
- Support tuning, regression testing, and enhancement of NLP performance using live utterance data.
- Ensure adherence to enterprise security and compliance requirements.
Required Skills & Qualifications:
- 3+ years developing conversational AI experiences, including at least 1–2 years with Google Dialogflow CX or ES.
- Solid experience across GCP services (Functions, Cloud Run, IAM, BigQuery, etc.).
- Strong skills in Python or Node.js for webhook fulfillment and backend logic.
- Knowledge of NLU/NLP modeling, training, and performance tuning.
- Prior experience working in client-facing or embedded roles via consulting or staff augmentation.
- Ability to communicate effectively with technical and business stakeholders.
Nice to Have:
- Hands-on experience with Agent Assist, CCAI Insights, or third-party CCaaS tools (Genesys, Twilio Flex, NICE CXone).
- Familiarity with Vertex AI, AutoML, or other GCP ML services.
- Experience in regulated industries (healthcare, finance, insurance, etc.).
- Google Cloud certification in Cloud Developer or CCAI Engineer.
We are looking for a highly skilled DevOps/Cloud Engineer with over 6 years of experience in infrastructure automation, cloud platforms, networking, and security. If you are passionate about designing scalable systems and love solving complex cloud and DevOps challenges—this opportunity is for you.
Key Responsibilities
- Design, deploy, and manage cloud-native infrastructure using Kubernetes (K8s), Helm, Terraform, and Ansible
- Automate provisioning and orchestration workflows for cloud and hybrid environments
- Manage and optimize deployments on AWS, Azure, and GCP for high availability and cost efficiency
- Troubleshoot and implement advanced network architectures including VPNs, firewalls, load balancers, and routing protocols
- Implement and enforce security best practices: IAM, encryption, compliance, and vulnerability management
- Collaborate with development and operations teams to improve CI/CD workflows and system observability
Required Skills & Qualifications
- 6+ years of experience in DevOps, Infrastructure as Code (IaC), and cloud-native systems
- Expertise in Helm, Terraform, and Kubernetes
- Strong hands-on experience with AWS and Azure
- Solid understanding of networking, firewall configurations, and security protocols
- Experience with CI/CD tools like Jenkins, GitHub Actions, or similar
- Strong problem-solving skills and a performance-first mindset
Why Join Us?
- Work on cutting-edge cloud infrastructure across diverse industries
- Be part of a collaborative, forward-thinking team
- Flexible hybrid work model – work from anywhere while staying connected
- Opportunity to take ownership and lead critical DevOps initiatives
Job Title : Data Engineer – GCP + Spark + DBT
Location : Bengaluru (On-site at Client Location | 3 Days WFO)
Experience : 8 to 12 Years
Level : Associate Architect
Type : Full-time
Job Overview :
We are looking for a seasoned Data Engineer to join the Data Platform Engineering team supporting a Unified Data Platform (UDP). This role requires hands-on expertise in DBT, GCP, BigQuery, and PySpark, with a solid foundation in CI/CD, data pipeline optimization, and agile delivery.
Mandatory Skills : GCP, DBT, Google Dataform, BigQuery, PySpark/Spark SQL, Advanced SQL, CI/CD, Git, Agile Methodologies.
Key Responsibilities :
- Design, build, and optimize scalable data pipelines using BigQuery, DBT, and PySpark.
- Leverage GCP-native services like Cloud Storage, Pub/Sub, Dataproc, Cloud Functions, and Composer for ETL/ELT workflows.
- Implement and maintain CI/CD for data engineering projects with Git-based version control.
- Collaborate with cross-functional teams including Infra, Security, and DataOps for reliable, secure, and high-quality data delivery.
- Lead code reviews, mentor junior engineers, and enforce best practices in data engineering.
- Participate in Agile sprints, backlog grooming, and Jira-based project tracking.
Must-Have Skills :
- Strong experience with DBT, Google Dataform, and BigQuery
- Hands-on expertise with PySpark/Spark SQL
- Proficient in GCP for data engineering workflows
- Solid knowledge of SQL optimization, Git, and CI/CD pipelines
- Agile team experience and strong problem-solving abilities
Nice-to-Have Skills :
- Familiarity with Databricks, Delta Lake, or Kafka
- Exposure to data observability and quality frameworks (e.g., Great Expectations, Soda)
- Knowledge of MDM patterns, Terraform, or IaC is a plus


Job Title : AI Architect
Location : Pune (On-site | 3 Days WFO)
Experience : 6+ Years
Shift : US or flexible shifts
Job Summary :
We are looking for an experienced AI Architect to design and deploy AI/ML solutions that align with business goals.
The role involves leading end-to-end architecture, model development, deployment, and integration using modern AI/ML tools and cloud platforms (AWS/Azure/GCP).
Key Responsibilities :
- Define AI strategy and identify business use cases
- Design scalable AI/ML architectures
- Collaborate on data preparation, model development & deployment
- Ensure data quality, governance, and ethical AI practices
- Integrate AI into existing systems and monitor performance
Must-Have Skills :
- Machine Learning, Deep Learning, NLP, Computer Vision
- Data Engineering, Model Deployment (CI/CD, MLOps)
- Python Programming, Cloud (AWS/Azure/GCP)
- Distributed Systems, Data Governance
- Strong communication & stakeholder collaboration
Good to Have :
- AI certifications (Azure/GCP/AWS)
- Experience in big data and analytics


Requirements
- 7+ years of experience with Python
- Strong expertise in Python frameworks (Django, Flask, or FastAPI)
- Experience with GCP, Terraform, and Kubernetes
- Deep understanding of REST API development and GraphQL
- Strong knowledge of SQL and NoSQL databases
- Experience with microservices architecture
- Proficiency with CI/CD tools (Jenkins, CircleCI, GitLab)
- Experience with container orchestration using Kubernetes
- Understanding of cloud architecture and serverless computing
- Experience with monitoring and logging solutions
- Strong background in writing unit and integration tests
- Familiarity with AI/ML concepts and integration points
Responsibilities
- Design and develop scalable backend services for our AI platform
- Architect and implement complex systems with high reliability
- Build and maintain APIs for internal and external consumption
- Work closely with AI engineers to integrate ML functionality
- Optimize application performance and resource utilization
- Make architectural decisions that balance immediate needs with long-term scalability
- Mentor junior engineers and promote best practices
- Contribute to the evolution of our technical standards and processes

About Moative
Moative, an Applied AI company, designs and builds transformation AI solutions for traditional industries in energy, utilities, healthcare & lifesciences, and more. Through Moative Labs, we build AI micro-products and launch AI startups with partners in vertical markets that align with our theses.
Our Past: We have built and sold two companies, one of which was an AI company. Our founders and leaders are Math PhDs, Ivy League University Alumni, Ex-Googlers, and successful entrepreneurs.
Our Team: Our team of 20+ employees consist of data scientists, AI/ML Engineers, and mathematicians from top engineering and research institutes such as IITs, CERN, IISc, UZH, Ph.Ds. Our team includes academicians, IBM Research Fellows, and former founders.
Work you’ll do
As a Data Scientist at Moative, you’ll play a crucial role in extracting valuable insights from data to drive informed decision-making. You’ll work closely with cross-functional teams to build predictive models and develop solutions to complex business problems. You will also be involved in conducting experiments, building POCs and prototypes.
Responsibilities
- Support end-to-end development and deployment of ML/ AI models - from data preparation, data analysis and feature engineering to model development, validation and deployment
- Gather, prepare and analyze data, write code to develop and validate models, and continuously monitor and update them as needed.
- Collaborate with domain experts, engineers, and stakeholders in translating business problems into data-driven solutions
- Document methodologies and results, present findings and communicate insights to non-technical audiences
Skills & Requirements
- Proficiency in Python and familiarity with basic Python libraries for data analysis and ML algorithms (such as NumPy, Pandas, ScikitLearn, NLTK).
- Strong understanding and experience with data analysis, statistical and mathematical concepts and ML algorithms
- Working knowledge of cloud platforms (e.g., AWS, Azure, GCP).
- Broad understanding of data structures and data engineering.
- Strong communication skills
- Strong collaboration skills, continuous learning attitude and a problem solving mind-set
Working at Moative
Moative is a young company, but we believe strongly in thinking long-term, while acting with urgency. Our ethos is rooted in innovation, efficiency and high-quality outcomes. We believe the future of work is AI-augmented and boundary less. Here are some of our guiding principles:
- Think in decades. Act in hours. As an independent company, our moat is time. While our decisions are for the long-term horizon, our execution will be fast – measured in hours and days, not weeks and months.
- Own the canvas. Throw yourself in to build, fix or improve – anything that isn’t done right, irrespective of who did it. Be selfish about improving across the organization – because once the rot sets in, we waste years in surgery and recovery.
- Use data or don’t use data. Use data where you ought to but not as a ‘cover-my-back’ political tool. Be capable of making decisions with partial or limited data. Get better at intuition and pattern-matching. Whichever way you go, be mostly right about it.
- Avoid work about work. Process creeps on purpose, unless we constantly question it. We are deliberate about committing to rituals that take time away from the actual work. We truly believe that a meeting that could be an email, should be an email and you don’t need a person with the highest title to say that loud.
- High revenue per person. We work backwards from this metric. Our default is to automate instead of hiring. We multi-skill our people to own more outcomes than hiring someone who has less to do. We don’t like squatting and hoarding that comes in the form of hiring for growth. High revenue per person comes from high quality work from everyone. We demand it.
If this role and our work is of interest to you, please apply here. We encourage you to apply even if you believe you do not meet all the requirements listed above.
That said, you should demonstrate that you are in the 90th percentile or above. This may mean that you have studied in top-notch institutions, won competitions that are intellectually demanding, built something of your own, or rated as an outstanding performer by your current or previous employers.
The position is based out of Chennai. Our work currently involves significant in-person collaboration and we expect you to be present in the city. We intend to move to a hybrid model in a few months time.
What You’ll Do:
We’re looking for a skilled DevOps Engineer to help us build and maintain reliable, secure, and scalable infrastructure. You will work closely with our development, product, and security teams to streamline deployments, improve performance, and ensure cloud infrastructure resilience.
Responsibilities:
● Deploy, manage, and monitor infrastructure on Google Cloud Platform (GCP)
● Build CI/CD pipelines using Jenkins and integrate them with Git workflows
● Design and manage Kubernetes clusters and helm-based deployments
● Manage infrastructure as code using Terraform
● Set up logging, monitoring, and alerting (Stackdriver, Prometheus, Grafana)
● Ensure security best practices across cloud resources, networks, and secrets
● Automate repetitive operations and improve system reliability
● Collaborate with developers to troubleshoot and resolve issues in staging/production environments
What We’re Looking For:
Required Skills:
● 1–3 years of hands-on experience in a DevOps or SRE role
● Strong knowledge of GCP services (IAM, GKE, Cloud Run, VPC, Cloud Build, etc.)
● Proficiency in Kubernetes (deployment, scaling, troubleshooting)
● Experience with Terraform for infrastructure provisioning
● CI/CD pipeline setup using Jenkins, GitHub Actions, or similar tools
● Understanding of DevSecOps principles and cloud security practices
● Good command over Linux, shell scripting, and basic networking concepts
Nice to have:
● Experience with Docker, Helm, ArgoCD
● Exposure to other cloud platforms (AWS, Azure)
● Familiarity with incident response and disaster recovery planning
● Knowledge of logging and monitoring tools like ELK, Prometheus, Grafana

· Design, develop, and implement AI/ML models and algorithms.
· Focus on building Proof of Concept (POC) applications to demonstrate the feasibility and value of AI solutions.
· Write clean, efficient, and well-documented code.
· Collaborate with data engineers to ensure data quality and availability for model training and evaluation.
· Work closely with senior team members to understand project requirements and contribute to technical solutions.
· Troubleshoot and debug AI/ML models and applications.
· Stay up-to-date with the latest advancements in AI/ML.
· Utilize machine learning frameworks (e.g., TensorFlow, PyTorch, Scikit-learn) to develop and deploy models.
· Develop and deploy AI solutions on Google Cloud Platform (GCP).
· Implement data preprocessing and feature engineering techniques using libraries like Pandas and NumPy.
· Utilize Vertex AI for model training, deployment, and management.
· Integrate and leverage Google Gemini for specific AI functionalities.
Qualifications:
· Bachelor’s degree in computer science, Artificial Intelligence, or a related field.
· 3+ years of experience in developing and implementing AI/ML models.
· Strong programming skills in Python.
· Experience with machine learning frameworks such as TensorFlow, PyTorch, or Scikit-learn.
· Good understanding of machine learning concepts and techniques.
· Ability to work independently and as part of a team.
· Strong problem-solving skills.
· Good communication skills.
· Experience with Google Cloud Platform (GCP) is preferred.
· Familiarity with Vertex AI is a plus.
Job Opening: Cloud and Observability Engineer
📍 Location: Work From Office – Gurgaon (Sector 43)
🕒 Experience: 2+ Years
💼 Employment Type: Full-Time
Role Overview:
As a Cloud and Observability Engineer, you will play a critical role in helping customers transition and optimize their monitoring and observability infrastructure. You'll be responsible for building high-quality extension packages for alerts, dashboards, and parsing rules using the organization Platform. Your work will directly impact the reliability, scalability, and efficiency of monitoring across cloud-native environments.
This is a work-from-office role requiring collaboration with global customers and internal stakeholders.
Key Responsibilities:
- Extension Delivery:
- Develop, enhance, and maintain extension packages for alerts, dashboards, and parsing rules to improve monitoring experience.
- Conduct in-depth research to create world-class observability solutions (e.g., for cloud-native and container technologies).
- Customer & Internal Support:
- Act as a technical advisor to both internal teams and external clients.
- Respond to queries, resolve issues, and incorporate feedback related to deployed extensions.
- Observability Solutions:
- Design and implement optimized monitoring architectures.
- Migrate and package dashboards, alerts, and rules based on customer environments.
- Automation & Deployment:
- Use CI/CD tools and version control systems to package and deploy monitoring components.
- Continuously improve deployment workflows.
- Collaboration & Enablement:
- Work closely with DevOps, engineering, and customer success teams to gather requirements and deliver solutions.
- Deliver technical documentation and training for customers.
Requirements:
Professional Experience:
- Minimum 2 years in Systems Engineering or similar roles.
- Focus on monitoring, observability, and alerting tools.
- Cloud & Container Tech:
- Hands-on experience with AWS, Azure, or GCP.
- Experience with Kubernetes, EKS, GKE, or AKS.
- Cloud DevOps certifications (preferred).
Observability Tools:
- Practical experience with at least two observability platforms (e.g., Prometheus, Grafana, Datadog, etc.).
- Strong understanding of alerting, dashboards, and infrastructure monitoring.
Scripting & Automation:
- Familiarity with CI/CD, deployment pipelines, and version control.
- Experience in packaging and managing observability assets.
- Technical Skills:
- Working knowledge of PromQL, Grafana, and related query languages.
- Willingness to learn Dataprime and Lucene syntax.
- Soft Skills:
- Excellent problem-solving and debugging abilities.
- Strong verbal and written communication in English.
- Ability to work across US and European time zones as needed.
Why Join Us?
- Opportunity to work on cutting-edge observability platforms.
- Collaborate with global teams and top-tier clients.
- Shape the future of cloud monitoring and performance optimization.
- Growth-oriented, learning-focused environment.



Job Title: .NET Full Stack Developer
Experience: 3 to 6 Years
Work Mode: Hybrid (2-3 days from office)
Location: Gurgaon
Joiners: Immediate joiners or candidates who have completed their notice period preferred
Key Responsibilities
- Design, develop, and maintain web applications using .NET (C#) on the backend and Angular/React on the frontend.
- Develop RESTful APIs and integrate them with front-end components.
- Collaborate with UI/UX designers, backend developers, and product managers to deliver high-quality features.
- Write clean, maintainable, and efficient code following best practices.
- Participate in code reviews and contribute to continuous improvement of development processes.
- Troubleshoot and debug issues across the application stack.
- Work with DevOps teams to support CI/CD pipelines and deployment.
- Ensure application scalability, performance, and security.
- Contribute to documentation, unit testing, and version control.
Required Skills
- Strong proficiency in C# and .NET Core/.NET Framework.
- Experience with JavaScript and modern front-end frameworks like Angular or React (preference for Angular).
- Exposure to cloud platforms – Azure (preferred), AWS, or GCP.
- Good understanding of HTML5, CSS3, and TypeScript.
- Experience in RESTful API development.
- Familiarity with Entity Framework and SQL-based databases like SQL Server.
- Understanding of version control systems like Git.
- Basic knowledge of CI/CD practices and tools like Azure DevOps or Jenkins.
Job Title: GCP Data Architect
Location: Madurai
Experience: 12+ Years
Notice Period: Immediate
About TechMango
TechMango is a rapidly growing IT Services and SaaS Product company that helps global businesses with digital transformation, modern data platforms, product engineering, and cloud-first initiatives. We are seeking a GCP Data Architect to lead data modernization efforts for our prestigious client, Livingston, in a highly strategic project.
Role Summary
As a GCP Data Architect, you will be responsible for designing and implementing scalable, high-performance data solutions on Google Cloud Platform. You will work closely with stakeholders to define data architecture, implement data pipelines, modernize legacy data systems, and guide data strategy aligned with enterprise goals.
Key Responsibilities:
- Lead end-to-end design and implementation of scalable data architecture on Google Cloud Platform (GCP)
- Define data strategy, standards, and best practices for cloud data engineering and analytics
- Develop data ingestion pipelines using Dataflow, Pub/Sub, Apache Beam, Cloud Composer (Airflow), and BigQuery
- Migrate on-prem or legacy systems to GCP (e.g., from Hadoop, Teradata, or Oracle to BigQuery)
- Architect data lakes, warehouses, and real-time data platforms
- Ensure data governance, security, lineage, and compliance (using tools like Data Catalog, IAM, DLP)
- Guide a team of data engineers and collaborate with business stakeholders, data scientists, and product managers
- Create documentation, high-level design (HLD) and low-level design (LLD), and oversee development standards
- Provide technical leadership in architectural decisions and future-proofing the data ecosystem
Required Skills & Qualifications:
- 10+ years of experience in data architecture, data engineering, or enterprise data platforms
- Minimum 3–5 years of hands-on experience in GCP Data Service
- Proficient in:BigQuery, Cloud Storage, Dataflow, Pub/Sub, Composer, Cloud SQL/Spanner
- Python / Java / SQL
- Data modeling (OLTP, OLAP, Star/Snowflake schema)
- Experience with real-time data processing, streaming architectures, and batch ETL pipelines
- Good understanding of IAM, networking, security models, and cost optimization on GCP
- Prior experience in leading cloud data transformation projects
- Excellent communication and stakeholder management skills
Preferred Qualifications:
- GCP Professional Data Engineer / Architect Certification
- Experience with Terraform, CI/CD, GitOps, Looker / Data Studio / Tableau for analytics
- Exposure to AI/ML use cases and MLOps on GCP
- Experience working in agile environments and client-facing roles
What We Offer:
- Opportunity to work on large-scale data modernization projects with global clients
- A fast-growing company with a strong tech and people culture
- Competitive salary, benefits, and flexibility
- Collaborative environment that values innovation and leadership
DevOps Engineer
AiSensy
Gurugram, Haryana, India (On-site)
About AiSensy
AiSensy is a WhatsApp based Marketing & Engagement platform helping businesses like Adani, Delhi Transport Corporation, Yakult, Godrej, Aditya Birla Hindalco., Wipro, Asian Paints, India Today Group Skullcandy, Vivo, Physicswallah, Cosco grow their revenues via WhatsApp.
- Enabling 100,000+ Businesses with WhatsApp Engagement & Marketing
- 400Crores + WhatsApp Messages done between Businesses and Users via AiSensy per year
- Working with top brands like Delhi Transport Corporation, Vivo, Physicswallah & more
- High Impact as Businesses drive 25-80% Revenues using AiSensy Platform
- Mission-Driven and Growth Stage Startup backed by Marsshot.vc, Bluelotus.vc & 50+ Angel Investors
Now, we’re looking for a DevOps Engineer to help scale our infrastructure and optimize performance for millions of users. 🚀
What You’ll Do (Key Responsibilities)
🔹 CI/CD & Automation:
- Implement, manage, and optimize CI/CD pipelines using AWS CodePipeline, GitHub Actions, or Jenkins.
- Automate deployment processes to improve efficiency and reduce downtime.
🔹 Infrastructure Management:
- Use Terraform, Ansible, Chef, Puppet, or Pulumi to manage infrastructure as code.
- Deploy and maintain Dockerized applications on Kubernetes clusters for scalability.
🔹 Cloud & Security:
- Work extensively with AWS (Preferred) or other cloud platforms to build and maintain cloud infrastructure.
- Optimize cloud costs and ensure security best practices are in place.
🔹 Monitoring & Troubleshooting:
- Set up and manage monitoring tools like CloudWatch, Prometheus, Datadog, New Relic, or Grafana to track system performance and uptime.
- Proactively identify and resolve infrastructure-related issues.
🔹 Scripting & Automation:
- Use Python or Bash scripting to automate repetitive DevOps tasks.
- Build internal tools for system health monitoring, logging, and debugging.
What We’re Looking For (Must-Have Skills)
✅ Version Control: Proficiency in Git (GitLab / GitHub / Bitbucket)
✅ CI/CD Tools: Hands-on experience with AWS CodePipeline, GitHub Actions, or Jenkins
✅ Infrastructure as Code: Strong knowledge of Terraform, Ansible, Chef, or Pulumi
✅ Containerization & Orchestration: Experience with Docker & Kubernetes
✅ Cloud Expertise: Hands-on experience with AWS (Preferred) or other cloud providers
✅ Monitoring & Alerting: Familiarity with CloudWatch, Prometheus, Datadog, or Grafana
✅ Scripting Knowledge: Python or Bash for automation
Bonus Skills (Good to Have, Not Mandatory)
➕ AWS Certifications: Solutions Architect, DevOps Engineer, Security, Networking
➕ Experience with Microsoft/Linux/F5 Technologies
➕ Hands-on knowledge of Database servers


Engineering Head / Tech Lead (React + Node.js)
About MrPropTek
MrPropTek is building the future of real estate technology. We're looking for a hands-on Engineering Head / Tech Lead to drive our tech strategy and lead the development of scalable web applications across frontend and backend using React and Node.js.
Responsibilities
- Lead and mentor a team of full-stack developers
- Architect and build scalable, high-performance applications
- Drive end-to-end development using React (frontend) and Node.js (backend)
- Collaborate with product, design, and business teams to align on priorities
- Enforce code quality, best practices, and agile processes
- Oversee deployment, performance, and security of systems
Requirements
- 7+ years in software development; 3+ years in a tech lead or engineering management role
- Deep expertise in React.js, Node.js, JavaScript/TypeScript
- Experience with REST APIs, cloud platforms (AWS/GCP), and databases (SQL/NoSQL)
- Strong leadership, communication, and decision-making skills
- Startup or fast-paced team experience preferred
Job Location- Mohali, Delhi/ NCR
Job Type: Full-time
About the Company
We are hiring for a fast-growing, well-funded product startup backed by a leadership team with a proven track record of building billion-dollar digital businesses. The company is focused on delivering enterprise-grade SaaS products in the Cybersecurity domain for B2B markets. You’ll be part of a passionate and dynamic engineering team building innovative solutions using modern tech stacks.
Key Responsibilities
- Design and develop scalable microservices using Java and Spring Boot
- Build and manage robust RESTful APIs
- Collaborate with cross-functional teams in an Agile setup
- Lead and mentor junior engineers, driving technical excellence
- Contribute to architecture discussions and code reviews
- Work with PostgreSQL, implement data integrity and consistency
- Deploy and manage services on cloud platforms like GCP or Azure
- Utilize Docker/Kubernetes for containerization and orchestration
Must-Have Skills
- Strong backend experience with Java, Spring Boot, REST APIs
- Proficiency in frontend development with React.js
- Experience with PostgreSQL and database optimization
- Hands-on with cloud platforms (GCP or Azure)
- Familiarity with Docker and Kubernetes
- Strong understanding of:
- API Gateways
- Hibernate & JPA
- Transaction management & ACID properties
- Multi-threading and context switching
Good to Have
- Experience in Cybersecurity or Healthcare domain
Exposure to CI/CD pipelines and DevOps practices

We are looking for a skilled and motivated Data Engineer with strong experience in Python programming and Google Cloud Platform (GCP) to join our data engineering team. The ideal candidate will be responsible for designing, developing, and maintaining robust and scalable ETL (Extract, Transform, Load) data pipelines. The role involves working with various GCP services, implementing data ingestion and transformation logic, and ensuring data quality and consistency across systems.
Key Responsibilities:
- Design, develop, test, and maintain scalable ETL data pipelines using Python.
- Work extensively on Google Cloud Platform (GCP) services such as:
- Dataflow for real-time and batch data processing
- Cloud Functions for lightweight serverless compute
- BigQuery for data warehousing and analytics
- Cloud Composer for orchestration of data workflows (based on Apache Airflow)
- Google Cloud Storage (GCS) for managing data at scale
- IAM for access control and security
- Cloud Run for containerized applications
- Perform data ingestion from various sources and apply transformation and cleansing logic to ensure high-quality data delivery.
- Implement and enforce data quality checks, validation rules, and monitoring.
- Collaborate with data scientists, analysts, and other engineering teams to understand data needs and deliver efficient data solutions.
- Manage version control using GitHub and participate in CI/CD pipeline deployments for data projects.
- Write complex SQL queries for data extraction and validation from relational databases such as SQL Server, Oracle, or PostgreSQL.
- Document pipeline designs, data flow diagrams, and operational support procedures.
Required Skills:
- 4–8 years of hands-on experience in Python for backend or data engineering projects.
- Strong understanding and working experience with GCP cloud services (especially Dataflow, BigQuery, Cloud Functions, Cloud Composer, etc.).
- Solid understanding of data pipeline architecture, data integration, and transformation techniques.
- Experience in working with version control systems like GitHub and knowledge of CI/CD practices.
- Strong experience in SQL with at least one enterprise database (SQL Server, Oracle, PostgreSQL, etc.).
About the Role
We are looking for a highly motivated DevOps Engineer with a strong background in cloud technologies, big data ecosystems, and software development lifecycles to lead cross-functional teams in delivering high-impact projects. The ideal candidate will combine excellent project management skills with technical acumen in GCP, DevOps, and Python-based applications.
Key Responsibilities
- Lead end-to-end project planning, execution, and delivery, ensuring alignment
- Create and maintain project documentation including detailed timelines, sprint boards, risk logs, and weekly status reports.
- Facilitate Agile ceremonies: daily stand-ups, sprint planning, retrospectives, and backlog grooming.
- Actively manage risks, scope changes, resource allocation, and project dependencies to ensure delivery without disruptions.
- Ensure compliance with QA processes and security/compliance standards throughout the SDLC.
- Collaborate with stakeholders and senior leadership to communicate progress, blockers, and key milestones.
- Provide mentorship and support to cross-functional team members to drive continuous improvement and team performance.
- Coordinate with clients and act as a key point of contact for requirement gathering, updates, and escalations.
Required Skills & Experience
Cloud & DevOps
- Proficient in Google Cloud Platform (GCP) services: Compute, Storage, Networking, IAM.
- Hands-on experience with cloud deployments and infrastructure as code.
- Strong working knowledge of CI/CD pipelines, Docker, Kubernetes, and Terraform (or similar tools).
Big Data & Data Engineering
- Experience with large-scale data processing using tools like PySpark, Hadoop, Hive, HDFS, and Spark Streaming (preferred).
- Proven experience in managing and optimizing big data pipelines and ensuring high performance.
Programming & Frameworks
- Strong proficiency in Python with experience in Django (REST APIs, ORM, deployment workflows).
- Familiarity with Git and version control best practices.
- Basic knowledge of Linux administration and shell scripting.
Nice to Have
- Knowledge or prior experience in the Media & Advertising domain.
- Experience in client-facing roles and handling stakeholder communications.
- Proven ability to manage technical teams (5–6 members).
Why Join Us?
- Work on cutting-edge cloud and data engineering projects
- Collaborate with a talented, fast-paced team
- Flexible work setup and culture of ownership
1. Solid Databricks & pyspark experience
2. Must have worked in projects dealing with data at terabyte scale
3. Must have knowledge of spark optimization techniques
4. Must have experience setting up job pipelines in Databricks
5. Basic knowledge of gcp and big query is required
6. Understanding LLMs and vector db


Key Responsibilities:
● Design, develop, and maintain scalable web applications using .NET Core, .NET
Framework, C#, and related technologies.
● Participate in all phases of the SDLC, including requirements gathering, architecture
design, coding, testing, deployment, and support.
● Build and integrate RESTful APIs, and work with SQL Server, Entity Framework, and
modern front-end technologies such as Angular, React, and JavaScript.
● Conduct thorough code reviews, write unit tests, and ensure adherence to coding
standards and best practices.
● Lead or support .NET Framework to .NET Core migration initiatives, ensuring
minimal disruption and optimal performance.
● Implement and manage CI/CD pipelines using tools like Azure DevOps, Jenkins, or
GitLab CI/CD.
● Containerize applications using Docker and deploy/manage them on orchestration
platforms like Kubernetes or GKE.
● Lead and execute database migration projects, particularly transitioning from SQL
Server to PostgreSQL.
● Manage and optimize Cloud SQL for PostgreSQL, including configuration, tuning, and
ongoing maintenance.
● Leverage Google Cloud Platform (GCP) services such as GKE, Cloud SQL, Cloud
Run, and Dataflow to build and maintain cloud-native solutions.
● Handle schema conversion and data transformation tasks as part of migration and
modernization efforts.
Required Skills & Experience:
● 5+ years of hands-on experience with C#, .NET Core, and .NET Framework.
● Proven experience in application modernization and cloud-native development.
● Strong knowledge of containerization (Docker) and orchestration tools like
Kubernetes/GKE.
● Expertise in implementing and managing CI/CD pipelines.
● Solid understanding of relational databases and experience in SQL Server to
PostgreSQL migrations.
● Familiarity with cloud infrastructure, especially GCP services relevant to application
hosting and data processing.
● Excellent problem-solving, communication,

About You: ● Education ranging from a Bachelor’s of Science degree in computer science or related engineering degree. ● 12+ years of high level API, abstraction layers, and application software development experience. ● 5+ years experience building scalable, serverless solutions in GCP or AWS ● 4+ years of experience in Python, MongoDB ● Experience with large-scale distributed systems and streaming data services. ● Experience building, developing, and maintaining cloud native infrastructure, serverless architecture, micro-operations, and workflow automation. ● You are a hardworking problem-solver who thrives in finding solutions to difficult technical challenges. ● Experience with modern high-level languages and databases including Javascript, MongoDB, and Python. ● Experience in Github, Gitlab, CI/CD, Jira, unit testing, integration testing, regression testing, and collaborative documentation. ● Expertise with GCP, Kubernetes, Docker, or containerization, is a great plus. ● Ability to write and assess clean, functional, high quality and testable code for each of our projects. ● Positive and proactive, solution-focused contributor and team motivation.
Job Title: Sr. DevOps Engineer
Experience Required: 2 to 4 years in DevOps or related fields
Employment Type: Full-time
About the Role:
We are seeking a highly skilled and experienced Lead DevOps Engineer. This role will focus on driving the design, implementation, and optimization of our CI/CD pipelines, cloud infrastructure, and operational processes. As a Lead DevOps Engineer, you will play a pivotal role in enhancing the scalability, reliability, and security of our systems while mentoring a team of DevOps engineers to achieve operational excellence.
Key Responsibilities:
Infrastructure Management: Architect, deploy, and maintain scalable, secure, and resilient cloud infrastructure (e.g., AWS, Azure, or GCP).
CI/CD Pipelines: Design and optimize CI/CD pipelines, to improve development velocity and deployment quality.
Automation: Automate repetitive tasks and workflows, such as provisioning cloud resources, configuring servers, managing deployments, and implementing infrastructure as code (IaC) using tools like Terraform, CloudFormation, or Ansible.
Monitoring & Logging: Implement robust monitoring, alerting, and logging systems for enterprise and cloud-native environments using tools like Prometheus, Grafana, ELK Stack, NewRelic or Datadog.
Security: Ensure the infrastructure adheres to security best practices, including vulnerability assessments and incident response processes.
Collaboration: Work closely with development, QA, and IT teams to align DevOps strategies with project goals.
Mentorship: Lead, mentor, and train a team of DevOps engineers to foster growth and technical expertise.
Incident Management: Oversee production system reliability, including root cause analysis and performance tuning.
Required Skills & Qualifications:
Technical Expertise:
Strong proficiency in cloud platforms like AWS, Azure, or GCP.
Advanced knowledge of containerization technologies (e.g., Docker, Kubernetes).
Expertise in IaC tools such as Terraform, CloudFormation, or Pulumi.
Hands-on experience with CI/CD tools, particularly Bitbucket Pipelines, Jenkins, GitLab CI/CD, Github Actions or CircleCI.
Proficiency in scripting languages (e.g., Python, Bash, PowerShell).
Soft Skills:
Excellent communication and leadership skills.
Strong analytical and problem-solving abilities.
Proven ability to manage and lead a team effectively.
Experience:
4 years + of experience in DevOps or Site Reliability Engineering (SRE).
4+ years + in a leadership or team lead role, with proven experience managing distributed teams, mentoring team members, and driving cross-functional collaboration.
Strong understanding of microservices, APIs, and serverless architectures.
Nice to Have:
Certifications like AWS Certified Solutions Architect, Kubernetes Administrator, or similar.
Experience with GitOps tools such as ArgoCD or Flux.
Knowledge of compliance standards (e.g., GDPR, SOC 2, ISO 27001).
Perks & Benefits:
Competitive salary and performance bonuses.
Comprehensive health insurance for you and your family.
Professional development opportunities and certifications, including sponsored certifications and access to training programs to help you grow your skills and expertise.
Flexible working hours and remote work options.
Collaborative and inclusive work culture.
Join us to build and scale world-class systems that empower innovation and deliver exceptional user experiences.
You can directly contact us: Nine three one six one two zero one three two

About the Role
We are looking for a highly motivated Project Manager with a strong background in cloud technologies, big data ecosystems, and software development lifecycles to lead cross-functional teams in delivering high-impact projects. The ideal candidate will combine excellent project management skills with technical acumen in GCP, DevOps, and Python-based applications.
Key Responsibilities
- Lead end-to-end project planning, execution, and delivery, ensuring alignment with business goals and timelines.
- Create and maintain project documentation including detailed timelines, sprint boards, risk logs, and weekly status reports.
- Facilitate Agile ceremonies: daily stand-ups, sprint planning, retrospectives, and backlog grooming.
- Actively manage risks, scope changes, resource allocation, and project dependencies to ensure delivery without disruptions.
- Ensure compliance with QA processes and security/compliance standards throughout the SDLC.
- Collaborate with stakeholders and senior leadership to communicate progress, blockers, and key milestones.
- Provide mentorship and support to cross-functional team members to drive continuous improvement and team performance.
- Coordinate with clients and act as a key point of contact for requirement gathering, updates, and escalations.
Required Skills & Experience
Cloud & DevOps
- Proficient in Google Cloud Platform (GCP) services: Compute, Storage, Networking, IAM.
- Hands-on experience with cloud deployments and infrastructure as code.
- Strong working knowledge of CI/CD pipelines, Docker, Kubernetes, and Terraform (or similar tools).
Big Data & Data Engineering
- Experience with large-scale data processing using tools like PySpark, Hadoop, Hive, HDFS, and Spark Streaming (preferred).
- Proven experience in managing and optimizing big data pipelines and ensuring high performance.
Programming & Frameworks
- Strong proficiency in Python with experience in Django (REST APIs, ORM, deployment workflows).
- Familiarity with Git and version control best practices.
- Basic knowledge of Linux administration and shell scripting.
Nice to Have
- Knowledge or prior experience in the Media & Advertising domain.
- Experience in client-facing roles and handling stakeholder communications.
- Proven ability to manage technical teams (5–6 members).
Why Join Us?
- Work on cutting-edge cloud and data engineering projects
- Collaborate with a talented, fast-paced team
- Flexible work setup and culture of ownership
- Continuous learning and upskilling environment
- Inclusive health benefits included

Job Title: AI Solutioning Architect – Healthcare IT
Role Summary:
The AI Solutioning Architect leads the design and implementation of AI-driven solutions across the organization, ensuring alignment with business goals and healthcare IT standards. This role defines the AI/ML architecture, guides technical execution, and fosters innovation using platforms like Google Cloud (GCP).
Key Responsibilities:
- Architect scalable AI solutions from data ingestion to deployment.
- Align AI initiatives with business objectives and regulatory requirements (HIPAA).
- Collaborate with cross-functional teams to deliver AI projects.
- Lead POCs, evaluate AI tools/platforms, and promote GCP adoption.
- Mentor technical teams and ensure best practices in MLOps.
- Communicate complex concepts to diverse stakeholders.
Qualifications:
- Bachelor’s/Master’s in Computer Science or related field.
- 12+ years in software development/architecture with strong AI/ML focus.
- Experience in healthcare IT and compliance (HIPAA).
- Proficient in Python/Java and ML frameworks (TensorFlow, PyTorch).
- Hands-on with GCP (preferred) or other cloud platforms.
- Strong leadership, problem-solving, and communication skills.

Job Title : Lead Web Developer / Frontend Engineer
Experience Required : 10+ Years
Location : Bangalore (Hybrid – 3 Days Work From Office)
Work Timings : 11:00 AM to 8:00 PM IST
Notice Period : Immediate or Up to 30 Days (Preferred)
Work Mode : Hybrid
Interview Mode : Face-to-Face mandatory (for Round 2)
Role Overview :
We are hiring a Lead Frontend Engineer with 10+ Years of experience to drive the development of scalable, modern, and high-performance web applications.
This is a hands-on technical leadership role focused on React.js, micro-frontends, and Backend for Frontend (BFF) architecture, requiring both coding expertise and team leadership skills.
Mandatory Skills :
React.js, JavaScript/TypeScript, HTML, CSS, micro-frontend architecture, Backend for Frontend (BFF), Webpack, Jenkins (CI/CD), GCP, RDBMS/SQL, Git, and team leadership.
Core Responsibilities :
- Design and develop cloud-based web applications using React.js, HTML, CSS.
- Collaborate with UX/UI designers and backend engineers to implement seamless user experiences.
- Lead and mentor a team of frontend developers.
- Write clean, well-documented, scalable code using modern JavaScript/TypeScript practices.
- Implement CI/CD pipelines using Jenkins, deploy applications to CDNs.
- Integrate with GCP services, optimize front-end performance.
- Stay updated with modern frontend technologies and design patterns.
- Use Git for version control and collaborative workflows.
- Implement JavaScript libraries for web analytics and performance monitoring.
Key Requirements :
- 10+ Years of experience as a frontend/web developer.
- Strong proficiency in React.js, JavaScript/TypeScript, HTML, CSS.
- Experience with micro-frontend architecture and Backend for Frontend (BFF) patterns.
- Proficiency in frontend design frameworks and libraries (jQuery, Node.js).
- Strong understanding of build tools like Webpack, CI/CD using Jenkins.
- Experience with GCP and deploying to CDNs.
- Solid experience in RDBMS, SQL.
- Familiarity with Git and agile development practices.
- Excellent debugging, problem-solving, and communication skills.
- Bachelor’s/Master’s in Computer Science or a related field.
Nice to Have :
- Experience with Node.js.
- Previous experience working with web analytics frameworks.
- Exposure to JavaScript observability tools.
Interview Process :
1. Round 1 : Online Technical Interview (via Geektrust – 1 Hour)
2. Round 2 : Face-to-Face Interview with the Indian team in Bangalore (3 Hours – Mandatory)
3. Round 3 : Online Interview with CEO (30 Minutes)
Important Notes :
- Face-to-face interview in Bangalore is mandatory for Round 2.
- Preference given to candidates currently in Bangalore or willing to travel for interviews.
- Remote applicants who cannot attend the in-person round will not be considered.

Why This Role Matters
We’re looking for a Principal Engineer to lead the architecture and execution of our GenAI-powered, self-serve marketing platforms. You will work directly with the CEO to shape, build, and scale products that change how marketers interact with data and AI. This is intrapreneurship in action — not a sandbox innovation lab, but a real-world product with traction, velocity, and high stakes.
What You'll Do
- Co-own product architecture and direction alongside the CEO.
- Build GenAI-native, full-stack platforms from MVP to scale — powered by LLMs, agents, and predictive AI.
- Own the full stack: React (frontend), Node.js/Python (backend), GCP (infra), BigQuery (data), and vector databases (AI).
- Lead a lean, high-caliber team with a hands-on, unblock-and-coach mindset.
- Drive rapid iteration with rigor, balancing short-term delivery with long-term resilience.
- Ensure scalability, observability, and fault tolerance in multi-tenant, cloud-native environments.
- Bridge business and tech — aligning execution with evolving user and market insights.
What You Bring
- 8–12 years of experience building and scaling full-stack, data-heavy or AI-driven products.
- Fluency in React, Node.js, and Google Cloud (Functions, BigQuery, Cloud SQL, Airflow, etc.).
- Hands-on experience with GenAI tools (LangChain, OpenAI APIs, LlamaIndex) is a bonus.
- Track record of shipping products from ambiguity to impact.
- Strong product mindset — your goal is user value, not just elegant code.
- Architectural leadership with ownership of engineering rigor and scaling best practices.
- Startup or founder DNA — you’ve built things from scratch and know how to move fast without breaking things.
Who You Are
- A former founder, senior IC, or tech lead who’s done zero-to-one and 1-to-n scaling.
- Hungry for ownership and velocity — frustrated by bureaucracy or stagnation.
- You code because you care about solving real problems for real users.
- You’re pragmatic, hands-on, and grounded in first principles.
- You understand that great software isn't just shipped — it's hardened, maintained, and evolves with minimal manual effort.
- You’re open to evolving into a founding engineer role with influence over the tech vision and culture.
What You Get
- Equity in a high-growth product-led startup.
- A chance to build global products out of India with full-stack and GenAI innovation.
- Access to high-context decision-making and direct collaboration with the CEO.
- A tight, ego-free team and a culture that values clarity, ownership, learning, and candor.
Why YOptima?
YOptima is redefining how leading marketers unlock growth through full-funnel, AI-powered media solutions. As part of our growth journey, this is your opportunity to own the growth charter for leading brands and agencies globally and shape the narrative of a next-generation marketing platform.
Ready to lead, build, and scale?
We’d love to hear from you.


About the Role:
We’re looking for a skilled developer to build and maintain web and mobile apps using React, React Native, and Node.js. You’ll work on both the frontend and backend, collaborating with our team to deliver high-quality products.
What You’ll Do:
- Build and maintain full stack applications for web and mobile
- Write clean, efficient code with React, React Native, and Node.js
- Work with designers and other developers to deliver new features
- Debug, troubleshoot, and optimize existing apps
- Stay updated on the latest tech and best practices
What We’re Looking For:
- Solid experience with React, React Native, and Node.js
- Comfortable building both web and mobile applications
- Good understanding of REST APIs and databases
- Familiar with Git and agile workflows
- Team player with clear communication skills
Nice to Have:
- Experience with testing and CI/CD
- Knowledge of UI/UX basics
Requirements
- Bachelors/Masters in Computer Science or a related field
- 5-8 years of relevant experience
- Proven track record of Team Leading/Mentoring a team successfully.
- Experience with web technologies and microservices architecture both frontend and backend.
- Java, Spring framework, hibernate
- MySQL, Mongo, Solr, Redis,
- Kubernetes, Docker
- Strong understanding of Object-Oriented Programming, Data Structures, and Algorithms.
- Excellent teamwork skills, flexibility, and ability to handle multiple tasks.
- Experience with API Design, ability to architect and implement an intuitive customer and third-party integration story
- Ability to think and analyze both breadth-wise (client, server, DB, control flow) and depth-wise (threads, sessions, space-time complexity) while designing and implementing services
- Exceptional design and architectural skills
- Experience of cloud providers/platforms like GCP and AWS
Roles & Responsibilities
- Develop new user-facing features.
- Work alongside the product to understand our requirements, and design, develop and iterate, think through the complex architecture.
- Writing clean, reusable, high-quality, high-performance, maintainable code.
- Encourage innovation and efficiency improvements to ensure processes are productive.
- Ensure the training and mentoring of the team members.
- Ensure the technical feasibility of UI/UX designs and optimize applications for maximum speed.
- Research and apply new technologies, techniques, and best practices.
- Team mentorship and leadership.

Product company for financial operations automation platform

Mandatory Criteria :
- Candidate must have Strong hands-on experience with Kubernetes of atleast 2 years in production environments.
- Candidate should have Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI).
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Candidate should have strong Backend experience.
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
About the Role
We are looking for a highly skilled and motivated Cloud Backend Engineer with 4–7 years of experience, who has worked extensively on at least one major cloud platform (GCP, AWS, Azure, or OCI). Experience with multiple cloud providers is a strong plus. As a Senior Development Engineer, you will play a key role in designing, building, and scaling backend services and infrastructure on cloud-native platforms.
# Experience with Kubernetes is mandatory.
Key Responsibilities
- Design and develop scalable, reliable backend services and cloud-native applications.
- Build and manage RESTful APIs, microservices, and asynchronous data processing systems.
- Deploy and operate workloads on Kubernetes with best practices in availability, monitoring, and cost-efficiency.
- Implement and manage CI/CD pipelines and infrastructure automation.
- Collaborate with frontend, DevOps, and product teams in an agile environment.
- Ensure high code quality through testing, reviews, and documentation.
Required Skills
- Strong hands-on experience with Kubernetes of atleast 2 years in production environments (mandatory).
- Expertise in at least one public cloud platform [GCP (Preferred), AWS, Azure, or OCI].
- Proficient in backend programming with Python, Java, or Kotlin (at least one is required).
- Solid understanding of distributed systems, microservices, and cloud-native architecture.
- Experience with containerization using Docker and Kubernetes-native deployment workflows.
- Working knowledge of SQL and relational databases.
Preferred Qualifications
- Experience working across multiple cloud platforms.
- Familiarity with infrastructure-as-code tools like Terraform or CloudFormation.
- Exposure to monitoring, logging, and observability stacks (e.g., Prometheus, Grafana, Cloud Monitoring).
- Hands-on experience with BigQuery or Snowflake for data analytics and integration.
Nice to Have
- Knowledge of NoSQL databases or event-driven/message-based architectures.
- Experience with serverless services, managed data pipelines, or data lake platforms.
Key Skills Required:
- Strong hands-on experience with Terraform
- Proficiency in CI/CD tools (e.g., Jenkins, GitLab CI, Azure DevOps, etc.)
- Experience working on Azure or GCP cloud platforms (at least one is mandatory)
- Good understanding of DevOps practices


Job title - Python developer
Exp – 4 to 6 years
Location – Pune/Mum/B’lore
PFB JD
Requirements:
- Proven experience as a Python Developer
- Strong knowledge of core Python and Pyspark concepts
- Experience with web frameworks such as Django or Flask
- Good exposure to any cloud platform (GCP Preferred)
- CI/CD exposure required
- Solid understanding of RESTful APIs and how to build them
- Experience working with databases like Oracle DB and MySQL
- Ability to write efficient SQL queries and optimize database performance
- Strong problem-solving skills and attention to detail
- Strong SQL programing (stored procedure, functions)
- Excellent communication and interpersonal skill
Roles and Responsibilities
- Design, develop, and maintain data pipelines and ETL processes using pyspark
- Work closely with data scientists and analysts to provide them with clean, structured data.
- Optimize data storage and retrieval for performance and scalability.
- Collaborate with cross-functional teams to gather data requirements.
- Ensure data quality and integrity through data validation and cleansing processes.
- Monitor and troubleshoot data-related issues to ensure data pipeline reliability.
- Stay up to date with industry best practices and emerging technologies in data engineering.
We are looking for an experienced GCP Cloud Engineer to design, implement, and manage cloud-based solutions on Google Cloud Platform (GCP). The ideal candidate should have expertise in GKE (Google Kubernetes Engine), Cloud Run, Cloud Loadbalancer, Cloud function, Azure DevOps, and Terraform, with a strong focus on automation, security, and scalability.
You will work closely with development, operations, and security teams to ensure robust cloud infrastructure and CI/CD pipelines while optimizing performance and cost.
Key Responsibilities:
1. Cloud Infrastructure Design & Management
· Architect, deploy, and maintain GCP cloud resources via terraform/other automation.
· Implement Google Cloud Storage, Cloud SQL, file store, for data storage and processing needs.
· Manage and configure Cloud Load Balancers (HTTP(S), TCP/UDP, and SSL Proxy) for high availability and scalability.
· Optimize resource allocation, monitoring, and cost efficiency across GCP environments.
2. Kubernetes & Container Orchestration
· Deploy, manage, and optimize workloads on Google Kubernetes Engine (GKE).
· Work with Helm charts, Istio, and service meshes for microservices deployments.
· Automate scaling, rolling updates, and zero-downtime deployments.
3. Serverless & Compute Services
· Deploy and manage applications on Cloud Run and Cloud Functions for scalable, serverless workloads.
· Optimize containerized applications running on Cloud Run for cost efficiency and performance.
4. CI/CD & DevOps Automation
· Design, implement, and manage CI/CD pipelines using Azure DevOps.
· Automate infrastructure deployment using Terraform, Bash and Power shell scripting
· Integrate security and compliance checks into the DevOps workflow (DevSecOps).
Required Skills & Qualifications:
✔ Experience: 8+ years in Cloud Engineering, with a focus on GCP.
✔ Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions).
✔ Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm.
✔ DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation.
✔ Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources.
✔ Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation.
✔ Security & Compliance: Knowledge of cloud security principles, IAM, and compliance standards.


LendFlow is an AI-powered home loan assessment platform that helps mortgage brokers and lenders save hours by automating document analysis, income validation, and serviceability assessment. We turn complex financial documents into clear insights—fast.
We’re building a smart assistant that ingests client docs (bank statements, payslips, loan summaries) and uses modular AI agents to extract, classify, and summarize financial data in minutes, not hours. Think OCR + AI agents + compliance-ready outputs.
🛠️ What You’ll Be Building
As part of our early technical team, you’ll help us develop and launch our MVP. Key modules include:
- Document ingestion and OCR processing (Textract, Document AI)
- AI agent workflows using LangChain or CrewAI
- Serviceability calculators with business rule engines
- React + Next.js frontend for brokers and analysts
- FastAPI backend with PostgreSQL
- Security, encryption, audit logging (privacy-first design)
🎯 We’re Looking For:
Must-Have Skills:
- Strong experience with Python (FastAPI, OCR, LLMs, prompt engineering)
- Familiarity with AI agent frameworks (LangChain, CrewAI, Autogen, or similar)
- Frontend skills in React.js / Next.js
- Experience with PostgreSQL and cloud storage (AWS/GCP)
- Understanding of financial documents and data privacy best practices
Bonus Points:
- Experience with OCR tools like Amazon Textract, Tesseract, or Document AI
- Building ML/NLP pipelines in real-world apps
- Prior work in fintech, lending, or proptech sectors
Must have handled at least one project of medium to high complexity of migrating ETL pipelines and data warehouses to cloud.
Min 3 years of experience with premium consulting companies.
mandatory experience in GCP.
Strong proficiency in at least one major cloud platform - Azure, GCP(Primary) or AWS. Azure + GCP (Secondary) preferred. Having proficiency in all 3 is a significant plus.
· Design, develop, and maintain cloud-based applications and infrastructure across various cloud platforms.
· Select and configure appropriate cloud services based on specific project requirements and constraints.
· Implement infrastructure automation with tools like Terraform and Ansible.
· Write clean, efficient, and well-documented code using various programming languages. (Python (Required), Knowledge of Java, C#, JavaScript is a plus).
· Implement RESTful APIs and microservices architectures.
· Utilize DevOps practices for continuous integration and continuous delivery (CI/CD).
· Design, configure, and manage scalable and secure cloud infrastructure for MLOps.
· Monitor and optimize cloud resources for performance and cost efficiency.
· Implement security best practices throughout the development lifecycle.
· Collaborate with developers, operations, and security teams to ensure seamless integration and successful deployments.
· Stay up-to-date on the latest cloud technologies, MLOps tools and trends
· Strong analytical and problem-solving skills.
Looking for Fresher developers
Responsibilities:
- Implement integrations requested by customers
- Deploy updates and fixes
- Provide Level 2 technical support
- Build tools to reduce occurrences of errors and improve customer experience
- Develop software to integrate with internal back-end systems
- Perform root cause analysis for production errors
- Investigate and resolve technical issues
- Develop scripts to automate visualization
- Design procedures for system troubleshooting and maintenance
Requirements and skill:
Knowledge in DevOps Engineer or similar software engineering role
Good knowledge of Terraform, Kubernetes
Working knowledge on AWS, Google Cloud
You can directly contact me on nine three one six one two zero one three two
Job Description for Database Consultant-I (PostgreSQL)
Job Title: Database Consultant-I (PostgreSQL)
Company: Mydbops
About Us:
Mydbops is a trusted leader with 8+ years of excellence in open-source database management. We deliver best-in-class services across MySQL, MariaDB, MongoDB, PostgreSQL, TiDB, Cassandra, and more. Our focus is on building scalable, secure, and high-performance database solutions for global clients. As a PCI DSS-certified and ISO-certified organisation, we are committed to operational excellence and data security.
Role Overview:
As a Database Consultant – I (PostgreSQL Team), you will take ownership of PostgreSQL database environments, offering expert-level support to our clients. This role involves proactive monitoring, performance tuning, troubleshooting, high availability setup, and guiding junior team members. You will play a key role in customer-facing technical delivery, solution design, and implementation.
Key Responsibilities:
- Manage PostgreSQL production environments for performance, stability, and scalability.
- Handle complex troubleshooting, performance analysis, and query optimisation.
- Implement backup strategies, recovery solutions, replication, and failover techniques.
- Set up and manage high availability architectures (Streaming Replication, Patroni, etc.).
- Work with DevOps/cloud teams for deployment and automation.
- Support upgrades, patching, and migration projects across environments.
- Use monitoring tools to proactively detect and resolve issues.
- Mentor junior engineers and guide troubleshooting efforts.
- Interact with clients to understand requirements and deliver solutions.
Requirements:
- 3–5 years of hands-on experience in PostgreSQL database administration.
- Strong Linux OS knowledge and scripting skills (Bash/Python).
- Proficiency in SQL tuning, performance diagnostics, and explain plans.
- Experience with tools like pgBackRest, Barman for backup and recovery.
- Familiarity with high availability, failover, replication, and clustering.
- Good understanding of AWS RDS, Aurora PostgreSQL, and GCP Cloud SQL.
- Experience with monitoring tools like pg_stat_statements, PMM, Nagios, or custom dashboards.
- Knowledge of automation/configuration tools like Ansible or Terraform is a plus.
- Strong communication and problem-solving skills.
Preferred Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Engineering, or equivalent.
- PostgreSQL certification (EDB/Cloud certifications preferred).
- Past experience in a consulting, customer support, or managed services role.
- Exposure to multi-cloud environments and database-as-a-service platforms.
- Prior experience with database migrations or modernisation projects.
Why Join Us:
- Opportunity to work in a dynamic and growing industry.
- Learning and development opportunities to enhance your career.
- A collaborative work environment with a supportive team.
Job Details:
- Job Type: Full-time
- Work Days: 5 Days
- Work Mode: Work From Home
- Experience Required: 3–5 years

Position : Senior Data Analyst
Experience Required : 5 to 8 Years
Location : Hyderabad or Bangalore (Work Mode: Hybrid – 3 Days WFO)
Shift Timing : 11:00 AM – 8:00 PM IST
Notice Period : Immediate Joiners Only
Job Summary :
We are seeking a highly analytical and experienced Senior Data Analyst to lead complex data-driven initiatives that influence key business decisions.
The ideal candidate will have a strong foundation in data analytics, cloud platforms, and BI tools, along with the ability to communicate findings effectively across cross-functional teams. This role also involves mentoring junior analysts and collaborating closely with business and tech teams.
Key Responsibilities :
- Lead the design, execution, and delivery of advanced data analysis projects.
- Collaborate with stakeholders to identify KPIs, define requirements, and develop actionable insights.
- Create and maintain interactive dashboards, reports, and visualizations.
- Perform root cause analysis and uncover meaningful patterns from large datasets.
- Present analytical findings to senior leaders and non-technical audiences.
- Maintain data integrity, quality, and governance in all reporting and analytics solutions.
- Mentor junior analysts and support their professional development.
- Coordinate with data engineering and IT teams to optimize data pipelines and infrastructure.
Must-Have Skills :
- Strong proficiency in SQL and Databricks
- Hands-on experience with cloud data platforms (AWS, Azure, or GCP)
- Sound understanding of data warehousing concepts and BI best practices
Good-to-Have :
- Experience with AWS
- Exposure to machine learning and predictive analytics
- Industry-specific analytics experience (preferred but not mandatory)