Cutshort logo
Google Cloud Platform (GCP) Jobs in Mumbai

50+ Google Cloud Platform (GCP) Jobs in Mumbai | Google Cloud Platform (GCP) Job openings in Mumbai

Apply to 50+ Google Cloud Platform (GCP) Jobs in Mumbai on CutShort.io. Explore the latest Google Cloud Platform (GCP) Job opportunities across top companies like Google, Amazon & Adobe.

icon
Quantiphi

at Quantiphi

3 candid answers
1 video
Nikita Sinha
Posted by Nikita Sinha
Mumbai, Bengaluru (Bangalore), Trivandrum
7 - 10 yrs
Upto ₹35L / yr (Varies
)
skill icon.NET
Microservices
Architecture
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
+1 more

We are looking for a Mid-Level .NET Developer to design, develop, and maintain scalable microservices for enterprise applications. The role involves working on high-performance, reliable systems deployed in containerized environments.


Key Responsibilities:

  • Develop and maintain scalable .NET microservices
  • Build robust Web APIs with proper validation, error handling, and security
  • Write unit and integration tests to ensure code quality
  • Design portable and environment-agnostic solutions
  • Collaborate with cross-functional teams and client stakeholders
  • Optimize performance and implement caching strategies
  • Follow security best practices for enterprise applications
  • Participate in code reviews and maintain coding standards
  • Support deployment and troubleshoot issues in client environments

Must-Have Skills:

Core Technical Expertise:

  • 4+ years of experience with .NET Core (3.1+) / .NET 5+ and C# (8+)
  • Strong hands-on experience with ASP.NET Core Web API & Entity Framework Core
  • Experience building REST APIs and middleware
  • Strong understanding of SOLID principles, Dependency Injection, Repository pattern
  • Experience with unit testing (xUnit / NUnit / MSTest), Moq, integration testing

Microservices & Deployment:

  • Hands-on experience with Docker
  • Understanding of microservices architecture & distributed systems
  • Experience with configuration management (appsettings.json, IConfiguration)
  • Knowledge of NuGet and dependency management

Good-to-Have Skills:

Advanced Technical:

  • Experience with .NET 6/7/8, Minimal APIs, gRPC, SignalR
  • Advanced EF Core, Dapper, database migrations
  • Kubernetes and container orchestration
  • Cloud platforms: Azure / GCP / Alibaba Cloud
  • Message brokers: Azure Service Bus, RabbitMQ, Kafka
  • Databases: PostgreSQL, MySQL, MongoDB, Cassandra
  • API Gateways: Azure API Management, Kong

Development & Operations:

  • CI/CD tools: Azure DevOps, Jenkins, GitHub Actions
  • Monitoring: Application Insights, Serilog, Prometheus
  • Security: HTTPS, CORS, input validation, secure coding
  • Background services: Hangfire, Quartz.NET

Client-Facing Experience:

  • Experience in service-based organizations
  • Ability to adapt to multiple domains
  • Understanding of industry standards and compliance 


Read more
Quantiphi

at Quantiphi

3 candid answers
1 video
Nikita Sinha
Posted by Nikita Sinha
Mumbai, Bengaluru (Bangalore), Trivandrum
4 - 6 yrs
Upto ₹25L / yr (Varies
)
skill icon.NET
Microservices
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
Windows Azure

We are looking for a Mid-Level .NET Developer to design, develop, and maintain scalable microservices for enterprise applications. The role involves working on high-performance, reliable systems deployed in containerized environments.


Key Responsibilities:

  • Develop and maintain scalable .NET microservices
  • Build robust Web APIs with proper validation, error handling, and security
  • Write unit and integration tests to ensure code quality
  • Design portable and environment-agnostic solutions
  • Collaborate with cross-functional teams and client stakeholders
  • Optimize performance and implement caching strategies
  • Follow security best practices for enterprise applications
  • Participate in code reviews and maintain coding standards
  • Support deployment and troubleshoot issues in client environments

Must-Have Skills:

Core Technical Expertise:

  • 4+ years of experience with .NET Core (3.1+) / .NET 5+ and C# (8+)
  • Strong hands-on experience with ASP.NET Core Web API & Entity Framework Core
  • Experience building REST APIs and middleware
  • Strong understanding of SOLID principles, Dependency Injection, Repository pattern
  • Experience with unit testing (xUnit / NUnit / MSTest), Moq, integration testing

Microservices & Deployment:

  • Hands-on experience with Docker
  • Understanding of microservices architecture & distributed systems
  • Experience with configuration management (appsettings.json, IConfiguration)
  • Knowledge of NuGet and dependency management

Good-to-Have Skills:

Advanced Technical:

  • Experience with .NET 6/7/8, Minimal APIs, gRPC, SignalR
  • Advanced EF Core, Dapper, database migrations
  • Kubernetes and container orchestration
  • Cloud platforms: Azure / GCP / Alibaba Cloud
  • Message brokers: Azure Service Bus, RabbitMQ, Kafka
  • Databases: PostgreSQL, MySQL, MongoDB, Cassandra
  • API Gateways: Azure API Management, Kong

Development & Operations:

  • CI/CD tools: Azure DevOps, Jenkins, GitHub Actions
  • Monitoring: Application Insights, Serilog, Prometheus
  • Security: HTTPS, CORS, input validation, secure coding
  • Background services: Hangfire, Quartz.NET

Client-Facing Experience:

  • Experience in service-based organizations
  • Ability to adapt to multiple domains
  • Understanding of industry standards and compliance 
Read more
Quantiphi

at Quantiphi

3 candid answers
1 video
Nikita Sinha
Posted by Nikita Sinha
Bengaluru (Bangalore), Mumbai, Trivandrum
3 - 6 yrs
Upto ₹30L / yr (Varies
)
Google Cloud Platform (GCP)
DevOps
CI/CD
skill iconKubernetes
skill iconGitHub
+2 more

Role & Responsibilities

  • Develop and deliver automation software to build and improve platform functionality
  • Ensure reliability, availability, and manageability of applications and cloud platforms
  • Champion adoption of Infrastructure as Code (IaC) practices
  • Design and build self-service, self-healing, monitoring, and alerting platforms
  • Automate development and testing workflows through CI/CD pipelines (Git, Jenkins, SonarQube, Artifactory, Docker containers)
  • Build and manage container hosting platforms using Kubernetes

Requirements

  • Strong experience deploying and maintaining GCP cloud infrastructure
  • Well-versed in service-oriented and cloud-based architecture design patterns
  • Knowledge of cloud services including compute, storage, networking, messaging, and automation tools (e.g., CloudFormation/Terraform equivalents)
  • Experience with relational and NoSQL databases (Postgres, Cassandra)
  • Hands-on experience with automation/configuration tools (Puppet, Chef, Ansible, Terraform)

Additional Skills

  • Strong Linux system administration and troubleshooting skills
  • Programming/scripting exposure (Bash, Python, Core Java, or Scala)
  • CI/CD pipeline experience (Jenkins, Git, Maven, etc.)
  • Experience integrating solutions in multi-region environments
  • Familiarity with Agile/Scrum/DevOps methodologies
Read more
Quantiphi

at Quantiphi

3 candid answers
1 video
Nikita Sinha
Posted by Nikita Sinha
Bengaluru (Bangalore), Mumbai, Trivandrum
4 - 7 yrs
Upto ₹35L / yr (Varies
)
Large Language Models (LLM) tuning
skill iconDeep Learning
Google Cloud Platform (GCP)
Google Vertex AI
Windows Azure
+2 more

Build, deploy, and maintain production-grade AI/ML solutions for Fortune 500 enterprise clients on Google Cloud Platform. Hands-on role focused on shipping scalable AI systems across GenAI, agentic workflows, traditional ML, and computer vision.


Key Responsibilities:


Generative AI & Agentic Systems

  • Design and build GenAI applications (RAG, agentic workflows, multi-agent systems)
  • Develop intelligent systems with memory, planning, and reasoning capabilities
  • Implement prompt engineering, context optimization, and evaluation frameworks
  • Build observable and reliable multi-agent architectures

Traditional ML & Computer Vision

  • Develop ML pipelines (forecasting, recommendation, classification, regression)
  • Build production-grade computer vision solutions (document AI, image analysis)
  • Perform feature engineering, model optimization, and benchmarking

MLOps & Production Engineering

  • Own end-to-end ML lifecycle (CI/CD, testing, versioning, deployment)
  • Build scalable APIs, microservices, and data pipelines
  • Monitor models, detect drift, and implement A/B testing frameworks

Knowledge Solutions

  • Architect knowledge graphs and semantic search systems
  • Implement hybrid retrieval (vector + keyword search)

Client Collaboration

  • Present technical solutions to enterprise clients
  • Collaborate with architects, data engineers, and business teams

Required Skills & Experience

  • 3–6 years of hands-on ML Engineering experience
  • Strong Python and software engineering fundamentals
  • Experience shipping production ML systems on cloud (GCP preferred)
  • Experience across GenAI, Traditional ML, Computer Vision
  • MLOps experience and RAG-based systems

Preferred

  • GCP Professional ML Engineer certification
  • Knowledge graphs / semantic search experience
  • Experience in regulated industries (Healthcare / BFSI)
  • Open-source or technical publications
Read more
Arcis India
Sarita Jena
Posted by Sarita Jena
Mumbai
6 - 8 yrs
₹12L - ₹20L / yr
skill iconJava
skill iconSpring Boot
Quarkus
Microservices
Webservices
+17 more

6 + years of hands-on development experience and in-depth knowledge of , Spring Java, Spring boot, Quarkus and nice to have front-end technologies like Angular, React JS

● Excellent Engineering skills in designing and implementing scalable solutions

● Good knowledge of CI/CD Pipeline with strong focus on TDD

Strong communication skills and ownership

● Exposure to Cloud, Kubernetes, Docker, Microservices is highly desired.

● Experience in working on public cloud environments like AWS, Azure, GCP w.r.t. solutions development, deployment & adoption of cloud-based technology components like IaaS / PaaS offerings

● Proficiency in PL/SQL and Database development.

Strong in J2EE & OOPS Design Patterns.

Read more
Searce Inc

at Searce Inc

3 recruiters
Srishti Dani
Posted by Srishti Dani
Mumbai, Pune, Bengaluru (Bangalore)
7 - 10 yrs
Best in industry
Data migration
Datawarehousing
ETL
SQL
Google Cloud Platform (GCP)
+7 more

Lead Data Engineer


What are we looking for

real solver?

Solver? Absolutely. But not the usual kind. We're searching for the architects of the audacious & the pioneers of the possible. If you're the type to dismantle assumptions, re-engineer ‘best practices,’ and build solutions that make the future possible NOW, then you're speaking our language.


Your Responsibilities

What you will wake up to solve.

  • Lead Technical Design & Data Architecture: Architect and lead the end-to-end development of scalable, cloud-native data platforms. You’ll guide the squad on critical architectural decisions—choosing between Batch vs. Streaming or ETL vs. ELT—while remaining 100% hands-on, contributing high-quality, production-grade code.
  • Build High-Velocity Data Pipelines: Drive the implementation of robust data transports and ingestion frameworks using Python, SQL, and Spark. You will build integration layers that connect heterogeneous sources (SaaS, RDBMS, NoSQL) into unified, high-availability environments like BigQuery, Snowflake, or Redshift.
  • Mentor & Elevate the Squad: Foster a culture of technical excellence by mentoring and inspiring a team of data analysts and engineers. Lead deep-dive code reviews, promote best-practice data modeling (Star/Snowflake schema), and ensure the squad adopts modern engineering standards like CI/CD for data.
  • Drive AI-Ready Data Strategy: Be the expert in designing data foundations optimized for AI and Machine Learning. You will champion the use of GCP (Dataflow, Pub/Sub, BigQuery) and AWS (Lambda, Glue, EMR) to create "clean room" environments that fuel advanced analytics and generative AI models.
  • Partner with Clients as a Technical DRI: Act as the Directly Responsible Individual for client success. Translate ambiguous business questions into elegant data services, manage project deliverables using Agile methodologies, and ensure that the data provided is accurate, consistent, and mission-critical.
  • Troubleshoot & Optimize for Scale: Own the reliability of the reporting layer. You will proactively monitor pipelines, troubleshoot complex transformation bottlenecks, and propose ways to improve platform performance and cost-efficiency.
  • Innovate and Build Reusable IP: Spearhead the creation of reusable data frameworks, custom operators, and transformation libraries that accelerate future projects and establish Searce’s unique technical advantage in the market.


Welcome to Searce


The AI-Native tech consultancy that's rewriting the rules.

Searce is an AI-native, engineering-led, modern tech consultancy that empowers clients to futurify their business by delivering intelligent, impactful, real business outcomes. Searce solvers co-innovate with clients as their trusted transformational partners ensuring sustained competitive advantage. Searce clients realize smarter, faster, better business outcomes delivered by AI-native Searce solver squads. 


Functional Skills 

the solver personas.

  • The Data Architect: This persona deconstructs ambiguous business goals into scalable, elegant data blueprints. They don't just move data; they design the foundation—from schema design to partitioning strategies—that allows data scientists and analysts to thrive, foreseeing technical bottlenecks and making pragmatic trade-offs.
  • The Player-Coach: As a hands-on leader, this persona leads from the front by writing exemplary, production-grade SQL and Python while simultaneously mentoring and elevating the skills of the squad. Their success is measured by the team's ability to deliver high-quality, maintainable code and their growth as engineers.
  • The Pragmatic Innovator: This individual balances a passion for modern data tech (like Generative AI and Real-time Streaming) with a sharp focus on business outcomes. They champion new tools where they add real value but are disciplined enough to choose stable, cost-effective solutions to meet deadlines and deliver robust products.
  • The Client-Facing Technologist: This persona acts as the crucial technical bridge between the data squad and the client. They build trust by listening actively, explaining complex data concepts (like data latency or idempotency) in simple terms, and demonstrating how engineering decisions align with the client’s strategic goals.
  • The Quality Craftsman: This individual possesses an unwavering commitment to data integrity and treats data engineering as a craft. They are the guardian of the reporting layer, advocating for robust testing, data validation frameworks, and clean, modular code to ensure the long-term reliability of the data platform.


Experience & Relevance 

  • Engineering Depth: 7-10 years of professional experience in end-to-end data product development. You have a portfolio that proves your ability to build complex, high-velocity pipelines for both Batch and Streaming workloads.
  • Cloud-Native Fluency: Deep, hands-on experience designing and deploying scalable data solutions on at least one major cloud platform (AWS, GCP, or Azure). You are comfortable navigating the nuances of EMR, BigQuery, or Synapse at scale.
  • AI-Native Workflow: You don’t just build for AI; you build with AI. You must be proficient in using AI coding assistants (e.g., GitHub Copilot) to accelerate your delivery and have a track record of building the data foundations required for Generative AI.
  • Architectural Portfolio: Evidence of leading 2-3 large-scale transformations—including platform migrations, data lakehouse builds, or real-time analytics architectures.
  • Client-Facing Acumen: You have direct experience in a consultative, client-facing role. You can confidently translate a CEO’s business vision into a Lead Engineer’s technical specification without losing anything in translation.


Join the ‘real solvers’

ready to futurify?

If you are excited by the possibilities of what an AI-native engineering-led, modern tech consultancy can do to futurify businesses, apply here and experience the ‘Art of the possible’. Don’t Just Send a Resume. Send a Statement.


Read more
Searce Inc

at Searce Inc

3 recruiters
Jatin Gereja
Posted by Jatin Gereja
Bengaluru (Bangalore), Mumbai, Pune
10 - 18 yrs
Best in industry
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
Enterprise Data Warehouse (EDW)
Data modeling
Big Data
+9 more

Director - Data engineering


What are we looking for

real solver?

Solver? Absolutely. But not the usual kind. We're searching for the architects of the audacious & the pioneers of the possible. If you're the type to dismantle assumptions, re-engineer ‘best practices,’ and build solutions that make the future possible NOW, then you're speaking our language.


Your Responsibilities

what you will wake up to solve.

1. Delivery & Tactical Rigor

  • Methodology Implementation: Implement and manage a unified, 'DataOps-First' methodology for data engineering delivery (ETL/ELT pipelines, Data Modeling, MLOps, Data Governance) within assigned business units. This ensures predictable outcomes and trusted data integrity by reducing architecture variability at the project level.
  • Operational Stewardship: Drive initiatives to optimize team utilization and enhance operational efficiency within the practice. You manage the commercial success of your squads, ensuring data delivery models (from migration to modern data stack implementation) are executed profitably, scalably, and cost-effectively.
  • Execution & Technical Resolution
  • Technical Escalation: Serve as the primary escalation point for delivery issues, personally leading the resolution of complex data integration bottlenecks and pipeline failures to protect client timelines and data reliability standards.
  • Quality Enforcement
  • Quality Oversight: Execute and monitor technical data quality standards, ensuring engineering teams adhere to strict policies regarding data lineage, automated quality checks (observability), security/privacy compliance (GDPR/CCPA/PII), and active catalog management.

2. Strategic Growth & Practice Scaling

  • Talent & Scaling Execution: Execute the strategy for data engineering talent acquisition and development within your business units. Implement objective metrics to assess and grow the 'Data-Native' DNA of your teams, ensuring squads are consistently equipped to handle petabyte-scale environments and high-impact delivery.
  • Offerings Alignment: Drive the adoption of standardized regional offerings (e.g., Modern Data Platform, Data Mesh, Lakehouse Implementation). Ensure your teams leverage the profitable frameworks defined by the practice to accelerate time-to-insight and eliminate architectural fragmentation in client environments.
  • Innovation & IP Development: Lead the practical integration of Vector Databases and LLM-ready architectures into project delivery. Champion the hands-on development of IP and reusable accelerators (e.g., automated ingestion engines) that improve delivery speed and enhance data availability across your portfolio.

3. Leadership & Unit Management

  • Unit Leadership: Directly lead, mentor, and manage the Engineering Managers and Lead Architects within your business unit. Hold your teams accountable for project-level operational consistency, technical talent development, and strict adherence to the practice's data governance standards.
  • Stakeholder Communication: Clearly articulate the business unit’s operational performance, technical quality metrics, and delivery progress to the C-suite Stakeholders and regional client leadership, bridging the gap between technical execution and business value.
  • Ecosystem Alignment: Maintain strong technical relationships with key partner contacts (Snowflake, Databricks, AWS/GCP). Align team delivery capabilities with current product roadmaps and ensure squad-level participation in training, certifications, and partner-led enablement opportunities.


Welcome to Searce

The ‘process-first’, AI-native modern tech consultancy that's rewriting the rules.

We don’t do traditional.

As an engineering-led consultancy, we are dedicated to relentlessly improving the real business outcomes. Our solvers co-innovate with clients to futurify operations and make processes smarter, faster & better.


Functional Skills

1. Delivery Management & Operational Excellence

  • Methodology Execution: Expert capability in implementing and enforcing a unified delivery methodology (DataOps, Agile, Mesh Principles) within specific business units. Proven track record of auditing squad-level adherence to ensure consistency across the project lifecycle.
  • Operational Performance: High proficiency in managing day-to-day operational metrics, including squad utilization, resource forecasting, and productivity tracking. Skilled at optimizing team performance to meet profitability and efficiency targets.
  • SOW & Risk Mitigation: Proven experience in operationalizing Statement of Work (SOW) requirements and identifying technical delivery risks early. Expert at mitigating scope creep and data-specific bottlenecks (e.g., latency, ingestion gaps) before they impact client outcomes.
  • Technical Escalation Leadership: Demonstrated ability to lead "war room" efforts to resolve complex pipeline failures or data integrity issues. Skilled at providing clear, rapid remediation plans and communicating technical status directly to regional stakeholders.

2. Architectural Implementation & Technical Oversight

  • Modern Stack Proficiency: Deep, hands-on expertise in implementing Cloud-Native architectures (Lakehouse, Data Mesh, MPP) on Snowflake, Databricks, or hyperscalers. Ability to conduct deep-dive architectural reviews and course-correct design decisions at the squad level to ensure scalability.
  • Operationalizing Governance: Proven experience in embedding data quality and observability (completeness, freshness, accuracy) directly into the CI/CD pipeline. Responsible for technical enforcement of regulatory compliance (GDPR/PII) and maintaining the integrity of data catalogs across active projects.
  • Applied Domain Expertise: Practical experience leading the delivery of high-growth solutions, specifically Generative AI infrastructure (RAG, Vector DBs), Real-Time Streaming, and large-scale platform migrations with a focus on zero-downtime execution.
  • DataOps & Engineering Standards: Expert-level mastery of DataOps, including the setup and management of orchestration frameworks (Airflow, Dagster) and Infrastructure as Code (IaC). You ensure that automation is a baseline requirement, not an afterthought, for all delivery teams.

3. Unit Management & Commercial Execution

  • Unit & Team Management: Proven success in leading and mentoring Engineering Managers and Lead Architects. Responsible for the operational metrics, technical output, and career development of the business unit's talent pool.
  • Offerings Implementation & Scoping: Expertise in translating service offerings (e.g., Data Maturity Assessments, Lakehouse Builds) into accurate project scopes, technical estimates, and resource plans to ensure delivery is both profitable and competitive.
  • Talent Growth & Mentorship: Functional ability to implement growth frameworks for data engineering roles. Focus on hands-on coaching and scaling high-performance technical talent to meet the demands of complex, petabyte-scale environments.
  • Partner Enablement: Functional competence in managing regional technical relationships with major partners (Snowflake, Databricks, GCP/AWS). Drives squad-level certifications, joint technical enablement, and alignment with partner product roadmaps.

Tech Superpowers

  • Modern Data Architect – Reimagines business with the Modern Data Stack (MDS) to deliver data mesh implementations, insights, & real value to clients.
  • End-to-End Ecosystem Thinker – Builds modular, reusable data products across ingestion, transformation (ETL/ELT), governance, and consumption layers.
  • Distributed Compute Savant – Crafts resilient, high-throughput architectures that survive petabyte-scale volume and data skew without breaking the bank.
  • Governance & Integrity Guardian – Embeds data quality, complete lineage, and privacy-by-design (GDPR/PII) into every table, view, and pipeline.
  • AI-Ready Orchestrator – Engineers pipelines that bridge structured data with Unstructured/Vector stores, powering RAG models and Generative AI workflows.
  • Product-Minded Strategist – Balances architectural purity with time-to-insight; treats every dataset as a measurable "Data Product" with clear ROI.
  • Pragmatic Stack Curator – Chooses the simplest tools that compound reliability; fluent in SQL, Python, Spark, dbt, and Cloud Warehouses.
  • Builder @ Heart – Writes, reviews, and optimizes queries daily; proves architectures with cost-performance benchmarks, not slideware. Business-first, data-second, outcome focused technology leader.

Experience & Relevance

  • Executive Experience: Minimum 10+ years of progressive experience in data engineering and analytics, with at least 3 years in a Senior Manager or Director -level role managing multiple technical teams and owning significant operational and efficiency metrics for a large data service line.
  • Delivery Standardization: Demonstrated success in defining and implementing globally consistent, repeatable delivery methodologies (DataOps/Agile Data Warehousing) across diverse teams.
  • Architectural Depth: Must retain deep, current expertise in Modern Data Stack architectures (Lakehouse, MPP, Mesh) and maintain the ability to personally validate high-level architectural and data pipeline design decisions.
  • Operational Leadership: Proven expertise in managing and scaling large professional services organizations, demonstrated ability to optimize utilization, resource allocation, and operational expense.
  • Domain Expertise: Strong background in Enterprise Data Platforms, Applied AI/ML, Generative AI integration, or large-scale Cloud Data Migration.
  • Communication: Exceptional executive-level presentation and negotiation skills, particularly in communicating complex operational, data quality, and governance metrics to C-level stakeholders.

Join the ‘real solvers’

ready to futurify?

If you are excited by the possibilities of what an AI-native engineering-led, modern tech consultancy can do to futurify businesses, apply here and experience the ‘Art of the possible’. Don’t Just Send a Resume. Send a Statement.

Read more
Searce Inc

at Searce Inc

3 recruiters
Tejashree Kokare
Posted by Tejashree Kokare
Bengaluru (Bangalore), Pune, Mumbai
6 - 15 yrs
Best in industry
Google Cloud Platform (GCP)
Data engineering
Data warehouse architecture
Data architecture
Data modeling
+6 more

Solutions Architect - Data Engineering


Modern tech solutions advisory & 'futurify' consulting as a Searce lead fds (‘forward deployed solver’) architecting scalable data platforms and robust data engineering solutions that power intelligent insights and fuel AI innovation.

If you’re a tech-savvy, consultative seller with the brain of a strategist, the heart of a builder, and the charisma of a storyteller — we’ve got a seat for you at the front of the table.

You're not a sales lead. You're the transformation driver.


What are we looking for

real solver?

Solver? Absolutely. But not the usual kind. We're searching for the architects of the audacious & the pioneers of the possible. If you're the type to dismantle assumptions, re-engineer ‘best practices,’ and build solutions that make the future possible NOW, then you're speaking our language.

  • Improver. Solver. Futurist.
  • Great sense of humor.
  • ‘Possible. It is.’ Mindset.
  • Compassionate collaborator. Bold experimenter. Tireless iterator.
  • Natural creativity that doesn’t just challenge the norm, but solves to design what’s better.
  • Thinks in systems. Solves at scale.


This Isn’t for Everyone. But if you’re the kind who questions why things are done a certain way— and then identifies 3 better ways to do it — we’d love to chat with you.


Your Responsibilities

what you will wake up to solve.


You are not just a Solutions Architect; you are a futurifier of our data universe and the primary enabler of our AI ambitions. With a deep-seated passion for data engineering, you will architect and build the foundational data infrastructure that powers the customers entire data intelligence ecosystem.

As the Directly Responsible Individual (DRI) for our enterprise-grade data platforms, you own the outcome, end-to-end. You are the definitive solver for our customer's most complex data challenges, leveraging a powerful tech stack including Snowflake, Databricks, etc. and core GCP & AWS services (BigQuery, Spanner, Airflow, Kafka). This is a hands-on-keys role where you won't just design solutions—you'll build them, break them, and perfect them.


  • Solution Design & Pre-sales Excellence:Collaborate with cross-functional teams, including sales, engineering, and operations, to ensure successful project delivery.
  • Design Core Data Engineering: Master data modeling, architecting high-performance data ingestion pipelines and ensuring data quality and governance throughout the data lifecycle.
  • Enable Cloud & AI: Design and implement solutions utilizing core GCP data services, building foundational data platforms that efficiently support advanced analytics and AI/ML initiatives.
  • Optimize Performance & Cost: Continuously optimize data architectures and implementations for performance, efficiency, and cost-effectiveness within the cloud environment.
  • Bridge Business & Tech: Translate complex business requirements into clear technical designs, providing technical leadership and guidance to data engineering teams.
  • Stay Ahead of the Curve: Continuously research and evaluate new data technologies, architectural patterns, and industry trends to keep our data platforms at the cutting edge.


Functional Skills:


  • Enterprise Data Architecture Design: Expert ability to design holistic, scalable, and resilient data architectures for complex enterprise environments.
  • Cloud Data Platform Strategy: Proven capability to strategize, design, and implement cloud-native data platforms.
  • Pre-Sales & Technical Storyteller: Crafts compelling, client-ready proposals, architectural decks, and technical demonstrations. Doesn't just present; shapes the strategic technical narrative behind every proposed solution.
  • Advanced Data Modelling: Mastery in designing various data models for analytical, operational, and transactional use cases.
  • Data Ingestion & Pipeline Orchestration: Strong expertise in designing and optimizing robust data ingestion and transformation pipelines.
  • Stakeholder Communication: Exceptional skills in articulating complex technical concepts and architectural decisions to both technical and non-technical stakeholders.
  • Performance & Cost Optimization: Adept at optimizing data solutions for performance, efficiency, and cost within a cloud environment.


Tech Superpowers:


  • Cloud Data Mastery: You're a wizard at leveraging public cloud data services, with deep expertise in GCP (BigQuery, Spanner, etc.) and expert proficiency in modern data warehouse solutions like Snowflake.
  • Data Engineering Core: Highly skilled in designing, implementing, and managing data workflows using tools like Apache Airflow and Apache Kafka. You're also an authority on advanced data modeling and ETL/ELT patterns.
  • AI/ML Data Foundation: You instinctively design data pipelines and structures that efficiently feed and empower Machine Learning and Artificial Intelligence applications.
  • Programming for Data: You have a strong command over key programming languages (Python, SQL) for scripting, automation, and building data processing applications.


Experience & Relevance:


  • Architectural Leadership (8+ Years): You bring extensive experience (7+ years) specifically in a Solutions Architect role, focused on data engineering and platform building.
  • Cloud Data Expertise: You have a proven track record of designing and implementing production-grade data solutions leveraging major public cloud platforms, with significant experience in Google Cloud Platform (GCP).
  • Data Warehousing & Data Platform: Demonstrated hands-on experience in the end-to-end design, implementation, and optimization of modern data warehouses and comprehensive data platforms.
  • Databricks & BigQuery Mastery: You possess significant practical experience with Databricks as a core data warehouse and GCP BigQuery for analytical workloads.
  • Data Ingestion & Orchestration: Proven experience designing and implementing complex data ingestion pipelines and workflow orchestration using tools like Airflow and real-time streaming technologies like Kafka.
  • AI/ML Data Enablement: Experience in building data foundations specifically geared towards supporting Machine Learning and Artificial Intelligence initiatives.


Join the ‘real solvers’

ready to futurify?

If you are excited by the possibilities of what an AI-native engineering-led, modern tech consultancy can do to futurify businesses, apply here and experience the ‘Art of the possible’.


Don’t Just Send a Resume. Send a Statement.


So, If you are passionate about tech, future & what you read above (we really are!), apply here to experience the ‘Art of Possible’

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Dhruti Parikh
Posted by Dhruti Parikh
Bengaluru (Bangalore), Mumbai, Pune
10 - 18 yrs
₹15L - ₹35L / yr
Google Cloud Platform (GCP)
Iaac
IAM
GKE
Iac

Job Summary: 

We are looking for an experienced GCP Cloud Engineer to design, implement, and manage cloud-based solutions on Google Cloud Platform (GCP). The ideal candidate should have expertise in GKE (Google Kubernetes Engine), Cloud Run, Cloud Loadbalancer, Cloud function, Azure DevOps, and Terraform, with a strong focus on automation, security, and scalability. 

You will work closely with development, operations, and security teams to ensure robust cloud infrastructure and CI/CD pipelines while optimizing performance and cost. 

 

Key Responsibilities: 

1. Cloud Infrastructure Design & Management 

  • Architect, deploy, and maintain GCP cloud resources via terraform/other automation. 
  • Implement Google Cloud Storage, Cloud SQL, filestore, for data storage and processing needs. 
  • Manage and configure Cloud Load Balancers (HTTP(S), TCP/UDP, and SSL Proxy) for high availability and scalability. 
  • Optimize resource allocation, monitoring, and cost efficiency across GCP environments. 

2. Kubernetes & Container Orchestration 

  • Deploy, manage, and optimize workloads on Google Kubernetes Engine (GKE). 
  • Work with Helm charts for microservices deployments. 
  • Automate scaling, rolling updates, and zero-downtime deployments. 

 

3. Serverless & Compute Services 

  • Deploy and manage applications on Cloud Run and Cloud Functions for scalable, serverless workloads. 
  • Optimize containerized applications running on Cloud Run for cost efficiency and performance. 

 

4. CI/CD & DevOps Automation 

  • Design, implement, and manage CI/CD pipelines using Azure DevOps. 
  • Automate infrastructure deployment using Terraform, Bash and Powershell scripting 
  • Integrate security and compliance checks into the DevOps workflow (DevSecOps). 

 

 

Required Skills & Qualifications: 

✔ Experience: 4+ years in Cloud Engineering, with a focus on GCP. 

✔ Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions). 

✔ Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm. 

✔ DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation. 

✔ Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources. 

✔ Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation. 

✔ Security & Compliance: Knowledge of cloud security principles, IAM, and compliance standards. 

Read more
Quantiphi

at Quantiphi

3 candid answers
1 video
Nikita Sinha
Posted by Nikita Sinha
Bengaluru (Bangalore), Mumbai, Trivandrum
4 - 7 yrs
Upto ₹30L / yr (Varies
)
Google Cloud Platform (GCP)
SQL
ETL
Datawarehousing
Data-flow analysis

We are looking for a skilled Data Engineer / Data Warehouse Engineer to design, develop, and maintain scalable data pipelines and enterprise data warehouse solutions. The role involves close collaboration with business stakeholders and BI teams to deliver high-quality data for analytics and reporting.


Key Responsibilities

  • Collaborate with business users and stakeholders to understand business processes and data requirements
  • Design and implement dimensional data models, including fact and dimension tables
  • Identify, design, and implement data transformation and cleansing logic
  • Build and maintain scalable, reliable, and high-performance ETL/ELT pipelines
  • Extract, transform, and load data from multiple source systems into the Enterprise Data Warehouse
  • Develop conceptual, logical, and physical data models, including metadata, data lineage, and technical definitions
  • Design, develop, and maintain ETL workflows and mappings using appropriate data load techniques
  • Provide high-level design, research, and effort estimates for data integration initiatives
  • Provide production support for ETL processes to ensure data availability and SLA adherence
  • Analyze and resolve data pipeline and performance issues
  • Partner with BI teams to design and develop reports and dashboards while ensuring data integrity and quality
  • Translate business requirements into well-defined technical data specifications
  • Work with data from ERP, CRM, HRIS, and other transactional systems for analytics and reporting
  • Define and document BI usage through use cases, prototypes, testing, and deployment
  • Support and enhance data governance and data quality processes
  • Identify trends, patterns, anomalies, and data quality issues, and recommend improvements
  • Train and support business users, IT analysts, and developers
  • Lead and collaborate with teams spread across multiple locations

Required Skills & Qualifications

  • Bachelor’s degree in Computer Science or a related field, or equivalent work experience
  • 3+ years of experience in Data Warehousing, Data Engineering, or Data Integration
  • Strong expertise in data warehousing concepts, tools, and best practices
  • Excellent SQL skills
  • Strong knowledge of relational databases such as SQL Server, PostgreSQL, and MySQL
  • Hands-on experience with Google Cloud Platform (GCP) services, including:
  1. BigQuery
  2. Cloud SQL
  3. Cloud Composer (Airflow)
  4. Dataflow
  5. Dataproc
  6. Cloud Functions
  7. Google Cloud Storage (GCS)
  • Experience with Informatica PowerExchange for Mainframe, Salesforce, and modern data sources
  • Strong experience integrating data using APIs, XML, JSON, and similar formats
  • In-depth understanding of OLAP, ETL frameworks, Data Warehousing, and Data Lakes
  • Solid understanding of SDLC, Agile, and Scrum methodologies
  • Strong problem-solving, multitasking, and organizational skills
  • Experience handling large-scale datasets and database design
  • Strong verbal and written communication skills
  • Experience leading teams across multiple locations

Good to Have

  • Experience with SSRS and SSIS
  • Exposure to AWS and/or Azure cloud platforms
  • Experience working with enterprise BI and analytics tools

Why Join Us

  • Opportunity to work on large-scale, enterprise data platforms
  • Exposure to modern cloud-native data engineering technologies
  • Collaborative environment with strong stakeholder interaction
  • Career growth and leadership opportunities
Read more
Mumbai
3 - 6 yrs
₹7L - ₹15L / yr
skill iconDocker
skill iconKubernetes
DevOps
Google Cloud Platform (GCP)

Lightning Job By Cutshort ⚡

 

As part of this feature, you can expect status updates about your application and replies within 72 hours (once the screening questions are answered)


Job Overview:


We are seeking an experienced DevOps Engineer to join our team. The successful candidate will be responsible for designing, implementing, and maintaining the infrastructure and software systems required to support our development and production environments. The ideal candidate should have a strong background in Linux, GitHub, Actions/Jenkins, ArgoCD, AWS, Kubernetes, Helm, Datadog, MongoDB, Envoy Proxy, Cert-Manager, Terraform, ELK, Cloudflare, and BigRock.


Kindly apply at https://wohlig.keka.com/careers/jobdetails/54566


Responsibilities:


• Design, implement and maintain CI/CD pipelines using GitHub, Actions/Jenkins, Kubernetes, Helm, and ArgoCD.

• Deploy and manage Kubernetes clusters using AWS.

• Configure and maintain Envoy Proxy and Cert-Manager to automate deployment and manage application environments.

• Monitor system performance using Datadog, ELK, and Cloudflare tools.

• Automate infrastructure management and maintenance tasks using Terraform, Ansible, or similar tools.

• Collaborate with development teams to design, implement and test infrastructure changes.

• Troubleshoot and resolve infrastructure issues as they arise.

• Participate in on-call rotation and provide support for production issues.


Qualifications:

• Bachelor's or Master's degree in Computer Science, Engineering or a related field.

• 3+ years of experience in DevOps engineering with a focus on Linux, GitHub, Actions/CodeFresh, ArgoCD, AWS, Kubernetes, Helm, Datadog, MongoDB, Envoy Proxy, Cert-Manager, Terraform, ELK, Cloudflare, and BigRock.

• Strong understanding of Linux administration and shell scripting.

• Experience with automation tools such as Terraform, Ansible, or similar.

• Ability to write infrastructure as code using tools such as Terraform, Ansible, or similar.

• Experience with container orchestration platforms such as Kubernetes.

• Familiarity with container technologies such as Docker.

• Experience with cloud providers such as AWS.

• Experience with monitoring tools such as Datadog and ELK.



Skills:

• Strong analytical and problem-solving skills.

• Excellent communication and collaboration skills.

• Ability to work independently or in a team environment.

• Strong attention to detail.

• Ability to learn and apply new technologies quickly.

• Ability to work in a fast-paced and dynamic environment.

• Strong understanding of DevOps principles and methodologies.

Read more
AI Industry

AI Industry

Agency job
via Peak Hire Solutions by Dharati Thakkar
Mumbai, Bengaluru (Bangalore), Hyderabad, Gurugram
6 - 10 yrs
₹32L - ₹42L / yr
ETL
SQL
Google Cloud Platform (GCP)
Data engineering
ELT
+17 more

Role & Responsibilities:

We are looking for a strong Data Engineer to join our growing team. The ideal candidate brings solid ETL fundamentals, hands-on pipeline experience, and cloud platform proficiency — with a preference for GCP / BigQuery expertise.


Responsibilities:

  • Design, build, and maintain scalable data pipelines and ETL/ELT workflows
  • Work with Dataform or DBT to implement transformation logic and data models
  • Develop and optimize data solutions on GCP (BigQuery, GCS) or AWS/Azure
  • Support data migration initiatives and data mesh architecture patterns
  • Collaborate with analysts, scientists, and business stakeholders to deliver reliable data products
  • Apply data governance and quality best practices across the data lifecycle
  • Troubleshoot pipeline issues and drive proactive monitoring and resolution


Ideal Candidate:

  • Strong Data Engineer Profile
  • Must have 6+ years of hands-on experience in Data Engineering, with strong ownership of end-to-end data pipeline development.
  • Must have strong experience in ETL/ELT pipeline design, transformation logic, and data workflow orchestration.
  • Must have hands-on experience with any one of the following: Dataform, dbt, or BigQuery, with practical exposure to data transformation, modeling, or cloud data warehousing.
  • Must have working experience on any cloud platform: GCP (preferred), AWS, or Azure, including object storage (GCS, S3, ADLS).
  • Must have strong SQL skills with experience in writing complex queries and optimizing performance.
  • Must have programming experience in Python and/or SQL for data processing.
  • Must have experience in building and maintaining scalable data pipelines and troubleshooting data issues.
  • Exposure to data migration projects and/or data mesh architecture concepts.
  • Experience with Spark / PySpark or large-scale data processing frameworks.
  • Experience working in product-based companies or data-driven environments.
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or related field.


NOTE:

  • There will be an interview drive scheduled on 28th and 29th March 2026, and if shortlisted, they will be expected to be available on these Interview dates. Only Immediate joiners are considered.
Read more
Recruiting Bond

at Recruiting Bond

2 candid answers
Pavan Kumar
Posted by Pavan Kumar
Bengaluru (Bangalore), Mumbai
10 - 16 yrs
₹75L - ₹130L / yr
Distributed Systems
Microservices
Enterprise architecture
System Design & Architecture
Event-Driven Architecture
+29 more

🚨 We’re Building a “Top 1% Engineering Org”


We’re building a high-talent-density, AI-first R&D organization from scratch — inside a publicly listed company undergoing a full-scale transformation.

Think:

→ Rewriting legacy systems into AI-native architectures

→ Embedding LLMs + Agentic AI into core workflows

→ Reimagining platforms, infra, and data systems for the next decade

This is the kind of shift you’d expect from Google, Microsoft, or Meta —

Except you get to build it from day 0 → scale it globally.


About the Role / Team

We are building a next-generation AI-first R&D organization in Bengaluru, focused on solving complex problems across LLMs, Agentic AI systems, distributed computing, and enterprise-scale architectures.


This initiative is part of a publicly listed global company investing heavily in AI-driven transformation, re-architecting its platforms into intelligent, autonomous systems powered by large language models, workflows, and decision engines.


You will be working on:

  • Agentic AI systems & LLM-powered workflows
  • Distributed, scalable backend systems
  • Enterprise-grade AI platforms
  • Automation-first engineering environments


🚀 The Mandate

Own and evolve the technical backbone of an AI-first enterprise platform.


You will define architecture across LLM-powered systems, distributed services, and data platforms — and lead critical transformations from legacy → AI-native systems.


🧩 What You’ll Do

  • Architect large-scale distributed systems powering AI-driven workflows
  • Lead 0→1 and 1→N platform builds (LLM integrations, agentic systems, orchestration layers)
  • Redesign legacy systems into scalable, modular, AI-native architectures
  • Drive system design excellence across teams (APIs, infra, observability, reliability)
  • Make high-stakes decisions on trade-offs (latency, cost, scalability, model performance)
  • Mentor senior engineers and influence engineering culture/org standards
  • Partner with product, data, and leadership on long-term technical strategy


🧠 What We’re Looking For

  • Proven track record building high-scale backend or platform systems
  • Deep expertise in distributed systems, microservices, cloud (AWS/GCP/Azure)
  • Strong exposure to data systems/infra / Data / real-time architectures
  • Experience or strong interest in LLMs, GenAI, or AI system design
  • Exceptional system design, abstraction, and problem-solving ability
  • High ownership mindset — you think in terms of systems, not tickets
  • Strong coding skills in Python / Java / Go / Node.js
  • Solid understanding of data structures, system design basics, and backend architecture
  • Experience building scalable APIs and services
  • Familiarity or curiosity around AI/LLMs, async systems, or event-driven design
  • Strong debugging, problem-solving, and ownership mindset
  • Solve hard system problems (latency, scale, reliability)
  • Drive cross-team technical decisions and standards
  • Mentor senior engineers and influence org-wide architecture 
  • Design large-scale distributed systems and backend platforms
  • Mentorship & Technical Leadership 
  • Expertise in system design, scalability, and performance optimization


Nice to Have

  • Experience integrating LLMs, vector databases, or AI pipelines
  • Contributions to architecture at scale
  • Experience with Agentic AI / LLM orchestration frameworks
  • Background in product engineering or platform companies
  • Exposure to global-scale systems (millions of users / high throughput)


🔥 What Sets You Apart

  • Built platforms used by millions of users / high-throughput systems
  • Experience with event-driven systems, stream processing, or infra platforms
  • Prior work on AI/ML platforms, model serving, or intelligent systems
Read more
NeoGenCode Technologies Pvt Ltd
Mumbai
5 - 10 yrs
₹12L - ₹24L / yr
DevOps
skill iconAmazon Web Services (AWS)
Windows Azure
Google Cloud Platform (GCP)
skill iconKubernetes
+12 more

Job Title : Senior DevOps Engineer (Only Mumbai Candidates)

Experience : 5+ Years

Location : Mumbai (On-site)

Notice Period : Immediate to 15 Days

Interview Process : 1 Internal Round + 1 Client Round


Mandatory Skills :

Multi-Cloud (AWS/GCP/Azure – any two), Kubernetes, Terraform, Helm (writing Helm Charts), CI/CD (GitLab CI/Jenkins/GitHub Actions), GitOps (ArgoCD/FluxCD), Multi-tenant deployments, Stateful microservices on Kubernetes, Enterprise Linux.


Role Overview :

We are looking for a Senior DevOps Engineer to design, build, and manage scalable cloud infrastructure and DevOps pipelines for product-based platforms.

The ideal candidate should have strong experience with Kubernetes, Terraform, Helm Charts, CI/CD, and GitOps practices.


Key Responsibilities :

  • Design and manage scalable cloud infrastructure across AWS/GCP/Azure.
  • Deploy and manage microservices on Kubernetes clusters.
  • Build and maintain Infrastructure as Code using Terraform and Helm.
  • Implement CI/CD pipelines using GitLab CI, Jenkins, or GitHub Actions.
  • Implement GitOps workflows using ArgoCD or FluxCD.
  • Ensure secure, scalable, and reliable DevOps architecture.
  • Implement monitoring and logging using Prometheus, Grafana, or ELK.

Good to Have :

  • Packer, OpenShift/Rancher/K3s, On-prem deployments, PaaS experience, scripting (Bash/Python), Terraform modules.
Read more
Wohlig Transformations Pvt Ltd
Apoorva Lakshkar
Posted by Apoorva Lakshkar
Mumbai
7 - 10 yrs
₹15L - ₹23L / yr
Google Cloud Platform (GCP)
skill iconAmazon Web Services (AWS)
DevOps
skill iconKubernetes

Job Overview 


We are seeking an experienced Senior Solution Architect to join our dynamic DevOps organization. The ideal candidate will have a strong background in cloud technologies, with expertise in migration projects across platforms such as GCP, AWS, and Azure. The candidate should possess a deep understanding of DevOps principles, Kubernetes orchestration, Data migration & management and automation tools like CI/CD pipelines and Terraform.The individual should be highly skilled in designing scalable application architectures capable of handling substantial workloads while ensuring the highest standards of quality.


Key Responsibilities 


  • Lead and drive cloud migration projects from on-premises data centers or other cloud platforms to GCP, AWS, or Azure.
  • Design and implement migration strategies that ensure minimal downtime and maximum efficiency.
  • Demonstrate proficiency in GCP, AWS, and Azure, with the ability to choose and optimize solutions based on specific business requirements.
  • Provide guidance on selecting the appropriate cloud services for various workloads.
  • Design, implement, and optimize CI/CD pipelines to streamline software delivery.
  • Utilize Terraform for infrastructure as code (IaC) to automate deployment processes.
  • Collaborate with development and operations teams to enhance the overall DevOps culture.
  • Possess in-depth knowledge and practical experience with Kubernetes orchestration for containerized applications.
  • Architect and optimize Kubernetes clusters for high availability and scalability.
  • Engage in research and development activities to stay abreast of industry trends and emerging technologies.
  • Evaluate and introduce new tools and methodologies to enhance the efficiency and effectiveness of cloud solutions.
  • Architect solutions that can handle large-scale workloads and provide guidance on scaling strategies.
  • Ensure high-performance levels and reliability in production environments.
  • Design scalable and high-performance database architectures tailored to meet business needs.
  • Execute database migrations with a keen focus on data consistency, integrity, and performance.
  • Develop and implement database pipelines to automate processes such as data migrations, schema changes, and backups.
  • Optimize database workflows to enhance efficiency and reliability.
  • Work closely with clients to assess and enhance the quality of existing architectures.
  • Implement best practices to ensure robust, secure, and well-architected solutions.
  • Drive migration projects, collaborating with cross-functional teams to ensure successful execution.
  • Provide technical leadership and mentorship to junior team members.


Required Skills and Qualifications: 


  • Bachelor's degree in Computer Science, Information Technology, or related field.
  • Relevant industry experience in a Solution Architect role.
  • Proven experience in leading cloud migration projects across GCP, AWS, and Azure.
  • Expertise in DevOps practices, CI/CD pipelines, and infrastructure automation.
  • In-depth knowledge of Kubernetes and container orchestration.
  • Strong background in scaling architectures to handle significant workloads.
  • Sound knowledge in database migrations
  • Excellent communication skills and the ability to articulate complex technical concepts to both technical and non-technical stakeholders.


Read more
Recruiting Bond

at Recruiting Bond

2 candid answers
Pavan Kumar
Posted by Pavan Kumar
Mumbai, Navi Mumbai
10 - 15 yrs
₹55L - ₹80L / yr
Distributed Systems
Systems design
Systems architecture
High-level design
LLD
+77 more

Location: Mumbai, Maharashtra, India

Sector: Technology, Information & Media

Company Size: 500 - 1,000 Employees

Employment: Full-Time, Permanent

Experience: 10 - 14 Years (Engineering Leadership)

Level: Engineering Manager / Group EM


ABOUT THIS MANDATE :


Recruiting Bond has been exclusively retained by one of India's most prominent and well-established digital platform organisations operating at the intersection of Technology, Information, and Media to identify and place an exceptional Engineering Manager who can lead engineering teams through an enterprise-wide AI adoption and digital transformation agenda.


This is a high-impact, hands-on leadership role at the nexus of people, product, and technology. The organisation is executing one of the most ambitious AI transformation programmes in its sector and this Engineering Manager will be a core driver of that change. You will lead multiple squads, own engineering delivery end-to-end, embed AI tooling and practices into the team's DNA, and shape the engineering culture of tomorrow.


We are seeking leaders who code when it matters, who build systems and teams with equal conviction, and who view AI not as a trend but as a fundamental shift in how great software is built.


THE OPPORTUNITY AT A GLANCE :


AI-First Engineering Culture :

  • Own AI adoption across your squads - from LLM tooling integration to automation-first delivery workflows. Make AI a default, not an afterthought.


Hands-On Engineering Leadership :

  • Stay close to the code. Lead architecture reviews, unblock engineers, and set the technical bar - not just the management agenda.


People & Org Builder :

  • Grow engineers into leaders. Build squads of 615 across functions. Drive hiring, career frameworks, and a culture of psychological safety.


KEY RESPONSIBILITIES :


1. Hands-On Technical Engagement :

  • Remain deeply embedded in the technical work participate in design reviews, architecture decisions, and critical code reviews
  • Set and uphold the engineering quality bar : performance benchmarks, security standards, test coverage, and release quality
  • Provide technical direction on backend platform strategy, API design, service decomposition, and data architecture
  • Identify and resolve systemic technical debt and architectural risks across team-owned services
  • Unblock engineers by diving into complex problems debugging, pair programming, and system analysis when it matters
  • Own key technical decisions in collaboration with Tech Leads and Principal Engineers; balance pragmatism with long-term sustainability


2. AI Adoption, Integration & Transformation (2026 Mandate) :

  • Define and execute the team's AI adoption roadmap - from developer tooling to product-facing AI features
  • Champion the integration of GenAI tools (GitHub Copilot, Cursor, Claude, ChatGPT) across the full engineering workflow coding, testing, documentation, incident response
  • Embed LLM-powered capabilities into the product : recommendation engines, intelligent search, conversational interfaces, content generation, and predictive systems
  • Lead evaluation and adoption of AI-assisted SDLC practices : automated code review, AI-generated test suites, intelligent observability, and anomaly detection
  • Partner with Data Science and ML Platform teams to productionise ML models with robust MLOps pipelines
  • Build team literacy in prompt engineering, RAG (Retrieval-Augmented Generation), and AI agent frameworks
  • Create an experimentation culture : run structured AI pilots, measure productivity impact, and scale what works
  • Stay ahead of the AI tooling landscape and advise senior leadership on strategic AI investments and engineering implications


3. People Leadership & Team Development :

  • Lead, manage, and grow squads of 6 - 15 engineers across seniority levels (L2 through L6 / Junior through Staff)
  • Conduct structured 1 : 1s, career growth conversations, and development planning with every direct report
  • Design and execute personalised AI upskilling programmes ensure every engineer develops practical AI fluency by end of 2026
  • Build and maintain a high-performance team culture : clarity of ownership, accountability, fast feedback loops, and psychological safety
  • Drive performance management fairly and rigorously recognise top performers, manage underperformance constructively
  • Lead technical hiring end-to-end : define job requirements, conduct bar-raising interviews, and make data-driven hire decisions
  • Contribute to engineering career frameworks and level definitions in partnership with the VP / Director of Engineering


4. Engineering Delivery & Execution Excellence :

  • Own end-to-end delivery for multiple product squads from planning and scoping through production release and post-launch stability
  • Implement and refine agile delivery frameworks (Scrum, Kanban, Shape Up) calibrated to squad needs and product cadence
  • Drive predictable delivery : maintain healthy sprint velocity, manage WIP limits, and ensure dependency resolution across teams.
  • Establish and own engineering KPIs : DORA metrics (deployment frequency, lead time, MTTR, change failure rate), uptime SLOs, and velocity trends
  • Lead incident management : build blameless post-mortem culture, own RCA processes, and drive systemic reliability improvements
  • Balance technical debt repayment with feature velocity negotiate prioritisation transparently with Product leadership


5. Strategic Leadership & Cross-Functional Influence :

  • Serve as the primary engineering partner for Product, Design, Data, and Business stakeholders translate ambiguity into executable engineering plans
  • Participate in quarterly roadmap planning, capacity forecasting, and OKR definition for engineering teams
  • Represent engineering in leadership forums articulate technical constraints, risks, and opportunities in business terms
  • Contribute to org-wide engineering strategy : platform investments, build-vs-buy decisions, and shared infrastructure priorities
  • Build relationships across geographies (Mumbai HQ + distributed teams) to maintain alignment and delivery cohesion
  • Act as a culture carrier and ambassador for engineering excellence, innovation, and responsible AI use


AI TRANSFORMATION LEADERSHIP 2026 EXPECTATIONS :


In 2026, Engineering Managers at this organisation are expected to be active architects of AI transformation not passive observers. The following outlines the specific AI leadership expectations for this role :


AI Developer Productivity

  • Drive measurable uplift in developer velocity through AI tooling adoption. Target : 30%+ reduction in code review cycle time and 40%+ increase in test coverage automation by Q3 2026.


LLM & GenAI Product Features

  • Own delivery of GenAI-powered product capabilities : intelligent content, semantic search, personalisation, and conversational UX in production, at scale.


AI-Augmented Observability

  • Implement AI-driven monitoring and anomaly detection pipelines. Reduce MTTR by leveraging predictive alerting, intelligent runbooks, and auto-remediation scripts.


Team AI Fluency :

  • Build mandatory AI literacy across all engineering levels.
  • Every engineer understands prompt engineering basics, AI ethics guardrails, and responsible AI deployment practices.


Responsible AI Governance :

  • Partner with Security, Legal, and Data Privacy to ensure all AI deployments meet compliance standards, bias mitigation requirements, and explainability benchmarks.


TECHNOLOGY STACK & DOMAIN FAMILIARITY REQUIRED :


  • Languages: Java/ Go/ Python/ Node.js /PHP /Rust (must be hands-on in at least 2)
  • Cloud: AWS / GCP / Azure (multi-cloud exposure strongly preferred)
  • AI & GenAI: OpenAI / Anthropic / Gemini APIs /LangChain /LlamaIndex / RAG / Vector DBs / GitHub
  • Copilot: Cursor /Hugging Face
  • Containers: Docker /Kubernetes /Helm /Service Mesh (Istio / Linkerd)
  • Databases: PostgreSQL /MongoDB / Redis / Cassandra / Elasticsearch / Pinecone (Vector DB)
  • Messaging: Apache Kafka /RabbitMQ /AWS SQS/SNS /Google Pub/Sub
  • MLOps & DataOps: MLflow /Kubeflow / SageMaker / Vertex AI /Airflow /dbt
  • Observability: Datadog /Prometheus /Grafana /OpenTelemetry / Jaeger /ELK Stack
  • CI/CD & IaC: GitHub Actions ArgoCD / Jenkins / Terraform /Ansible /Backstage (IDP)


QUALIFICATIONS & CANDIDATE PROFILE :

Education :

  • B.E. / B.Tech or M.E. / M.Tech from a Tier-I or Tier-II Institution - CS, IS, ECE, AI/ML streams strongly preferred
  • Demonstrated engineering depth and leadership impact may complement institution pedigree


Experience :

  • 10 to 14 years of progressive engineering experience, with at least 3 years in a formal Engineering Manager or equivalent people-leadership role
  • Proven track record of managing and scaling engineering teams (615+ engineers) in a fast-growing SaaS or digital product environment
  • Hands-on backend engineering background must be able to read, write, and critique production code
  • Direct experience driving AI/ML feature delivery or AI tooling adoption within engineering organisations
  • Exposure across start-up, mid-size, and large-scale product organisations, preferred adaptability is a core requirement
  • Strong CS fundamentals: distributed systems, algorithms, system design, and software architecture
  • Demonstrated career stability minimum of 2 years of average tenure per organisation.


The Ideal Engineering Manager in 2026 :

  • Leads with context, not control, empowers engineers while maintaining accountability and quality
  • Is fluent in both people language and technical language, switches registers naturally with engineers and executives alike
  • Sees AI as a force multiplier for the team, not a threat. Actively experiments with and advocates for AI tooling
  • Measures success by team outcomes, not personal output. Takes pride in what the team ships, not what they build alone
  • Creates feedback loops obsessively between product and engineering, between seniors and juniors, between metrics and decisions
  • Has strong opinions, loosely held, brings conviction to discussions but updates on evidence
  • Invests in engineering excellence as seriously as delivery velocity knows that quality and speed are not opposites


WHY THIS ROLE STANDS APART :


AI Transformation at Scale :

  • Lead one of the most significant AI adoption programmes in India's digital media sector.
  • Our decisions will shape how hundreds of engineers work in 2026 and beyond.


Hands-On & Strategic Balance :

  • A rare EM role that actively encourages technical depth.
  • Stay close to the code while owning the people agenda - the best of both worlds.


Established Platform, Real Scale :

  • 5001,000 engineers, proven product-market fit, and the org maturity to execute.
  • This is not a greenfield startup gamble it is a serious company with serious ambition.


Clear Leadership Growth Path :

  • A visible, direct path toward Director / VP of Engineering.
  • Senior leadership is invested in growing its next generation of technology executives.


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Mumbai, Pune
7 - 13 yrs
Best in industry
skill iconJava
skill iconSpring Boot
Microservices
RESTful APIs
skill iconAmazon Web Services (AWS)
+1 more

JOB DESCRIPTION:


Location: Pune, Mumbai

Mode of Work : 3 days from Office


DSA(Collections, Hash maps, trees, Linkedlist and Arrays, etc), Core Oops concepts(Multithreading, Multi Processing, Polymorphism, Inheritance etc) Annotations in Spring and Spring boot, Java 8 Vital features, database Optimization, Microsevices and Rest API

  • Design, develop, and maintain low-latency, high-performance enterprise applications using Core Java (Java 5.0 and above).
  • Implement and integrate APIs using Spring Framework and Apache CXF.
  • Build microservices-based architecture for scalable and distributed systems.
  • Collaborate with cross-functional teams for high/low-level design, development, and deployment of software solutions.
  • Optimize performance through efficient multithreading, memory management, and algorithm design.
  • Ensure best coding practices, conduct code reviews, and perform unit/integration testing.
  • Work with RDBMS (preferably Sybase) for backend data integration.
  • Analyze complex business problems and deliver innovative technology solutions in the financial/trading domain.
  • Work in Unix/Linux environments for deployment and troubleshooting.
Read more
Digital transformation excellence provider

Digital transformation excellence provider

Agency job
via Peak Hire Solutions by Dharati Thakkar
Mumbai
12 - 20 yrs
₹30L - ₹40L / yr
Product Management
Business-to-business
Analytics
Product engineering
Procurement management
+26 more

 JOB DETAILS:

* Job Title: Head of Engineering/Senior Product Manager

* Industry: Digital transformation excellence provider

* Salary: Best in Industry

* Experience: 12-20 years

* Location: Mumbai

 

Job Description

Role Overview

The VP / Head of Technology will lead company’s technology function across engineering, product development, cloud infrastructure, security, and AI-led initiatives. This role focuses on delivering scalable, high-quality technology solutions across company’s core verticals including eCommerce, Procurement & e-Sourcing, ERP integrations, Sustainability/ESG, and Business Services.

This leader will drive execution, ensure technical excellence, modernize platforms, and collaborate closely with business and delivery teams.

 

Roles and Responsibilities:

Technology Execution & Architecture Leadership

·        Own and execute the technology roadmap aligned with business goals.

·        Build and maintain scalable architecture supporting multiple verticals.

·        Enforce engineering best practices, code quality, performance, and security.

·        Lead platform modernization including microservices, cloud-native architecture, API-first systems, and integration frameworks.

 

Product & Engineering Delivery

·        Manage multi-product engineering teams across eCommerce platforms, procurement systems, ERP integrations, analytics, and ESG solutions.

·        Own the full SDLC — requirements, design, development, testing, deployment, support.

·        Implement Agile, DevOps, CI/CD for faster releases and improved reliability.

·        Oversee product/platform interoperability across all company systems.

 

Vertical-Specific Technology Leadership

Procurement Tech:

·        Lead architecture and enhancements of procurement and indirect spend platforms.

·        Ensure interoperability with SAP Ariba, Coupa, Oracle, MS Dynamics, etc.

 

eCommerce:

·        Drive development of scalable B2B/B2C commerce platforms, headless commerce, marketplace integrations, and personalization capabilities.

 

Sustainability/ESG:

·        Support development of GHG tracking, reporting systems, and sustainability analytics platforms.

 

Business Services:

·        Enhance operational platforms with automation, workflow management, dashboards, and AI-driven efficiency tools.

 

Data, Cloud, Security & Infrastructure

·        Own cloud infrastructure strategy (Azure/AWS/GCP).

·        Ensure adherence to compliance standards (SOC2, ISO 27001, GDPR).

·        Lead cybersecurity policies, monitoring, threat detection, and recovery planning.

·        Drive observability, cost optimization, and system scalability.

 

AI, Automation & Innovation

·        Integrate AI/ML, analytics, and automation into product platforms and service delivery.

·        Build frameworks for workflow automation, supplier analytics, personalization, and operational efficiency.

·        Lead R&D for emerging tech aligned to business needs.

 

Leadership & Team Management

·        Lead and mentor engineering managers, architects, developers, QA, and DevOps.

·        Drive a culture of ownership, innovation, continuous learning, and performance accountability.

·        Build capability development frameworks and internal talent pipelines.

 

Stakeholder Collaboration

·        Partner with Sales, Delivery, Product, and Business Teams to align technology outcomes with customer needs.

·        Ensure transparent reporting on project status, risks, and technology KPIs.

·        Manage vendor relationships, technology partnerships, and external consultants.

 

Education, Training, Skills, and Experience Requirements:

Experience & Background

·        16+ years in technology execution roles, including 5–7 years in senior leadership.

·        Strong background in multi-product engineering for B2B platforms or enterprise systems.

·        Proven delivery experience across: eCommerce, ERP integrations, procurement platforms, ESG solutions, and automation.

 

Technical Skills

·        Expertise in cloud platforms (Azure/AWS/GCP), microservices architecture, API frameworks.

·        Strong grasp of procurement tech, ERP integrations, eCommerce platforms, and enterprise-scale systems.

·        Hands-on exposure to AI/ML, automation tools, data engineering, and analytics stacks.

·        Strong understanding of security, compliance, scalability, performance engineering.

 

Leadership Competencies

·        Execution-focused technology leadership.

·        Strong communication and stakeholder management skills.

·        Ability to lead distributed teams, manage complexity, and drive measurable outcomes.

·        Innovation mindset with practical implementation capability.

 

Education

·        Bachelor’s or Master’s in Computer Science/Engineering or equivalent.

·        Additional leadership education (MBA or similar) is a plus, not mandatory.

 

Travel Requirements

·        Occasional travel for client meetings, technology reviews, or global delivery coordination.

 

Must-Haves

·        10+ years of technology experience, with with at least 6 years leading large (50-100+) multi product engineering teams.

·        Must have worked on B2B Platforms. Experience in Procurement Tech or Supply Chain

·        Min. 10+ Years of Expertise in Cloud-Native Architecture, Expert-level design in Azure, AWS, or GCP using Microservices, Kubernetes (K8s), and Docker.

·        Min. 8+ Years of Expertise in Modern Engineering Practices, Advanced DevOps, CI/CD pipelines, and automated testing frameworks (Selenium, Cypress, etc.).

·        Hands-on leadership experience in Security & Compliance.

·        Min. 3+ Years of Expertise in AI & Data Engineering, Practical implementation of LLMs, Predictive Analytics, or AI-driven automation

·        Strong technology execution leadership, with ownership of end-to-end technology roadmaps aligned to business outcomes.

·        Min. 6+ Years of Expertise in B2B eCommerce Logic Architecture of Headless Commerce, marketplace integrations, and complex B2B catalog management.

·        Strong product management exposure

·        Proven experience in leading end-to-end team operations

·        Relevant experience in product-driven organizations or platforms

·        Strong Subject Matter Expertise (SME)

 

Education: - Master degree.

 

**************

Joining time / Notice Period: Immediate - 45days.

Location: - Andheri,

5 days working (3 - 2 days’ work from office)

Read more
MNC

MNC

Agency job
via rekha by Rekja Gorle
Mumbai
5 - 10 yrs
₹10L - ₹25L / yr
Windows Azure
DevOps
Microsoft Windows Azure
skill iconKubernetes
Google Cloud Platform (GCP)
+6 more

We are hiring a Senior DevOps Engineer (5–10 years experience) with strong hands-on expertise in AWS, CI/CD, Docker, Kubernetes, and Linux. The role involves designing, automating, and managing scalable cloud infrastructure and deployment pipelines. Experience with Terraform/Ansible, monitoring tools, and security best practices is required. Immediate joiners preferred.

Read more
Matchmaking platform

Matchmaking platform

Agency job
via Peak Hire Solutions by Dharati Thakkar
Mumbai
2 - 5 yrs
₹21L - ₹28L / yr
skill iconData Science
skill iconPython
Natural Language Processing (NLP)
MySQL
skill iconMachine Learning (ML)
+15 more

Review Criteria

  • Strong Data Scientist/Machine Learnings/ AI Engineer Profile
  • 2+ years of hands-on experience as a Data Scientist or Machine Learning Engineer building ML models
  • Strong expertise in Python with the ability to implement classical ML algorithms including linear regression, logistic regression, decision trees, gradient boosting, etc.
  • Hands-on experience in minimum 2+ usecaseds out of recommendation systems, image data, fraud/risk detection, price modelling, propensity models
  • Strong exposure to NLP, including text generation or text classification (Text G), embeddings, similarity models, user profiling, and feature extraction from unstructured text
  • Experience productionizing ML models through APIs/CI/CD/Docker and working on AWS or GCP environments
  • Preferred (Company) – Must be from product companies

 

Job Specific Criteria

  • CV Attachment is mandatory
  • What's your current company?
  • Which use cases you have hands on experience?
  • Are you ok for Mumbai location (if candidate is from outside Mumbai)?
  • Reason for change (if candidate has been in current company for less than 1 year)?
  • Reason for hike (if greater than 25%)?

 

Role & Responsibilities

  • Partner with Product to spot high-leverage ML opportunities tied to business metrics.
  • Wrangle large structured and unstructured datasets; build reliable features and data contracts.
  • Build and ship models to:
  • Enhance customer experiences and personalization
  • Boost revenue via pricing/discount optimization
  • Power user-to-user discovery and ranking (matchmaking at scale)
  • Detect and block fraud/risk in real time
  • Score conversion/churn/acceptance propensity for targeted actions
  • Collaborate with Engineering to productionize via APIs/CI/CD/Docker on AWS.
  • Design and run A/B tests with guardrails.
  • Build monitoring for model/data drift and business KPIs


Ideal Candidate

  • 2–5 years of DS/ML experience in consumer internet / B2C products, with 7–8 models shipped to production end-to-end.
  • Proven, hands-on success in at least two (preferably 3–4) of the following:
  • Recommender systems (retrieval + ranking, NDCG/Recall, online lift; bandits a plus)
  • Fraud/risk detection (severe class imbalance, PR-AUC)
  • Pricing models (elasticity, demand curves, margin vs. win-rate trade-offs, guardrails/simulation)
  • Propensity models (payment/churn)
  • Programming: strong Python and SQL; solid git, Docker, CI/CD.
  • Cloud and data: experience with AWS or GCP; familiarity with warehouses/dashboards (Redshift/BigQuery, Looker/Tableau).
  • ML breadth: recommender systems, NLP or user profiling, anomaly detection.
  • Communication: clear storytelling with data; can align stakeholders and drive decisions.


Read more
Sweet Dreams

at Sweet Dreams

1 candid answer
1 video
Gaurang Ruparelia
Posted by Gaurang Ruparelia
Mumbai
2 - 8 yrs
₹5L - ₹9L / yr
Shopify
skill iconNextJs (Next.js)
Google Cloud Platform (GCP)

What This Role Offers

  • Direct access to founders — plus the runway to build your own e-comm brand.
  • Creative freedom to experiment with formats, hooks, and trends — and ship fast.
  • Massive visibility — your work will reach millions across India and the world.
  • A front-row seat to scaling a fast-growing, publicly listed fashion brand.

Not the Right Fit If You

  • Don’t understand technical concepts like Browser mechanics, Database management, Security.
  • Need heavy scripting or step-by-step creative direction to ship.
  • Haven’t contributed to Open Source before.
  • Want a purely managerial/strategist role with minimal hands-on creation.

Pro Tip: share your portfolio of past tech products you’ve made and notice our speed at responding ;)



What Makes You the Perfect Fit:

  • Technical Foundation: BSc in Computer Science (or equivalent) and prior experience as a Software Engineer—comfortable reading and writing code when needed
  • Automation & No-Code Expertise: Proven track record implementing at least two enterprise-grade SaaS or no-code tools (e.g., Coda, Zapier, Make.com, n8n) across multiple teams
  • Process-First Mindset: You see operations as a canvas—love mapping workflows, spotting bottlenecks, and reimagining end-to-end processes
  • Cross-Functional Communicator: Skilled at translating technical capabilities into clear business solutions, training non-technical stakeholders, and fostering user adoption
  • Continuous Learner: Passionate about evaluating the latest AI and automation platforms—constantly asking, “How can this tool make us faster, smarter, and more resilient?”


Why Work With Us & What’s In It For You

  • End-to-End Problem Solving: You’ll get to deconstruct real-world retail challenges—from inventory bottlenecks to customer experience gaps—fueling your curiosity and sparking ideas for building your own SaaS ventures down the road.
  • Deep Customer Insight: By collaborating closely with frontline teams and engaging directly with customers, you’ll develop a keen understanding of business grievances and the “voice of the customer,” sharpening your ability to design solutions that truly matter.
  • Avant-Garde Tool Playground: This isn’t a sandbox—it’s a live environment. You’ll have the freedom and support to evaluate, pilot, and roll out cutting-edge AI and SaaS platforms in mission-critical contexts, measuring impact in real time.
  • High-Impact Role: You’ll be the catalyst for digital transformation across every function of a fast-growing public company.
  • End-to-End Visibility: From sourcing raw materials to delighting end-customers, you’ll see firsthand how your solutions ripple through each stage of our business.


What You’ll Do:

  • Drive Process Reforms: Map existing workflows across Sales, Sourcing, Retail, E-commerce, Marketing, Accounts, Finance, Legal, and other functions—identify inefficiencies and redesign for speed, accuracy, and scalability.
  • Document & Diagram Workflows: Create clear process documentation and visual workflow diagrams for every automation and tool integration to ensure transparency and maintainability.
  • Produce Training Content: Develop short, engaging video tutorials and introductory demos that walk teams through new SaaS/no-code tools and AI workflows, ensuring rapid adoption and self-service capability.
  • Implement Modern SaaS using Vibe Coding & No-Code Solutions: Lead the roll-out of platforms such as Coda, Make, Zapier, n8n, and generative-AI services—build custom dashboards, automated notifications, data integrations, and self-service portals for each department.
  • Cross-Department Collaboration: Partner with stakeholders from every team to define requirements, pilot new systems, gather feedback, and scale solutions—ensure adoption through workshops, documentation, and ongoing support.
  • Prototype & Experiment Rapidly: Develop proof-of-concept automations and AI workflows; conduct quick A/B tests to validate impact on KPIs like turnaround time, error rates, and employee satisfaction.
  • Scale & Standardize: Once a tool or process shows positive ROI, establish best practices, security guidelines, and governance frameworks to ensure consistency and compliance across the organization.



Read more
Deqode

at Deqode

1 recruiter
Samiksha Agrawal
Posted by Samiksha Agrawal
Mumbai
3 - 6 yrs
₹5L - ₹15L / yr
DevOps
Google Cloud Platform (GCP)
Terraform
skill iconJenkins
CI/CD
+2 more

Role: Senior Platform Engineer (GCP Cloud)

Experience Level: 3 to 6 Years

Work location: Mumbai

Mode : Hybrid


Role & Responsibilities:

  • Build automation software for cloud platforms and applications
  • Drive Infrastructure as Code (IaC) adoption
  • Design self-service, self-healing monitoring and alerting tools
  • Automate CI/CD pipelines (Git, Jenkins, SonarQube, Docker)
  • Build Kubernetes container platforms
  • Introduce new cloud technologies for business innovation

Requirements:

  • Hands-on experience with GCP Cloud
  • Knowledge of cloud services (compute, storage, network, messaging)
  • IaC tools experience (Terraform/CloudFormation)
  • SQL & NoSQL databases (Postgres, Cassandra)
  • Automation tools (Puppet/Chef/Ansible)
  • Strong Linux administration skills
  • Programming: Bash/Python/Java/Scala
  • CI/CD pipeline expertise (Jenkins, Git, Maven)
  • Multi-region deployment experience
  • Agile/Scrum/DevOps methodology


Read more
AI-First Company

AI-First Company

Agency job
via Peak Hire Solutions by Dharati Thakkar
Bengaluru (Bangalore), Mumbai, Hyderabad, Gurugram
5 - 17 yrs
₹30L - ₹45L / yr
Data engineering
Data architecture
SQL
Data modeling
GCS
+47 more

ROLES AND RESPONSIBILITIES:

You will be responsible for architecting, implementing, and optimizing Dremio-based data Lakehouse environments integrated with cloud storage, BI, and data engineering ecosystems. The role requires a strong balance of architecture design, data modeling, query optimization, and governance enablement in large-scale analytical environments.


  • Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
  • Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
  • Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
  • Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
  • Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
  • Support self-service analytics by enabling governed data products and semantic layers.
  • Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
  • Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.


IDEAL CANDIDATE:

  • Bachelor’s or Master’s in Computer Science, Information Systems, or related field.
  • 5+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
  • Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
  • Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
  • Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
  • Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
  • Excellent problem-solving, documentation, and stakeholder communication skills.


PREFERRED:

  • Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) and data catalogs (Collibra, Alation, Purview).
  • Exposure to Snowflake, Databricks, or BigQuery environments.
  • Experience in high-tech, manufacturing, or enterprise data modernization programs.
Read more
Matchmaking platform

Matchmaking platform

Agency job
via Peak Hire Solutions by Dharati Thakkar
Mumbai
2 - 5 yrs
₹15L - ₹28L / yr
skill iconData Science
skill iconPython
Natural Language Processing (NLP)
MySQL
skill iconMachine Learning (ML)
+15 more

Review Criteria

  • Strong Data Scientist/Machine Learnings/ AI Engineer Profile
  • 2+ years of hands-on experience as a Data Scientist or Machine Learning Engineer building ML models
  • Strong expertise in Python with the ability to implement classical ML algorithms including linear regression, logistic regression, decision trees, gradient boosting, etc.
  • Hands-on experience in minimum 2+ usecaseds out of recommendation systems, image data, fraud/risk detection, price modelling, propensity models
  • Strong exposure to NLP, including text generation or text classification (Text G), embeddings, similarity models, user profiling, and feature extraction from unstructured text
  • Experience productionizing ML models through APIs/CI/CD/Docker and working on AWS or GCP environments
  • Preferred (Company) – Must be from product companies

 

Job Specific Criteria

  • CV Attachment is mandatory
  • What's your current company?
  • Which use cases you have hands on experience?
  • Are you ok for Mumbai location (if candidate is from outside Mumbai)?
  • Reason for change (if candidate has been in current company for less than 1 year)?
  • Reason for hike (if greater than 25%)?

 

Role & Responsibilities

  • Partner with Product to spot high-leverage ML opportunities tied to business metrics.
  • Wrangle large structured and unstructured datasets; build reliable features and data contracts.
  • Build and ship models to:
  • Enhance customer experiences and personalization
  • Boost revenue via pricing/discount optimization
  • Power user-to-user discovery and ranking (matchmaking at scale)
  • Detect and block fraud/risk in real time
  • Score conversion/churn/acceptance propensity for targeted actions
  • Collaborate with Engineering to productionize via APIs/CI/CD/Docker on AWS.
  • Design and run A/B tests with guardrails.
  • Build monitoring for model/data drift and business KPIs


Ideal Candidate

  • 2–5 years of DS/ML experience in consumer internet / B2C products, with 7–8 models shipped to production end-to-end.
  • Proven, hands-on success in at least two (preferably 3–4) of the following:
  • Recommender systems (retrieval + ranking, NDCG/Recall, online lift; bandits a plus)
  • Fraud/risk detection (severe class imbalance, PR-AUC)
  • Pricing models (elasticity, demand curves, margin vs. win-rate trade-offs, guardrails/simulation)
  • Propensity models (payment/churn)
  • Programming: strong Python and SQL; solid git, Docker, CI/CD.
  • Cloud and data: experience with AWS or GCP; familiarity with warehouses/dashboards (Redshift/BigQuery, Looker/Tableau).
  • ML breadth: recommender systems, NLP or user profiling, anomaly detection.
  • Communication: clear storytelling with data; can align stakeholders and drive decisions.




Read more
CGI Inc

at CGI Inc

3 recruiters
Shruthi BT
Posted by Shruthi BT
Bengaluru (Bangalore), Mumbai, Pune, Hyderabad, Chennai
8 - 15 yrs
₹15L - ₹25L / yr
Google Cloud Platform (GCP)
Data engineering
Big query

Google Data Engineer - SSE


Position Description

Google Cloud Data Engineer

Notice Period: Immediate to 30 days serving

Job Description:

We are seeking a highly skilled Data Engineer with extensive experience in Google Cloud Platform (GCP) data services and big data technologies. The ideal candidate will be responsible for designing, implementing, and optimizing scalable data solutions while ensuring high performance, reliability, and security.

Key Responsibilities:


• Design, develop, and maintain scalable data pipelines and architectures using GCP data services.

• Implement and optimize solutions using BigQuery, Dataproc, Composer, Pub/Sub, Dataflow, GCS, and BigTable.

• Work with GCP databases such as Bigtable, Spanner, CloudSQL, AlloyDB, ensuring performance, security, and availability.

• Develop and manage data processing workflows using Apache Spark, Hadoop, Hive, Kafka, and other Big Data technologies.

• Ensure data governance and security using Dataplex, Data Catalog, and other GCP governance tooling.

• Collaborate with DevOps teams to build CI/CD pipelines for data workloads using Cloud Build, Artifact Registry, and Terraform.

• Optimize query performance and data storage across structured and unstructured datasets.

• Design and implement streaming data solutions using Pub/Sub, Kafka, or equivalent technologies.


Required Skills & Qualifications:


• 8-15 years of experience

• Strong expertise in GCP Dataflow, Pub/Sub, Cloud Composer, Cloud Workflow, BigQuery, Cloud Run, Cloud Build.

• Proficiency in Python and Java, with hands-on experience in data processing and ETL pipelines.

• In-depth knowledge of relational databases (SQL, MySQL, PostgreSQL, Oracle) and NoSQL databases (MongoDB, Scylla, Cassandra, DynamoDB).

• Experience with Big Data platforms such as Cloudera, Hortonworks, MapR, Azure HDInsight, IBM Open Platform.

• Strong understanding of AWS Data services such as Redshift, RDS, Athena, SQS/Kinesis.

• Familiarity with data formats such as Avro, ORC, Parquet.

• Experience handling large-scale data migrations and implementing data lake architectures.

• Expertise in data modeling, data warehousing, and distributed data processing frameworks.

• Deep understanding of data formats such as Avro, ORC, Parquet.

• Certification in GCP Data Engineering Certification or equivalent.


Good to Have:


• Experience in BigQuery, Presto, or equivalent.

• Exposure to Hadoop, Spark, Oozie, HBase.

• Understanding of cloud database migration strategies.

• Knowledge of GCP data governance and security best practices.

Read more
Tradelab Technologies

at Tradelab Technologies

1 candid answer
Aakanksha Yadav
Posted by Aakanksha Yadav
Mumbai, Bengaluru (Bangalore)
10 - 18 yrs
₹25L - ₹50L / yr
skill iconKubernetes
skill iconDocker
skill iconAmazon Web Services (AWS)
Microsoft Windows Azure
Google Cloud Platform (GCP)
+7 more

Type: Client-Facing Technical Architecture, Infrastructure Solutioning & Domain Consulting (India + International Markets)


Role Overview

Tradelab is seeking a senior Solution Architect who can interact with both Indian and international clients (Dubai, Singapore, London, US), helping them understand our trading systems, OMS/RMS/CMS stack, HFT platforms, feed systems, and Matching Engine. The architect will design scalable, secure, and ultra-low-latency deployments tailored to global forex markets, brokers, prop firms, liquidity providers, and market makers.


Key Responsibilities

1. Client Engagement (India + International Markets)

  • Engage with brokers, prop trading firms, liquidity providers, and financial institutions across India, Dubai, Singapore, and global hubs.
  • Explain Tradelab’s capabilities, architecture, and deployment options.
  • Understand region-specific latency expectations, connectivity options, and regulatory constraints.

2. Requirement Gathering & Solutioning

  • Capture client needs, throughput, order concurrency, tick volumes, and market data handling.
  • Assess infra readiness (cloud/on-prem/colo).
  • Propose architecture aligned with forex markets.

3. Global Architecture & Deployment Design

  • Design multi-region infrastructure using AWS/Azure/GCP.
  • Architect low-latency routing between India–Singapore–Dubai.
  • Support deployments in DCs like Equinix SG1/DX1.

4. Networking & Security Architecture

  • Architect multicast/unicast feeds, VPNs, IPSec tunnels, BGP routes.
  • Implement network hardening, segmentation, WAF/firewall rules.

5. DevOps, Cloud Engineering & Scalability

  • Build CI/CD pipelines, Kubernetes autoscaling, cost-optimized AWS multi-region deployments.
  • Design global failover models.

6. BFSI & Trading Domain Expertise

  • Indian broking, international forex, LP aggregation, HFT.
  • OMS/RMS, risk engines, LP connectivity, and matching engines.

7. Latency, Performance & Capacity Planning

  • Benchmark and optimize cross-region latency.
  • Tune performance for high tick volumes and volatility bursts.

8. Documentation & Consulting

  • Prepare HLDs, LLDs, SOWs, cost sheets, and deployment of playbooks.
  • Required Skills
  • AWS: EC2, VPC, EKS, NLB, MSK/Kafka, IAM, Global Accelerator.
  • DevOps: Kubernetes, Docker, Helm, Terraform.
  • Networking: IPSec, GRE, VPN, BGP, multicast (PIM/IGMP).
  • Message buses: Kafka, RabbitMQ, Redis Streams.

Domain Skills

  • Deep Broking Domain Understanding.
  • Indian broking + global forex/CFD.
  • FIX protocol, LP integration, market data feeds.
  • Regulations: SEBI, DFSA, MAS, ESMA.

Soft Skills

  • Excellent communication and client-facing ability.
  • Strong presales and solutioning mindset.
  • Preferred Qualifications
  • B.Tech/BE/M.Tech in CS or equivalent.
  • AWS Architect Professional, CCNP, CKA.

Why Join Us?

  • Experience in colocation/global trading infra.
  • Work with a team that expects and delivers excellence.
  • A culture where risk-taking is rewarded, and complacency is not.
  • Limitless opportunities for growth—if you can handle the pace.
  • A place where learning is currency, and outperformance is the only metric that matters.
  • The opportunity to build systems that move markets, execute trades in microseconds, and redefine fintech.

This isn’t just a job—it’s a proving ground. Ready to take the leap? Apply now.


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Praffull Shinde
Posted by Praffull Shinde
Pune, Mumbai, Bengaluru (Bangalore)
8 - 14 yrs
Best in industry
Google Cloud Platform (GCP)
Terraform
skill iconKubernetes
DevOps
skill iconPython

JD for Cloud engineer

 

Job Summary:


We are looking for an experienced GCP Cloud Engineer to design, implement, and manage cloud-based solutions on Google Cloud Platform (GCP). The ideal candidate should have expertise in GKE (Google Kubernetes Engine), Cloud Run, Cloud Loadbalancer, Cloud function, Azure DevOps, and Terraform, with a strong focus on automation, security, and scalability.


You will work closely with development, operations, and security teams to ensure robust cloud infrastructure and CI/CD pipelines while optimizing performance and cost.

 

Key Responsibilities:

1. Cloud Infrastructure Design & Management

  • Architect, deploy, and maintain GCP cloud resources via terraform/other automation.
  • Implement Google Cloud Storage, Cloud SQL, filestore,  for data storage and processing needs.
  • Manage and configure Cloud Load Balancers (HTTP(S), TCP/UDP, and SSL Proxy) for high availability and scalability.
  • Optimize resource allocation, monitoring, and cost efficiency across GCP environments.


2. Kubernetes & Container Orchestration

  • Deploy, manage, and optimize workloads on Google Kubernetes Engine (GKE).
  • Work with Helm charts for microservices deployments.
  • Automate scaling, rolling updates, and zero-downtime deployments.

 

3. Serverless & Compute Services

  • Deploy and manage applications on Cloud Run and Cloud Functions for scalable, serverless workloads.
  • Optimize containerized applications running on Cloud Run for cost efficiency and performance.

 

4. CI/CD & DevOps Automation

  • Design, implement, and manage CI/CD pipelines using Azure DevOps.
  • Automate infrastructure deployment using Terraform, Bash and Powershell scripting
  • Integrate security and compliance checks into the DevOps workflow (DevSecOps).

 

 

Required Skills & Qualifications:

Experience: 8+ years in Cloud Engineering, with a focus on GCP.

Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions).

Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm.

DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation.

Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources.

Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation.

Security & Compliance: Knowledge of cloud security principles, IAM, and compliance standards.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Bengaluru (Bangalore), Mumbai, Pune
7 - 12 yrs
₹1L - ₹45L / yr
Google Cloud Platform (GCP)
skill iconKubernetes
skill iconDocker
google kubernetes engineer
azure devops
+2 more

Required Skills & Qualifications:

✔ Experience: 4+ years in Cloud Engineering, with a focus on GCP.

✔ Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions).

✔ Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm.

✔ DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation.

✔ Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources.

✔ Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation.

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Bengaluru (Bangalore), Mumbai, Pune
3 - 7 yrs
Best in industry
Google Cloud Platform (GCP)
Shell Scripting
skill iconJava
skill iconRuby
Product Management

Job Title: Site Reliability Engineer (SRE) / Application Support Engineer

Experience: 3–7 Years

Location: Bangalore / Mumbai / Pune

About the Role

The successful candidate will join the S&C Site Reliability Engineering (SRE) Team, responsible for providing Tier 2/3 support to S&C business applications and environments. This role requires close collaboration with client-facing teams (Client Services, Product, and Research) as well as Infrastructure, Technology, and Application Development teams to maintain and support production and non-production environments.

Key Responsibilities

  • Provide Tier 2/3 product technical support and issue resolution.
  • Develop and maintain software tools to improve operations and support efficiency.
  • Manage system and software configurations; troubleshoot environment-related issues.
  • Identify opportunities to optimize system performance through configuration improvements or development suggestions.
  • Plan, document, and deploy software applications across Unix/Linux, Azure, and GCP environments.
  • Collaborate with Development and QA teams throughout the software release lifecycle.
  • Analyze and improve release and deployment processes to drive automation and efficiency.
  • Coordinate with infrastructure teams for maintenance, planned downtimes, and resource management across production and non-production environments.
  • Participate in on-call support (minimum one week per month) for off-hour emergencies and maintenance activities.

Required Skills & Qualifications

  • Education:
  • Bachelor’s degree in Computer Science, Engineering, or a related field (BE/MCA).
  • Master’s degree is a plus.
  • Experience:
  • 3–7 years in Production Support, Application Management, or Application Development (support/maintenance).
  • Technical Skills:
  • Strong Unix/Linux administration skills.
  • Excellent scripting skills — Shell, Python, Batch (mandatory).
  • Database expertise — Oracle (must have).
  • Understanding of Software Development Life Cycle (SDLC).
  • PowerShell knowledge is a plus.
  • Experience in Java or Ruby development is desirable.
  • Exposure to cloud platforms (GCP, Azure, or AWS) is an added advantage.
  • Soft Skills:
  • Excellent problem-solving and troubleshooting abilities.
  • Strong collaboration and communication skills.
  • Ability to work in a fast-paced, cross-functional environment.


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Bengaluru (Bangalore), Mumbai, Pune
3 - 7 yrs
Best in industry
Google Cloud Platform (GCP)
AZURE
skill iconJava
skill iconRuby
Oracle NoSQL Database
+5 more

Required skills and experience

• Bachelor's degree, Computer Science, Engineering or other similar concentration (BE/MCA)

• Master’s degree a plus

• 3-8 years’ experience in Production Support/ Application Management/ Application Development (support/ maintenance) role.

• Excellent problem-solving/troubleshooting skills, fast learner

• Strong knowledge of Unix Administration.

• Strong scripting skills in Shell, Python, Batch is must.

• Strong Database experience – Oracle

• Strong knowledge of Software Development Life Cycle

• Power shell is nice to have

• Software development skillsets in Java or Ruby.

• Worked upon any of the cloud platforms – GCP/Azure/AWS is nice to have

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Moulina Dey
Posted by Moulina Dey
Pune, Bengaluru (Bangalore), Mumbai
3 - 6 yrs
₹2L - ₹14L / yr
technical product support
Linux/Unix
Google Cloud Platform (GCP)
SRE
Reliability engineering

Department: S&C – Site Reliability Engineering (SRE)  

Experience Required: 4–8 Years  

Location: Bangalore / Pune /Mumbai 

Employment Type: Full-time


  • Provide Tier 2/3 technical product support to internal and external stakeholders. 
  • Develop automation tools and scripts to improve operational efficiency and support processes. 
  • Manage and maintain system and software configurations; troubleshoot environment/application-related issues. 
  • Optimize system performance through configuration tuning or development enhancements. 
  • Plan, document, and deploy applications in Unix/Linux, Azure, and GCP environments
  • Collaborate with Development, QA, and Infrastructure teams throughout the release and deployment of lifecycles
  • Drive automation initiatives for release and deployment processes. 
  • Coordinate with infrastructure teams to manage hardware/software resources, maintenance, and scheduled downtimes across production and non-production environments. 
  • Participate in on-call rotations (minimum one week per month) to address critical incidents and off-hour maintenance tasks. 

 

Key Competencies 

  • Strong analytical, troubleshooting, and critical thinking abilities. 
  • Excellent cross-functional collaboration skills. 
  • Strong focus on documentation, process improvement, and system reliability
  • Proactive, detail-oriented, and adaptable in a fast-paced work environment. 


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Bipasha Rath
Posted by Bipasha Rath
Mumbai, Pune
5 - 9 yrs
Best in industry
Google Cloud Platform (GCP)
Terraform
iac
Axure

We are seeking Cloud Developer with experience in GCP/Azure along with Terraform coding. They will help to manage & standardize the IaC module.

 

Experience:5 - 8 Years

Location:Mumbai & Pune

Mode of Work:Full Time

 Key Responsibilities:

  • Design,   develop, and maintain robust software applications using most   common and popular coding languages suitable for the application   design, with a strong focus on clean, maintainable, and efficient   code.
  • Develop,   maintain, and enhance Terraform modules to encapsulate common   infrastructure patterns and promote code reuse and standardization.
  • Develop   RESTful APIs and backend services aligned with modern architectural   practices.
  • Apply   object-oriented programming principles and design patterns to build   scalable systems.
  • Build   and maintain automated test frameworks and scripts to   ensure high product quality.
  • Troubleshoot   and resolve technical issues across application layers, from code to   infrastructure.
  • Work   with cloud platforms such as Azure or   Google Cloud Platform (GCP).
  • Use Git and   related version control practices effectively in a team-based   development environment.
  • Integrate   and experiment with AI development tools like GitHub   Copilot, Azure OpenAI, or similar to boost engineering   efficiency.

 

Requirements:

  • 5+   years of experience
  • Experience   with IaC Module
  • Terraform   coding experience along with Terraform   Module as a part of central platform team
  • Azure/GCP   cloud experience is a must
  • Experience   with C#/Python/Java Coding - is good to have

 

 If interested please share your updated resume with below details :

Total Experience -

Relevant Experience -

Current Location -

Current CTC -

Expected CTC -

Notice period -

Any offer in hand -


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Bipasha Rath
Posted by Bipasha Rath
Pune, Mumbai
5 - 8 yrs
Best in industry
Google Cloud Platform (GCP)
AZURE
Terraform
skill icon.NET
skill iconPython
+2 more

Job Description:


Position - Cloud Developer

Experience - 5 - 8 years

Location - Mumbai & Pune


Responsibilities:

  • Design, develop, and maintain robust software applications using most common and popular coding languages suitable for the application design, with a strong focus on clean, maintainable, and efficient code.
  • Develop, maintain, and enhance Terraform modules to encapsulate common infrastructure patterns and promote code reuse and standardization.
  • Develop RESTful APIs and backend services aligned with modern architectural practices.
  • Apply object-oriented programming principles and design patterns to build scalable systems.
  • Build and maintain automated test frameworks and scripts to ensure high product quality.
  • Troubleshoot and resolve technical issues across application layers, from code to infrastructure.
  • Work with cloud platforms such as Azure or Google Cloud Platform (GCP).
  • Use Git and related version control practices effectively in a team-based development environment.
  • Integrate and experiment with AI development tools like GitHub Copilot, Azure OpenAI, or similar to boost engineering efficiency.


Skills:

  • 5+ years of experience
  • Experience with IaC Module
  • Terraform coding experience along with Terraform Module as a part of central platform team
  • Azure/GCP cloud experience is a must
  • Experience with C#/Python/Java Coding - is good to have


Read more
Wissen Technology
Pune, Mumbai, Bengaluru (Bangalore)
4 - 10 yrs
Best in industry
Google Cloud Platform (GCP)
skill iconPython
skill iconKubernetes
Shell Scripting
SRE Engineer
+1 more

Dear Candidate,


Greetings from Wissen Technology. 

We have an exciting Job opportunity for GCP SRE Engineer Professionals. Please refer to the Job Description below and share your profile if interested.   

 About Wissen Technology:

  • The Wissen Group was founded in the year 2000. Wissen Technology, a part of Wissen Group, was established in the year 2015.
  • Wissen Technology is a specialized technology company that delivers high-end consulting for organizations in the Banking & Finance, Telecom, and Healthcare domains. We help clients build world class products.
  • Our workforce consists of 1000+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world.
  • Wissen Technology has grown its revenues by 400% in these five years without any external funding or investments.
  • Globally present with offices US, India, UK, Australia, Mexico, and Canada.
  • We offer an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation.
  • Wissen Technology has been certified as a Great Place to Work®.
  • Wissen Technology has been voted as the Top 20 AI/ML vendor by CIO Insider in 2020.
  • Over the years, Wissen Group has successfully delivered $650 million worth of projects for more than 20 of the Fortune 500 companies.
  • The technology and thought leadership that the company commands in the industry is the direct result of the kind of people Wissen has been able to attract. Wissen is committed to providing them the best possible opportunities and careers, which extends to providing the best possible experience and value to our clients.

We have served client across sectors like Banking, Telecom, Healthcare, Manufacturing, and Energy. They include likes of Morgan Stanley, MSCI, State Street Corporation, Flipkart, Swiggy, Trafigura, GE to name a few.



Job Description: 

Please find below details:


Experience - 4+ Years

Location- Bangalore/Mumbai/Pune


Team Responsibilities

The successful candidate shall be part of the S&C – SRE Team. Our team provides a tier 2/3 support to S&C Business. This position involves collaboration with the client facing teams like Client Services, Product and Research teams and Infrastructure/Technology and Application development teams to perform Environment and Application maintenance and support.

 

Resource's key Responsibilities


• Provide Tier 2/3 product technical support.

• Building software to help operations and support activities.

• Manage system\software configurations and troubleshoot environment issues.

• Identify opportunities for optimizing system performance through changes in configuration or suggestions for development.

• Plan, document and deploy software applications on our Unix/Linux/Azure and GCP based systems.

• Collaborate with development and software testing teams throughout the release process.

• Analyze release and deployment processes to identify key areas for automation and optimization.

• Manage hardware and software resources & coordinate maintenance, planned downtimes with

infrastructure group across all the environments. (Production / Non-Production).

• Must spend minimum one week a month as on call support to help with off-hour emergencies and maintenance activities.

 

Required skills and experience

• Bachelor's degree, Computer Science, Engineering or other similar concentration (BE/MCA)

• Master’s degree a plus

• 6-8 years’ experience in Production Support/ Application Management/ Application Development (support/ maintenance) role.

• Excellent problem-solving/troubleshooting skills, fast learner

• Strong knowledge of Unix Administration.

• Strong scripting skills in Shell, Python, Batch is must.

• Strong Database experience – Oracle

• Strong knowledge of Software Development Life Cycle

• Power shell is nice to have

• Software development skillsets in Java or Ruby.

• Worked upon any of the cloud platforms – GCP/Azure/AWS is nice to have




Read more
Wissen Technology

at Wissen Technology

4 recruiters
Sonali RajeshKumar
Posted by Sonali RajeshKumar
Bengaluru (Bangalore), Pune, Mumbai
4 - 9 yrs
Best in industry
Google Cloud Platform (GCP)
Reliability engineering
skill iconPython
Shell Scripting

Dear Candidate,


Greetings from Wissen Technology. 

We have an exciting Job opportunity for GCP SRE Engineer Professionals. Please refer to the Job Description below and share your profile if interested.   

 About Wissen Technology:

  • The Wissen Group was founded in the year 2000. Wissen Technology, a part of Wissen Group, was established in the year 2015.
  • Wissen Technology is a specialized technology company that delivers high-end consulting for organizations in the Banking & Finance, Telecom, and Healthcare domains. We help clients build world class products.
  • Our workforce consists of 1000+ highly skilled professionals, with leadership and senior management executives who have graduated from Ivy League Universities like Wharton, MIT, IITs, IIMs, and NITs and with rich work experience in some of the biggest companies in the world.
  • Wissen Technology has grown its revenues by 400% in these five years without any external funding or investments.
  • Globally present with offices US, India, UK, Australia, Mexico, and Canada.
  • We offer an array of services including Application Development, Artificial Intelligence & Machine Learning, Big Data & Analytics, Visualization & Business Intelligence, Robotic Process Automation, Cloud, Mobility, Agile & DevOps, Quality Assurance & Test Automation.
  • Wissen Technology has been certified as a Great Place to Work®.
  • Wissen Technology has been voted as the Top 20 AI/ML vendor by CIO Insider in 2020.
  • Over the years, Wissen Group has successfully delivered $650 million worth of projects for more than 20 of the Fortune 500 companies.
  • The technology and thought leadership that the company commands in the industry is the direct result of the kind of people Wissen has been able to attract. Wissen is committed to providing them the best possible opportunities and careers, which extends to providing the best possible experience and value to our clients.

We have served client across sectors like Banking, Telecom, Healthcare, Manufacturing, and Energy. They include likes of Morgan Stanley, MSCI, State Street Corporation, Flipkart, Swiggy, Trafigura, GE to name a few


Job Description: 

Please find below details:


Experience - 4+ Years

Location- Bangalore/Mumbai/Pune


Team Responsibilities

The successful candidate shall be part of the S&C – SRE Team. Our team provides a tier 2/3 support to S&C Business. This position involves collaboration with the client facing teams like Client Services, Product and Research teams and Infrastructure/Technology and Application development teams to perform Environment and Application maintenance and support.

 

Resource's key Responsibilities


• Provide Tier 2/3 product technical support.

• Building software to help operations and support activities.

• Manage system\software configurations and troubleshoot environment issues.

• Identify opportunities for optimizing system performance through changes in configuration or suggestions for development.

• Plan, document and deploy software applications on our Unix/Linux/Azure and GCP based systems.

• Collaborate with development and software testing teams throughout the release process.

• Analyze release and deployment processes to identify key areas for automation and optimization.

• Manage hardware and software resources & coordinate maintenance, planned downtimes with

infrastructure group across all the environments. (Production / Non-Production).

• Must spend minimum one week a month as on call support to help with off-hour emergencies and maintenance activities.

 

Required skills and experience

• Bachelor's degree, Computer Science, Engineering or other similar concentration (BE/MCA)

• Master’s degree a plus

• 6-8 years’ experience in Production Support/ Application Management/ Application Development (support/ maintenance) role.

• Excellent problem-solving/troubleshooting skills, fast learner

• Strong knowledge of Unix Administration.

• Strong scripting skills in Shell, Python, Batch is must.

• Strong Database experience – Oracle

• Strong knowledge of Software Development Life Cycle

• Power shell is nice to have

• Software development skillsets in Java or Ruby.

• Worked upon any of the cloud platforms – GCP/Azure/AWS is nice to have


Read more
Bengaluru (Bangalore), Mumbai, Delhi, Gurugram, Noida, Ghaziabad, Faridabad, Pune, Hyderabad, Mohali, Dehradun, Panchkula, Chennai
6 - 14 yrs
₹12L - ₹28L / yr
Test Automation (QA)
skill iconKubernetes
helm
skill iconDocker
skill iconAmazon Web Services (AWS)
+13 more

Job Title : Senior QA Automation Architect (Cloud & Kubernetes)

Experience : 6+ Years

Location : India (Multiple Offices)

Shift Timings : 12 PM to 9 PM (Noon Shift)

Working Days : 5 Days WFO (NO Hybrid)


About the Role :

We’re looking for a Senior QA Automation Architect with deep expertise in cloud-native systems, Kubernetes, and automation frameworks.

You’ll design scalable test architectures, enhance automation coverage, and ensure product reliability across hybrid-cloud and distributed environments.


Key Responsibilities :

  • Architect and maintain test automation frameworks for microservices.
  • Integrate automated tests into CI/CD pipelines (Jenkins, GitHub Actions).
  • Ensure reliability, scalability, and observability of test systems.
  • Work closely with DevOps and Cloud teams to streamline automation infrastructure.

Mandatory Skills :

  • Kubernetes, Helm, Docker, Linux
  • Cloud Platforms : AWS / Azure / GCP
  • CI/CD Tools : Jenkins, GitHub Actions
  • Scripting : Python, Pytest, Bash
  • Monitoring & Performance : Prometheus, Grafana, Jaeger, K6
  • IaC Practices : Terraform / Ansible

Good to Have :

  • Experience with Service Mesh (Istio/Linkerd).
  • Container Security or DevSecOps exposure.
Read more
Wissen Technology

at Wissen Technology

4 recruiters
Janane Mohanasankaran
Posted by Janane Mohanasankaran
Pune, Mumbai
4 - 7 yrs
Best in industry
skill iconJava
Windows Azure
skill iconAmazon Web Services (AWS)
Google Cloud Platform (GCP)

Job Description

      The ideal candidate will possess expertise in Core Java (at least Java 8), Spring framework, JDBC, threading, database management, and cloud platforms such as Azure and GCP. The candidate should also have strong debugging skills, the ability to understand multi-service flow, experience with large data processing, and excellent problem-solving abilities.

 

JD:

  1. Develop, and maintain Java applications using Core Java, Spring framework, JDBC, and threading concepts.
  2. Strong understanding of the Spring framework and its various modules.
  3. Experience with JDBC for database connectivity and manipulation
  4. Utilize database management systems to store and retrieve data efficiently.
  5. Proficiency in Core Java8 and thorough understanding of threading concepts and concurrent programming.
  6. Experience in in working with relational and nosql databases.
  7. Basic understanding of cloud platforms such as Azure and GCP and gain experience on DevOps practices is added advantage.
  8. Knowledge of containerization technologies (e.g., Docker, Kubernetes)
  9. Perform debugging and troubleshooting of applications using log analysis techniques.
  10. Understand multi-service flow and integration between components.
  11. Handle large-scale data processing tasks efficiently and effectively.
  12. Hands on experience using Spark is an added advantage.
  13. Good problem-solving and analytical abilities.
  14. Collaborate with cross-functional teams to identify and solve complex technical problems.
  15. Knowledge of Agile methodologies such as Scrum or Kanban
  16. Stay updated with the latest technologies and industry trends to improve development processes continuously and methodologies.

 

Read more
Pluginlive

at Pluginlive

1 recruiter
Harsha Saggi
Posted by Harsha Saggi
Mumbai, Chennai
1 - 3 yrs
₹5L - ₹8L / yr
skill iconPython
SQL
Data Structures
ETL
Dashboard
+3 more

About Us:

PluginLive is an all-in-one tech platform that bridges the gap between all its stakeholders - Corporates, Institutes Students, and Assessment & Training Partners. This ecosystem helps Corporates in brand building/positioning with colleges and the student community to scale its human capital, at the same time increasing student placements for Institutes, and giving students a real time perspective of the corporate world to help upskill themselves into becoming more desirable candidates.


Role Overview:

Entry-level Data Engineer position focused on building and maintaining data pipelines while developing visualization skills. You'll work alongside senior engineers to support our data infrastructure and create meaningful insights through data visualization.


Responsibilities:

  • Assist in building and maintaining ETL/ELT pipelines for data processing
  • Write SQL queries to extract and analyze data from various sources
  • Support data quality checks and basic data validation processes
  • Create simple dashboards and reports using visualization tools
  • Learn and work with Oracle Cloud services under guidance
  • Use Python for basic data manipulation and cleaning tasks
  • Document data processes and maintain data dictionaries
  • Collaborate with team members to understand data requirements
  • Participate in troubleshooting data issues with senior support
  • Contribute to data migration tasks as needed


Qualifications:

Required:

  • Bachelor's degree in Computer Science, Information Systems, or related field
  • around 2 years of experience in data engineering or related field
  • Strong SQL knowledge and database concepts
  • Comfortable with Python programming
  • Understanding of data structures and ETL concepts
  • Problem-solving mindset and attention to detail
  • Good communication skills
  • Willingness to learn cloud technologies


Preferred:

  • Exposure to Oracle Cloud or any cloud platform (AWS/GCP)
  • Basic knowledge of data visualization tools (Tableau, Power BI, or Python libraries like Matplotlib)
  • Experience with Pandas for data manipulation
  • Understanding of data warehousing concepts
  • Familiarity with version control (Git)
  • Academic projects or internships involving data processing


Nice-to-Have:

  • Knowledge of dbt, BigQuery, or Snowflake
  • Exposure to big data concepts
  • Experience with Jupyter notebooks
  • Comfort with AI-assisted coding tools (Copilot, GPTs)
  • Personal projects showcasing data work


What We Offer:

  • Mentorship from senior data engineers
  • Hands-on learning with modern data stack
  • Access to paid AI tools and learning resources
  • Clear growth path to mid-level engineer
  • Direct impact on product and data strategy
  • No unnecessary meetings — focused execution
  • Strong engineering culture with continuous learning opportunities
Read more
NTM

NTM

Agency job
Navi Mumbai
4 - 7 yrs
₹5L - ₹19L / yr
Google Cloud Platform (GCP)

Google BigQuery Engineer

Job Description Title : Google Big Query

Location : Navi Mumbai Experience : 4-7 Years


Roles & Responsibilities :


- Expected to perform independently and become an SME.

- Required active participation/contribution in team discussions.

- Contribute in providing solutions to work-related problems.

- Collaborate with cross-functional teams to analyze business requirements and design application solutions.

- Develop and maintain high-quality applications using Google BigQuery.

- Troubleshoot and debug application issues to ensure optimal performance and functionality.

- Perform code reviews and provide constructive feedback to improve code quality.

- Stay updated with the latest industry trends and technologies to continuously enhance skills and knowledge.


Professional & Technical Skills :

- Must To Have Skills: Proficiency in Google BigQuery.

- Strong understanding of data modeling and database design principles.

- Experience in developing scalable and efficient applications using Google BigQuery.

- Hands-on experience with SQL and query optimization techniques.

- Familiarity with data warehousing concepts and ETL processes.

Read more
iRage

at iRage

3 recruiters
Jyosana Jadhav
Posted by Jyosana Jadhav
Mumbai
4 - 7 yrs
₹25L - ₹40L / yr
skill iconReact.js
skill iconJavascript
DOM
WebSocket
Chart.js
+20 more

We are seeking a highly skilled React JS Developer with exceptional DOM manipulation expertise and real-time data handling experience to join our team. You'll be building and optimizing high-performance user interfaces for stock market trading applications where milliseconds matter and data flows continuously.


The ideal candidate thrives in fast-paced environments, understands the intricacies of browser performance, and has hands-on experience with WebSockets and real-time data streaming architectures.


Key Responsibilities


Core Development

  • Advanced DOM Operations: Implement complex, performance-optimized DOM manipulations for real-time trading interfaces
  • Real-time Data Management: Build robust WebSocket connections and handle high-frequency data streams with minimal latency
  • Performance Engineering: Create lightning-fast, scalable front-end applications that process thousands of market updates per second
  • Custom Component Architecture: Design and build reusable, high-performance React components optimized for trading workflows


Collaboration & Integration

  • Work closely with traders, quants, and backend developers to translate complex trading requirements into intuitive interfaces
  • Collaborate with UX/UI designers and product managers to create responsive, trader-focused experiences
  • Integrate with real-time market data APIs and trading execution systems


Technical Excellence

  • Implement sophisticated data visualizations and interactive charts using libraries like Chartjs, TradingView, or custom D3.js solutions
  • Ensure cross-browser compatibility and responsiveness across multiple devices and screen sizes
  • Debug and resolve complex performance issues, particularly in real-time data processing and rendering
  • Maintain high-quality code through reviews, testing, and comprehensive documentation


Required Skills & Experience


React & JavaScript Mastery

  • 5+ years of professional React.js development with deep understanding of React internals, hooks, and advanced patterns
  • Expert-level JavaScript (ES6+) with strong proficiency in asynchronous programming, closures, and memory management
  • Advanced HTML5 & CSS3 skills with focus on performance and cross-browser compatibility


Real-time & Performance Expertise

  • Proven experience with WebSockets and real-time data streaming protocols
  • Strong DOM manipulation skills - direct DOM access, virtual scrolling, efficient updates, and performance optimization
  • RESTful API integration with experience in handling high-frequency data feeds
  • Browser performance optimization - understanding of rendering pipeline, memory management, and profiling tools


Development Tools & Practices

  • Proficiency with modern build tools: Webpack, Babel, Vite, or similar
  • Experience with Git version control and collaborative development workflows
  • Agile/Scrum development environment experience
  • Understanding of testing frameworks (Jest, React Testing Library)


Financial Data Visualization

  • Experience with financial charting libraries: Chartjs, TradingView, D3.js, or custom visualization solutions
  • Understanding of market data structures, order books, and trading terminology
  • Knowledge of data streaming optimization techniques for financial applications


Nice-to-Have Skills


Domain Expertise

  • Prior experience in stock market, trading, or financial services - understanding of trading workflows, order management, risk systems
  • Algorithmic trading knowledge or exposure to quantitative trading systems
  • Financial market understanding - equities, derivatives, commodities


Technical Plus Points

  • Backend development experience with GoLang, Python, or Node.js
  • Database knowledge: SQL, NoSQL, time-series databases (InfluxDB, TimescaleDB)
  • Cloud platform experience: AWS, Azure, GCP for deploying scalable applications
  • Message queue systems: Redis, RabbitMQ, Kafka, NATS for real-time data processing
  • Microservices architecture understanding and API design principles


Advanced Skills

  • Service Worker implementation for offline-first applications
  • Progressive Web App (PWA) development
  • Mobile-first responsive design expertise


Qualifications

  • Bachelor's degree in Computer Science, Engineering, or related field (or equivalent professional experience)
  • 5+ years of professional React.js development with demonstrable experience in performance-critical applications
  • Portfolio or examples of complex real-time applications you've built
  • Financial services experience strongly preferred


Why You'll Love Working Here


We're a team that hustles—plain and simple. But we also believe life outside work matters. No cubicles, no suits—just great people doing great work in a space built for comfort and creativity.


What We Offer

💰 Competitive salary – Get paid what you're worth

🌴 Generous paid time off – Recharge and come back sharper

🌍 Work with the best – Collaborate with top-tier global talent

✈️ Adventure together – Annual offsites (mostly outside India) and regular team outings

🎯 Performance rewards – Multiple bonuses for those who go above and beyond

🏥 Health covered – Comprehensive insurance so you're always protected

Fun, not just work – On-site sports, games, and a lively workspace

🧠 Learn and lead – Regular knowledge-sharing sessions led by your peers

📚 Annual Education Stipend – Take any external course, bootcamp, or certification that makes you better at your craft

🏋️ Stay fit – Gym memberships with equal employer contribution to keep you at your best

🚚 Relocation support – Smooth move? We've got your back

🏆 Friendly competition – Work challenges and extracurricular contests to keep things exciting


We work hard, play hard, and grow together. Join us.



Read more
iRage

at iRage

3 recruiters
Jyosana Jadhav
Posted by Jyosana Jadhav
Mumbai
4 - 7 yrs
₹25L - ₹35L / yr
skill iconPython
DevOps
CI/CD
Infrastructure
Deployment management
+14 more

We are seeking an experienced Operations Lead to drive operational excellence and lead a dynamic team in our fast-paced environment. The ideal candidate will combine strong technical expertise in Python with proven leadership capabilities to optimize processes, ensure system reliability, and deliver results.


Key Responsibilities


  • Team & stakeholder leadership - Lead 3-4 operations professionals and work cross-functionally with developers, system administrators, quants, and traders


  • DevOps automation & deployment - Develop deployment pipelines, automate configuration management, and build Python-based tools for operational processes and system optimization


  • Technical excellence & standards - Drive code reviews, establish development standards, ensure regional consistency with DevOps practices, and maintain technical documentation


  • System operations & performance - Monitor and optimize system performance for high availability, scalability, and security while managing day-to-day operations


  • Incident management & troubleshooting - Coordinate incident response, resolve infrastructure and deployment issues, and implement automated solutions to prevent recurring problems


  • Strategic technical leadership - Make infrastructure decisions, identify operational requirements, design scalable architecture, and stay current with industry best practices


  • Reporting & continuous improvement - Report on operational metrics and KPIs to senior leadership while actively contributing to DevOps process improvements


Qualifications and Experience


  • Bachelor's degree in Computer Science, Engineering, or related technical field
  • Proven experience of at least 5 years as a Software Engineer including at least 2 years as a DevOps Engineer or similar role, working with complex software projects and environments.
  • Excellent knowledge with cloud technologies, containers and orchestration.
  • Proficiency in scripting and programming languages such as Python and Bash.
  • Experience with Linux operating systems and command-line tools.
  • Proficient in using Git for version control.


Good to Have


  • Experience with Nagios or similar monitoring and alerting systems
  • Backend and/or frontend development experience for operational tooling
  • Previous experience working in a trading firm or financial services environment
  • Knowledge of database management and SQL
  • Familiarity with cloud platforms (AWS, Azure, GCP)
  • Experience with DevOps practices and CI/CD pipelines
  • Understanding of network protocols and system administration


Why You’ll Love Working Here


We’re a team that hustles—plain and simple. But we also believe life outside work matters. No cubicles, no suits—just great people doing great work in a space built for comfort and creativity.


Here’s what we offer:

💰 Competitive salary – Get paid what you’re worth.

🌴 Generous paid time off – Recharge and come back sharper.

🌍 Work with the best – Collaborate with top-tier global talent.

✈️ Adventure together – Annual offsites (mostly outside India) and regular team outings.

🎯 Performance rewards – Multiple bonuses for those who go above and beyond.

🏥 Health covered – Comprehensive insurance so you’re always protected.

Fun, not just work – On-site sports, games, and a lively workspace.

🧠 Learn and lead – Regular knowledge-sharing sessions led by your peers.

📚 Annual Education Stipend – Take any external course, bootcamp, or certification that makes you better at your craft.

🏋️ Stay fit – Gym memberships with equal employer contribution to keep you at your best.

🚚 Relocation support – Smooth move? We’ve got your back.

🏆 Friendly competition – Work challenges and extracurricular contests to keep things exciting.


We work hard, play hard, and grow together. Join us.


(P.S. We hire for talent, not pedigree—but if you’ve worked at a top tech co or fintech startup, we’d love to hear how you’ve shipped great products.)


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Anurag Sinha
Posted by Anurag Sinha
Pune, Mumbai, Navi Mumbai
5 - 8 yrs
Best in industry
Google Cloud Platform (GCP)
Azure Cloud
Terraform
DevOps
  • Looking manage IaC module
  • Terraform experience is a must
  • Terraform Module as a part of central platform team
  • Azure/GCP exp is a must
  • C#/Python/Java coding – is good to have

 

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Amita Soni
Posted by Amita Soni
Pune, Bengaluru (Bangalore), Mumbai
4 - 7 yrs
Best in industry
Google Cloud Platform (GCP)
GKE
Microsoft Windows Azure
Terraform

Job Description


We are looking for an experienced GCP Cloud Engineer to design, implement, and manage cloud-based solutions on Google Cloud Platform (GCP). The ideal candidate should have expertise in GKE (Google Kubernetes Engine), Cloud Run, Cloud Loadbalancer, Cloud function, Azure DevOps, and Terraform, with a strong focus on automation, security, and scalability.


Work location: Pune/Mumbai/Bangalore


Experience: 4-7 Years 


Joining: Mid of October


You will work closely with development, operations, and security teams to ensure robust cloud infrastructure and CI/CD pipelines while optimizing performance and cost.


Key Responsibilities:

1. Cloud Infrastructure Design & Management

· Architect, deploy, and maintain GCP cloud resources via terraform/other automation.

· Implement Google Cloud Storage, Cloud SQL, file store, for data storage and processing needs.

· Manage and configure Cloud Load Balancers (HTTP(S), TCP/UDP, and SSL Proxy) for high availability and scalability.

· Optimize resource allocation, monitoring, and cost efficiency across GCP environments.

2. Kubernetes & Container Orchestration

· Deploy, manage, and optimize workloads on Google Kubernetes Engine (GKE).

· Work with Helm charts, Istio, and service meshes for microservices deployments.

· Automate scaling, rolling updates, and zero-downtime deployments.


3. Serverless & Compute Services

· Deploy and manage applications on Cloud Run and Cloud Functions for scalable, serverless workloads.

· Optimize containerized applications running on Cloud Run for cost efficiency and performance.


4. CI/CD & DevOps Automation

· Design, implement, and manage CI/CD pipelines using Azure DevOps.

· Automate infrastructure deployment using Terraform, Bash and Power shell scripting

· Integrate security and compliance checks into the DevOps workflow (DevSecOps).


Required Skills & Qualifications:

✔ Experience: 4+ years in Cloud Engineering, with a focus on GCP.

✔ Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions).

✔ Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm.

✔ DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation.

✔ Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources.

✔ Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation.

✔ Security & Compliance: Knowledge of cloud security principles, IAM, and compliance standards.


About Wissen Technology

Wissen Technology, established in 2015 and part of the Wissen Group (founded in 2000), is a specialized technology consulting company. We pride ourselves on delivering high-quality solutions for global organizations across Banking & Finance, Telecom, and Healthcare domains.

 

Here’s why Wissen Technology stands out:

 

Global Presence: Offices in US, India, UK, Australia, Mexico, and Canada.

Expert Team: Wissen Group comprises over 4000 highly skilled professionals worldwide, with Wissen Technology contributing 1400 of these experts. Our team includes graduates from prestigious institutions such as Wharton, MIT, IITs, IIMs, and NITs.

Recognitions: Great Place to Work® Certified.

Featured as a Top 20 AI/ML Vendor by CIO Insider (2020).

Impressive Growth: Achieved 400% revenue growth in 5 years without external funding.

Successful Projects: Delivered $650 million worth of projects to 20+ Fortune 500 companies.

 

For more details:

 

Website: www.wissen.com 

Wissen Thought leadership : https://www.wissen.com/articles/ 

 

LinkedIn: Wissen Technology

Read more
Deqode

at Deqode

1 recruiter
Shubham Das
Posted by Shubham Das
Mumbai, Chennai, Gurugram
6 - 9 yrs
₹12L - ₹17L / yr
skill iconAmazon Web Services (AWS)
Windows Azure
Google Cloud Platform (GCP)
DevOps
helm
+3 more

We are looking for a highly skilled DevOps/Cloud Engineer with over 6 years of experience in infrastructure automation, cloud platforms, networking, and security. If you are passionate about designing scalable systems and love solving complex cloud and DevOps challenges—this opportunity is for you.

Key Responsibilities

  • Design, deploy, and manage cloud-native infrastructure using Kubernetes (K8s), Helm, Terraform, and Ansible
  • Automate provisioning and orchestration workflows for cloud and hybrid environments
  • Manage and optimize deployments on AWS, Azure, and GCP for high availability and cost efficiency
  • Troubleshoot and implement advanced network architectures including VPNs, firewalls, load balancers, and routing protocols
  • Implement and enforce security best practices: IAM, encryption, compliance, and vulnerability management
  • Collaborate with development and operations teams to improve CI/CD workflows and system observability

Required Skills & Qualifications

  • 6+ years of experience in DevOps, Infrastructure as Code (IaC), and cloud-native systems
  • Expertise in Helm, Terraform, and Kubernetes
  • Strong hands-on experience with AWS and Azure
  • Solid understanding of networking, firewall configurations, and security protocols
  • Experience with CI/CD tools like Jenkins, GitHub Actions, or similar
  • Strong problem-solving skills and a performance-first mindset

Why Join Us?

  • Work on cutting-edge cloud infrastructure across diverse industries
  • Be part of a collaborative, forward-thinking team
  • Flexible hybrid work model – work from anywhere while staying connected
  • Opportunity to take ownership and lead critical DevOps initiatives


Read more
Cymetrix Software

at Cymetrix Software

2 candid answers
Netra Shettigar
Posted by Netra Shettigar
Mumbai
4 - 8 yrs
₹8L - ₹16L / yr
skill iconC#
skill icon.NET
.NET Compact Framework
SQL
Microsoft Windows Azure
+4 more

Key Responsibilities:

● Design, develop, and maintain scalable web applications using .NET Core, .NET

Framework, C#, and related technologies.

● Participate in all phases of the SDLC, including requirements gathering, architecture

design, coding, testing, deployment, and support.

● Build and integrate RESTful APIs, and work with SQL Server, Entity Framework, and

modern front-end technologies such as Angular, React, and JavaScript.

● Conduct thorough code reviews, write unit tests, and ensure adherence to coding

standards and best practices.

● Lead or support .NET Framework to .NET Core migration initiatives, ensuring

minimal disruption and optimal performance.

● Implement and manage CI/CD pipelines using tools like Azure DevOps, Jenkins, or

GitLab CI/CD.

● Containerize applications using Docker and deploy/manage them on orchestration

platforms like Kubernetes or GKE.

● Lead and execute database migration projects, particularly transitioning from SQL

Server to PostgreSQL.

● Manage and optimize Cloud SQL for PostgreSQL, including configuration, tuning, and

ongoing maintenance.

● Leverage Google Cloud Platform (GCP) services such as GKE, Cloud SQL, Cloud

Run, and Dataflow to build and maintain cloud-native solutions.

● Handle schema conversion and data transformation tasks as part of migration and

modernization efforts.


Required Skills & Experience:

● 5+ years of hands-on experience with C#, .NET Core, and .NET Framework.

● Proven experience in application modernization and cloud-native development.


● Strong knowledge of containerization (Docker) and orchestration tools like

Kubernetes/GKE.

● Expertise in implementing and managing CI/CD pipelines.

● Solid understanding of relational databases and experience in SQL Server to

PostgreSQL migrations.

● Familiarity with cloud infrastructure, especially GCP services relevant to application

hosting and data processing.

● Excellent problem-solving, communication,

Read more
Wissen Technology

at Wissen Technology

4 recruiters
Vishakha Walunj
Posted by Vishakha Walunj
Mumbai
5 - 10 yrs
Best in industry
Terraform
Windows Azure
Google Cloud Platform (GCP)
DevOps
CI/CD

Key Skills Required:

  • Strong hands-on experience with Terraform
  • Proficiency in CI/CD tools (e.g., Jenkins, GitLab CI, Azure DevOps, etc.)
  • Experience working on Azure or GCP cloud platforms (at least one is mandatory)
  • Good understanding of DevOps practices


Read more
Wissen Technology

at Wissen Technology

4 recruiters
Praffull Shinde
Posted by Praffull Shinde
Pune, Mumbai, Bengaluru (Bangalore)
4 - 8 yrs
₹14L - ₹26L / yr
skill iconPython
PySpark
skill iconDjango
skill iconFlask
RESTful APIs
+3 more

Job title - Python developer

Exp – 4 to 6 years

Location – Pune/Mum/B’lore

 

PFB JD

Requirements:

  • Proven experience as a Python Developer
  • Strong knowledge of core Python and Pyspark concepts
  • Experience with web frameworks such as Django or Flask
  • Good exposure to any cloud platform (GCP Preferred)
  • CI/CD exposure required
  • Solid understanding of RESTful APIs and how to build them
  • Experience working with databases like Oracle DB and MySQL
  • Ability to write efficient SQL queries and optimize database performance
  • Strong problem-solving skills and attention to detail
  • Strong SQL programing (stored procedure, functions)
  • Excellent communication and interpersonal skill

Roles and Responsibilities

  • Design, develop, and maintain data pipelines and ETL processes using pyspark
  • Work closely with data scientists and analysts to provide them with clean, structured data.
  • Optimize data storage and retrieval for performance and scalability.
  • Collaborate with cross-functional teams to gather data requirements.
  • Ensure data quality and integrity through data validation and cleansing processes.
  • Monitor and troubleshoot data-related issues to ensure data pipeline reliability.
  • Stay up to date with industry best practices and emerging technologies in data engineering.
Read more
Wissen Technology

at Wissen Technology

4 recruiters
Shikha Nagar
Posted by Shikha Nagar
Pune, Mumbai, Bengaluru (Bangalore)
8 - 10 yrs
Best in industry
Terraform
Google Cloud Platform (GCP)
skill iconKubernetes
DevOps
SQL Azure

We are looking for an experienced GCP Cloud Engineer to design, implement, and manage cloud-based solutions on Google Cloud Platform (GCP). The ideal candidate should have expertise in GKE (Google Kubernetes Engine), Cloud Run, Cloud Loadbalancer, Cloud function, Azure DevOps, and Terraform, with a strong focus on automation, security, and scalability.

You will work closely with development, operations, and security teams to ensure robust cloud infrastructure and CI/CD pipelines while optimizing performance and cost.


Key Responsibilities:

1. Cloud Infrastructure Design & Management

· Architect, deploy, and maintain GCP cloud resources via terraform/other automation.

· Implement Google Cloud Storage, Cloud SQL, file store, for data storage and processing needs.

· Manage and configure Cloud Load Balancers (HTTP(S), TCP/UDP, and SSL Proxy) for high availability and scalability.

· Optimize resource allocation, monitoring, and cost efficiency across GCP environments.

2. Kubernetes & Container Orchestration

· Deploy, manage, and optimize workloads on Google Kubernetes Engine (GKE).

· Work with Helm charts, Istio, and service meshes for microservices deployments.

· Automate scaling, rolling updates, and zero-downtime deployments.


3. Serverless & Compute Services

· Deploy and manage applications on Cloud Run and Cloud Functions for scalable, serverless workloads.

· Optimize containerized applications running on Cloud Run for cost efficiency and performance.


4. CI/CD & DevOps Automation

· Design, implement, and manage CI/CD pipelines using Azure DevOps.

· Automate infrastructure deployment using Terraform, Bash and Power shell scripting

· Integrate security and compliance checks into the DevOps workflow (DevSecOps).


Required Skills & Qualifications:

✔ Experience: 8+ years in Cloud Engineering, with a focus on GCP.

✔ Cloud Expertise: Strong knowledge of GCP services (GKE, Compute Engine, IAM, VPC, Cloud Storage, Cloud SQL, Cloud Functions).

✔ Kubernetes & Containers: Experience with GKE, Docker, GKE Networking, Helm.

✔ DevOps Tools: Hands-on experience with Azure DevOps for CI/CD pipeline automation.

✔ Infrastructure-as-Code (IaC): Expertise in Terraform for provisioning cloud resources.

✔ Scripting & Automation: Proficiency in Python, Bash, or PowerShell for automation.

✔ Security & Compliance: Knowledge of cloud security principles, IAM, and compliance standards.


Read more
Fatakpay

at Fatakpay

2 recruiters
Disha Gajra
Posted by Disha Gajra
Mumbai
8 - 12 yrs
₹20L - ₹50L / yr
skill iconAmazon Web Services (AWS)
Windows Azure
Google Cloud Platform (GCP)
Cyber Security

Job Title: IT Head – Fintech Industry

Department: Information Technology

Location: Andheri East

Reports to: COO

Job Type: Full-Time

Job Overview:

The IT Head in a fintech company is responsible for overseeing the entire information technology infrastructure, including the development, implementation, and maintenance of IT systems, networks, and software solutions. The role involves leading the IT team, managing technology projects, ensuring data security, and ensuring the smooth functioning of all technology operations. As the company scales, the IT Head will play a key role in enabling digital innovation, optimizing IT processes, and ensuring compliance with relevant regulations in the fintech sector.

Key Responsibilities:

1. IT Strategy and Leadership

  • Develop and execute the company’s IT strategy to align with the organization’s overall business goals and objectives, ensuring the integration of new technologies and systems.
  • Lead, mentor, and manage a team of IT professionals, setting clear goals, priorities, and performance expectations.
  • Stay up-to-date with industry trends and emerging technologies, providing guidance and recommending innovations to improve efficiency and security.
  • Oversee the design, implementation, and maintenance of IT systems that support fintech products, customer experience, and business operations.

2. IT Infrastructure Management

  • Oversee the management and optimization of the company’s IT infrastructure, including servers, networks, databases, and cloud services.
  • Ensure the scalability and reliability of IT systems to support the company’s growth and increasing demand for digital services.
  • Manage system updates, hardware procurement, and vendor relationships to ensure that infrastructure is cost-effective, secure, and high-performing.

3. Cybersecurity and Data Protection

  • Lead efforts to ensure the company’s IT infrastructure is secure, implementing robust cybersecurity measures to protect sensitive customer data, financial transactions, and intellectual property.
  • Develop and enforce data protection policies and procedures to ensure compliance with data privacy regulations (e.g., GDPR, CCPA, RBI, etc.).
  • Conduct regular security audits and vulnerability assessments, working with the security team to address potential risks proactively.

4. Software Development and Integration

  • Oversee the development and deployment of software applications and tools that support fintech operations, including payment gateways, loan management systems, and customer engagement platforms.
  • Collaborate with product teams to identify technological needs, integrate new features, and optimize existing products for improved performance and user experience.
  • Ensure the seamless integration of third-party platforms, APIs, and fintech partners into the company’s core systems.

5. IT Operations and Support

  • Ensure the efficient day-to-day operation of IT services, including helpdesk support, system maintenance, and troubleshooting.
  • Establish service level agreements (SLAs) for IT services, ensuring that internal teams and customers receive timely support and issue resolution.
  • Manage incident response, ensuring quick resolution of system failures, security breaches, or service interruptions.

6. Budgeting and Cost Control

  • Manage the IT department’s budget, ensuring cost-effective spending on technology, software, hardware, and IT services.
  • Analyze and recommend investments in new technologies and infrastructure that can improve business performance while optimizing costs.
  • Ensure the efficient use of IT resources and the appropriate allocation of budget to support business priorities.

7. Compliance and Regulatory Requirements

  • Ensure IT practices comply with relevant industry regulations and standards, such as financial services regulations, data privacy laws, and cybersecurity guidelines.
  • Work with legal and compliance teams to ensure that all systems and data handling procedures meet industry-specific regulatory requirements (e.g., PCI DSS, ISO 27001).
  • Provide input and guidance on IT-related regulatory audits and assessments, ensuring the organization is always in compliance.

8. Innovation and Digital Transformation

  • Drive innovation by identifying opportunities for digital transformation within the organization, using technology to streamline operations and enhance the customer experience.
  • Collaborate with other departments (marketing, customer service, product development) to introduce new fintech products and services powered by cutting-edge technology.
  • Oversee the implementation of AI, machine learning, and other advanced technologies to enhance business performance, operational efficiency, and customer satisfaction.

9. Vendor and Stakeholder Management

  • Manage relationships with external technology vendors, service providers, and consultants to ensure the company gets the best value for its investments.
  • Negotiate contracts, terms of service, and service level agreements (SLAs) with vendors and technology partners.
  • Ensure strong communication with business stakeholders, understanding their IT needs and delivering technology solutions that align with company objectives.

Qualifications and Skills:

Education:

  • Bachelor’s degree in Computer Science, Information Technology, Engineering, or a related field (Master’s degree or relevant certifications like ITIL, PMP, or CISSP are a plus).

Experience:

  • 8-12 years of experience in IT management, with at least 4 years in a leadership role, preferably within the fintech, banking, or technology industry.
  • Strong understanding of IT infrastructure, cloud computing, database management, and cybersecurity best practices.
  • Proven experience in managing IT teams and large-scale IT projects, especially in fast-paced, growth-driven environments.
  • Knowledge of fintech products and services, including digital payments, blockchain, and online lending platforms.

Skills:

  • Expertise in IT infrastructure management, cloud services (AWS, Azure, Google Cloud), and enterprise software.
  • Strong understanding of cybersecurity protocols, data protection laws, and IT governance frameworks.
  • Experience with software development and integration, particularly for fintech platforms.
  • Strong project management and budgeting skills, with a track record of delivering IT projects on time and within budget.
  • Excellent communication and leadership skills, with the ability to manage cross-functional teams and communicate complex technical concepts to non-technical stakeholders.
  • Ability to manage multiple priorities in a fast-paced, high-pressure environment.


Read more
Deqode

at Deqode

1 recruiter
Roshni Maji
Posted by Roshni Maji
Mumbai
3 - 6 yrs
₹8L - ₹13L / yr
skill iconAmazon Web Services (AWS)
Terraform
Ansible
skill iconDocker
Apache Kafka
+6 more

Must be:

  • Based in Mumbai
  • Comfortable with Work from Office
  • Available to join immediately


Responsibilities:

  • Manage, monitor, and scale production systems across cloud (AWS/GCP) and on-prem.
  • Work with Kubernetes, Docker, Lambdas to build reliable, scalable infrastructure.
  • Build tools and automation using Python, Go, or relevant scripting languages.
  • Ensure system observability using tools like NewRelic, Prometheus, Grafana, CloudWatch, PagerDuty.
  • Optimize for performance and low-latency in real-time systems using Kafka, gRPC, RTP.
  • Use Terraform, CloudFormation, Ansible, Chef, Puppet for infra automation and orchestration.
  • Load testing using Gatling, JMeter, and ensuring fault tolerance and high availability.
  • Collaborate with dev teams and participate in on-call rotations.


Requirements:

  • B.E./B.Tech in CS, Engineering or equivalent experience.
  • 3+ years in production infra and cloud-based systems.
  • Strong background in Linux (RHEL/CentOS) and shell scripting.
  • Experience managing hybrid infrastructure (cloud + on-prem).
  • Strong testing practices and code quality focus.
  • Experience leading teams is a plus.
Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort