13+ Google Cloud Storage Jobs in Bangalore (Bengaluru) | Google Cloud Storage Job openings in Bangalore (Bengaluru)
Apply to 13+ Google Cloud Storage Jobs in Bangalore (Bengaluru) on CutShort.io. Explore the latest Google Cloud Storage Job opportunities across top companies like Google, Amazon & Adobe.
Job Title: Infrastructure Engineer
Experience: 4.5Years +
Location: Bangalore
Employment Type: Full-Time
Joining: Immediate Joiner Preferred
💼 Job Summary
We are looking for a skilled Infrastructure Engineer to manage, maintain, and enhance our on-premise and cloud-based systems. The ideal candidate will have strong experience in server administration, virtualization, hybrid cloud environments, and infrastructure automation. This role requires hands-on expertise, strong troubleshooting ability, and the capability to collaborate with cross-functional teams.
Roles & Responsibilities
- Install, configure, and manage Windows and Linux servers.
- Maintain and administer Active Directory, DNS, DHCP, and file servers.
- Manage virtualization platforms such as VMware or Hyper-V.
- Monitor system performance, logs, and uptime to ensure high availability.
- Provide L2/L3 support, diagnose issues, and maintain detailed technical documentation.
- Deploy and manage cloud servers and resources in AWS, Azure, or Google Cloud.
- Design, build, and maintain hybrid environments (on-premises + cloud).
- Administer data storage systems and implement/test backup & disaster recovery plans.
- Handle cloud services such as cloud storage, networking, and identity (IAM, Azure AD).
- Ensure compliance with security standards like ISO, SOC, GDPR, PCI DSS.
- Integrate and manage monitoring and alerting tools.
- Support CI/CD pipelines and automation for infrastructure deployments.
- Collaborate with Developers, DevOps, and Network teams for seamless system integration.
- Troubleshoot and resolve complex infrastructure & system-level issues.
Key Skills Required
- Windows Server & Linux Administration
- VMware / Hyper-V / Virtualization technologies
- Active Directory, DNS, DHCP administration
- Knowledge of CI/CD and Infrastructure as Code
- Hands-on experience in AWS, Azure, or GCP
- Experience with cloud migration and hybrid cloud setups
- Proficiency in backup, replication, and disaster recovery tools
- Familiarity with automation tools (Terraform, Ansible, etc. preferred)
- Strong troubleshooting and documentation skills
- Understanding of networking concepts (TCP/IP, VPNs, firewalls, routing) is an added advantage
Job Details
- Job Title: Lead II - Software Engineering- AI, NLP, Python, Data science
- Industry: Technology
- Domain - Information technology (IT)
- Experience Required: 7-9 years
- Employment Type: Full Time
- Job Location: Bangalore
- CTC Range: Best in Industry
Job Description:
Role Proficiency:
Act creatively to develop applications by selecting appropriate technical options optimizing application development maintenance and performance by employing design patterns and reusing proven solutions. Account for others' developmental activities; assisting Project Manager in day-to-day project execution.
Additional Comments:
Mandatory Skills Data Science Skill to Evaluate AI, Gen AI, RAG, Data Science
Experience 8 to 10 Years
Location Bengaluru
Job Description
Job Title AI Engineer Mandatory Skills Artificial Intelligence, Natural Language Processing, python, data science Position AI Engineer – LLM & RAG Specialization Company Name: Sony India Software Centre About the role: We are seeking a highly skilled AI Engineer with 8-10 years of experience to join our innovation-driven team. This role focuses on the design, development, and deployment of advanced enterprise-scale Large Language Models (eLLM) and Retrieval Augmented Generation (RAG) solutions. You will work on end-to-end AI pipelines, from data processing to cloud deployment, delivering impactful solutions that enhance Sony’s products and services. Key Responsibilities: Design, implement, and optimize LLM-powered applications, ensuring high performance and scalability for enterprise use cases. Develop and maintain RAG pipelines, including vector database integration (e.g., Pinecone, Weaviate, FAISS) and embedding model optimization. Deploy, monitor, and maintain AI/ML models in production, ensuring reliability, security, and compliance. Collaborate with product, research, and engineering teams to integrate AI solutions into existing applications and workflows. Research and evaluate the latest LLM and AI advancements, recommending tools and architectures for continuous improvement. Preprocess, clean, and engineer features from large datasets to improve model accuracy and efficiency. Conduct code reviews and enforce AI/ML engineering best practices. Document architecture, pipelines, and results; present findings to both technical and business stakeholders. Job Description: 8-10 years of professional experience in AI/ML engineering, with at least 4+ years in LLM development and deployment. Proven expertise in RAG architectures, vector databases, and embedding models. Strong proficiency in Python; familiarity with Java, R, or other relevant languages is a plus. Experience with AI/ML frameworks (PyTorch, TensorFlow, etc.) and relevant deployment tools. Hands-on experience with cloud-based AI platforms such as AWS SageMaker, AWS Q Business, AWS Bedrock or Azure Machine Learning. Experience in designing, developing, and deploying Agentic AI systems, with a focus on creating autonomous agents that can reason, plan, and execute tasks to achieve specific goals. Understanding of security concepts in AI systems, including vulnerabilities and mitigation strategies. Solid knowledge of data processing, feature engineering, and working with large-scale datasets. Experience in designing and implementing AI-native applications and agentic workflows using the Model Context Protocol (MCP) is nice to have. Strong problem-solving skills, analytical thinking, and attention to detail. Excellent communication skills with the ability to explain complex AI concepts to diverse audiences. Day-to-day responsibilities: Design and deploy AI-driven solutions to address specific security challenges, such as threat detection, vulnerability prioritization, and security automation. Optimize LLM-based models for various security use cases, including chatbot development for security awareness or automated incident response. Implement and manage RAG pipelines for enhanced LLM performance. Integrate AI models with existing security tools, including Endpoint Detection and Response (EDR), Threat and Vulnerability Management (TVM) platforms, and Data Science/Analytics platforms. This will involve working with APIs and understanding data flows. Develop and implement metrics to evaluate the performance of AI models. Monitor deployed models for accuracy and performance and retrain as needed. Adhere to security best practices and ensure that all AI solutions are developed and deployed securely. Consider data privacy and compliance requirements. Work closely with other team members to understand security requirements and translate them into AI-driven solutions. Communicate effectively with stakeholders, including senior management, to present project updates and findings. Stay up to date with the latest advancements in AI/ML and security and identify opportunities to leverage new technologies to improve our security posture. Maintain thorough documentation of AI models, code, and processes. What We Offer Opportunity to work on cutting-edge LLM and RAG projects with global impact. A collaborative environment fostering innovation, research, and skill growth. Competitive salary, comprehensive benefits, and flexible work arrangements. The chance to shape AI-powered features in Sony’s next-generation products. Be able to function in an environment where the team is virtual and geographically dispersed
Education Qualification: Graduate
Skills: AI, NLP, Python, Data science
Must-Haves
Skills
AI, NLP, Python, Data science
NP: Immediate – 30 Days
- Own and Maintain our interface. Built in Go, React and an event-driven architecture under the hood
- We work with complex data types and relations. Design our database (postgres) and maintain our S3 buckets.
- Build an infra which will work super-fast on cloud servers and runs reliably on-prem
- Develop a user-centric platform by helping us with ruthless prioritization of features
- Have an eye for detail especially when it comes to visual appeal
What you bring to the table
- You have full-stack experience with relevant technologies (Go, React, Typescript)
- You know how to go from 0-to-1
- You have strong communication skills
- You thrive in a collaborative team setting
InViz is Bangalore Based Startup helping Enterprises simplifying the Search and Discovery experiences for both their end customers as well as their internal users. We use state-of-the-art technologies in Computer Vision, Natural Language Processing, Text Mining, and other ML techniques to extract information/concepts from data of different formats- text, images, videos and make them easily discoverable through simple human-friendly touchpoints.
TSDE - Data
Data Engineer:
- Should have total 3-6 Yrs of experience in Data Engineering.
- Person should have experience in coding data pipeline on GCP.
- Prior experience on Hadoop systems is ideal as candidate may not have total GCP experience.
- Strong on programming languages like Scala, Python, Java.
- Good understanding of various data storage formats and it’s advantages.
- Should have exposure on GCP tools to develop end to end data pipeline for various scenarios (including ingesting data from traditional data bases as well as integration of API based data sources).
- Should have Business mindset to understand data and how it will be used for BI and Analytics purposes.
- Data Engineer Certification preferred
|
Experience in Working with GCP tools like |
|
|
Store : CloudSQL , Cloud Storage, Cloud Bigtable, Bigquery, Cloud Spanner, Cloud DataStore |
|
|
Ingest : Stackdriver, Pub/Sub, AppEngine, Kubernete Engine, Kafka, DataPrep , Micro services |
|
|
Schedule : Cloud Composer |
|
|
Processing: Cloud Dataproc, Cloud Dataflow, Cloud Dataprep |
|
|
CI/CD - Bitbucket+Jenkinjs / Gitlab |
|
|
Atlassian Suite |
|
|
|
.
Want to work with an established & growing IT company? Join team Benison to have the right challenges that will help you accelerate your career growth to the next level, faster!
Benison Technologies was started in 2011 with a mission to revolutionize the silicon industry in India, with a host of amazing big clients like Google, Cisco, McAfee, Intel, and so on, you get to experience the best of both worlds. If you consider yourself an engineer who is capable to join our ever-growing team, then this is the right opportunity for you:
Why Benison Tech?
We have a partial acquisition from one of the biggest names in the world (well we can’t name them thanks to confidentiality) it’s one of the FAANG companies, and you can “Google” it if you like.
Oh! & one more thing, this did not happen by accident, our team put a ton of efforts to turn this gigantic dream into a reality.
Benison Tech has a consistent history of demonstrating growth through innovation time and again.
We don’t stop there, we then re-invest our profits back into the initiatives for the growth of our people, our culture and the company. Now enough with us, let’s talk about the job roles & responsibilities:
What you will be working on:
- Key contributor for developing product strategies and features.
- Software development for industries leading SaaS platform
- You will be involved closely in planning, designing, integration of client requirements.
- You will be working with one of the leaders in data resiliency and data protection.
Here are some technical skills require:
- Independently own features and create feature test plans/strategies based on development and feature completion milestones.
- Identify quality assurance process bottlenecks and suggest actions for improvement.
- Design automation framework for automating feature tests.
- Participate in test cases, test plans, s and code reviews.
- Resolve functional queries coming from other business units such as support, escalation, product management, etc.
- Participate in bug trailing, tracking quality assurance metrics.
- Hands-on experience with Python-Selenium or Cypress, will be preferred.
- Familiarity with Test Management systems like XRay and bug tracker like JIRA tools.
What we expect from you:
- 3-10 Years of relevant experience in QA Automation.
- Expert at test automation, creating test plans, test strategies for testing multiple product modules
- Should be able to quickly analyze failures and trace back to issues in the product or the automation suite.
- As a Software Development Engineer in Test you should be an expert at test automation for APIs as well as UI, creating test plans and test strategies for testing product features.
- You will guide and mentor junior team members by reviewing their automation code and test cases to ensure good coverage and quality of a feature
- Resolve functional queries coming from other business units such as support, escalation, product management, etc.
- Be a quick learner and be open to working on new technologies if needed.
- Excellent team player with strong verbal & written communication skills.
- Be able to step up when the situation demands such as meeting deadlines and critical production issues.
- Propose changes or enhancements to the framework for enabling new feature tests.
Few Skills which will add brownie points to your role
- Working knowledge of Dockers and Kubernetes will be an advantage
- Awareness of general manual and automation concepts and all types of testing methods
- Knowledge of the Backup or Storage domain will be an advantage.
If the above fits your skill-sets and tickles your interest then read below about the additional benefits that our company offers to talented folks like you:
Work Culture and Benefits
- Competitive salary and benefits package
(H1-B which means a chance to work onsite out of India) - A culture focused on talent development where you get promoted within the quarterly cycle of your anniversary.
- Opportunity to work with cutting-edge & challenging technologies including legacy tech.
- Open cafeteria to grab some munchies while you work, we make sure the space feels like your second home, you can also wear pyjamas if you like.
- Employee engagement initiatives such as project parties, flexible work hours, and long service awards, team bonding activities within the company, extra learning and personal development trainings, because why stop your learning at one thing!
- Insurance coverage: Group term life, personal accident, and Mediclaim hospitalization for self, spouse, two children, and your parents. (With some of the best insurance partners in India)
- Enjoy collaborative innovation (each members gets to innovate & think out of the box), along with highly experienced team managers who maintain diversity and work-life well-being.
- And of course, you get to work on projects from some of the most recognised brands within the networking and security space of the world, unlocking global opportunities to learn, grow & contribute in a way that is truly impactful yet purposeful at the same time.
Still not satisfied, and want more proof?
Head to our website https://benisontech.com/">https://benisontech.com to learn more.
As a MLOps Engineer in QuantumBlack you will:
Develop and deploy technology that enables data scientists and data engineers to build, productionize and deploy machine learning models following best practices. Work to set the standards for SWE and
DevOps practices within multi-disciplinary delivery teams
Choose and use the right cloud services, DevOps tooling and ML tooling for the team to be able to produce high-quality code that allows your team to release to production.
Build modern, scalable, and secure CI/CD pipelines to automate development and deployment
workflows used by data scientists (ML pipelines) and data engineers (Data pipelines)
Shape and support next generation technology that enables scaling ML products and platforms. Bring
expertise in cloud to enable ML use case development, including MLOps
Our Tech Stack-
We leverage AWS, Google Cloud, Azure, Databricks, Docker, Kubernetes, Argo, Airflow, Kedro, Python,
Terraform, GitHub actions, MLFlow, Node.JS, React, Typescript amongst others in our projects
Key Skills:
• Excellent hands-on expert knowledge of cloud platform infrastructure and administration
(Azure/AWS/GCP) with strong knowledge of cloud services integration, and cloud security
• Expertise setting up CI/CD processes, building and maintaining secure DevOps pipelines with at
least 2 major DevOps stacks (e.g., Azure DevOps, Gitlab, Argo)
• Experience with modern development methods and tooling: Containers (e.g., docker) and
container orchestration (K8s), CI/CD tools (e.g., Circle CI, Jenkins, GitHub actions, Azure
DevOps), version control (Git, GitHub, GitLab), orchestration/DAGs tools (e.g., Argo, Airflow,
Kubeflow)
• Hands-on coding skills Python 3 (e.g., API including automated testing frameworks and libraries
(e.g., pytest) and Infrastructure as Code (e.g., Terraform) and Kubernetes artifacts (e.g.,
deployments, operators, helm charts)
• Experience setting up at least one contemporary MLOps tooling (e.g., experiment tracking,
model governance, packaging, deployment, feature store)
• Practical knowledge delivering and maintaining production software such as APIs and cloud
infrastructure
• Knowledge of SQL (intermediate level or more preferred) and familiarity working with at least
one common RDBMS (MySQL, Postgres, SQL Server, Oracle)
- Good experience and exposure of cloud native architecture, development and deployment on public clouds AWS, Google Cloud etc
- Responsible for Linux server installation, maintain, monitoring, data backup and recovery, securtiy and administration
- Understanding of clusters, distributed architecture, container environment
- Experience in networking, including linux, software defined networking, network virtualization, open protocols, application acceleration and load balancing, DNS, virtual private networks
- Knowledge of several middleware such as MySQL, Apache etc
- Responsible for managing network storage
- Disaster recovery and incident response planning
- Configuring/monitoring firewalls, routers, switches and other network devices
Responsibilities and Duties
- Support the globally distributed cloud development and teams by mainitaning the cloud infrastructure labs hosted in a hybrid cloud environment
- Contribute towards optimization of preformance and acost of running the labs
• Bachelor or Master Degree in Computer Science, Software Engineering from a reputed
University.
• 5 - 8 Years of experience in building scalable, secure and compliant systems.
• More than 2 years of experience in working with GCP deployment for millions of daily visitors
• 5+ years hosting experience in a large heavy-traffic environment
• 5+ years production application support experience in a high uptime environment
• Software development and monitoring knowledge with Automated builds
• Technology:
o Cloud: AWS or Google Cloud
o Source Control: Gitlab or Bitbucket or Github
o Container Concepts: Docker, Microservices
o Continuous Integration: Jenkins, Bamboos
o Infrastructure Automation: Puppet, Chef or Ansible
o Deployment Automation: Jenkins, VSTS or Octopus Deploy
o Orchestration: Kubernets, Mesos, Swarm
o Automation: Node JS or Python
o Linux environment network administration, DNS, firewall and security management
• Ability to be adapt to the startup culture, handle multiple competing priorities, meet
deadlines and troubleshoot problems.




