
Job Role: Adaptive Autosar + Bootloader Developer
Mandatory Skills:
- Adaptive Autosar Development
- Bootloader Experience
- C++ Programming
- Hands-on experience with ISO 14229 (UDS Protocol)
- Experience in Flash Bootloader and Software Update topics
- Proficient in C++ and Python programming
- Application development experience in Service-Oriented Architectures
- Hands-on experience with QNX and Linux operating systems
- Familiarity with software development tools like CAN Analyzer, CANoe, and Debugger
- Strong problem-solving skills and the ability to work independently
- Exposure to the ASPICE Process is an advantage
- Excellent analytical and communication skills
Job Responsibilities:
- Engage in tasks related to the integration and development of Flash Bootloader (FBL) features and perform comprehensive testing activities.
- Collaborate continuously with counterparts in Germany to understand requirements and develop FBL features effectively.
- Create test specifications and meticulously document testing results.
Why Join InfoGrowth?
- Become part of an innovative team focused on transforming the automotive industry with cutting-edge technology.
- Work on exciting projects that challenge your skills and promote professional growth.
- Enjoy a collaborative environment that values teamwork and creativity.
š Apply Now to shape the future of automotive technology with InfoGrowth!

Similar jobs
Job Title:Ā Senior Devops EngineerĀ (Full-time)
Location:Ā Mumbai, Onsite
Experience Required:Ā 5+ Years
Required Qualifications
ā Experience:
ā 5+ years of hands-on experience as a DevOps Engineer or similar role, with
proven expertise in building and customizing Helm charts from scratch (not just
using pre-existing ones).
ā Demonstrated ability to design and whiteboard DevOps pipelines, including
CI/CD workflows for microservices applications.
ā Experience packaging and deploying applications with stateful dependencies
(e.g., databases, persistent storage) in varied environments: on-prem (air-gapped
and non-air-gapped), single-tenant cloud, multi-tenant cloud, and developer trials.
ā Proficiency in managing deployments in Kubernetes clusters, including offline
installations, upgrades via Helm, and adaptations for client restrictions (e.g., no
additional tools or VMs).
ā Track record of handling client interactions, such as asking probing questions
about infrastructure (e.g., OS versions, storage solutions, network restrictions)
and explaining technical concepts clearly.
āĀ Technical Skills:
ā Strong knowledge of Helm syntax and functionalities (e.g., Go templating, hooks,
subcharts, dependency management).
ā Expertise in containerization with Docker, including image management
(save/load, registries like Harbor or ECR).
ā Familiarity with CI/CD tools such as Jenkins, ArgoCD, GitHub Actions, and
GitOps for automated and manual deployments.
ā Understanding of storage solutions for on-prem and cloud, including object/file
storage (e.g., MinIO, Ceph, NFS, cloud-native like S3/EBS).
ā In-depth knowledge of Kubernetes concepts: StatefulSets, PersistentVolumes,
namespaces, HPA, liveness/readiness probes, network policies, and RBAC.
ā Solid grasp of cloud networking: VPCs (definition, boundaries, virtualization via
SDN, differences from private clouds), bare metal vs. virtual machines
(advantages like resource efficiency, flexibility, and scalability).
ā Ability to work in air-gapped environments, preparing offline artifacts and
ensuring self-contained deployment
DevOps Engineer
Location: Bangalore office
About Peliqan
Peliqan is an all-in-one data platform combining ELT/ETL pipelines, a built-in data warehouse, SQL and low-code Python transformations, reverse ETL, and AI-powered data activation. We connect 250+ data sources and serve enterprise teams, consultants, and SaaS companies. SOC 2 Type II certified and GDPR compliant.
The Role
Own and evolve the infrastructure powering Peliqan's multi-tenant data platform. You'll manage Kubernetes clusters, cloud resources, CI/CD pipelines, and monitoring ā keeping everything reliable, secure, and scalable. You'll be the go-to person for infrastructure support across the engineering team.
Responsibilities
Manage and optimise Kubernetes clusters running production workloads ā data pipelines, APIs, and customer-facing services.
Maintain Docker-based local development environments for the engineering team.
Administer cloud infrastructure on AWS and Google Cloud (compute, storage, networking, managed databases).
Build and maintain CI/CD pipelines for automated testing, building, and deploying across staging and production.
Set up and manage monitoring, alerting, and logging for platform health and incident response.
Manage release processes ā deployments, rollbacks, and release strategies.
- Maintain infrastructure-as-code using Helm charts.
- Support security hardening and compliance efforts (SOC 2, GDPR).
Requirements
3+ years in a DevOps, SRE, or Infrastructure Engineering role.
Strong hands-on experience with Kubernetes and Helm charts.
Deep familiarity with Docker for containerisation and local dev workflows.
Production experience with AWS and/or Google Cloud.
- Proficiency in Python and Bash scripting for automation and tooling.
- Solid grasp of DevOps principles: infrastructure-as-code, GitOps, observability, continuous delivery.
- Experience with CI/CD platforms (GitHub Actions, GitLab CI, or similar).
Nice to Have
- Experience supporting multi-tenant SaaS platforms or data infrastructure at scale.
- Knowledge of PostgreSQL, MySQL, or cloud-managed database administration.
- Exposure to security compliance frameworks (SOC 2, ISO 27001, GDPR).
DESIRED SKILLS AND EXPERIENCE
ļ· Strong analytical and problem-solving skills
ļ· Ability to work independently, learn quickly and be proactive
ļ· 3-5 years overall and at least 1-2 years of hands-on experience in designing and managing DevOps Cloud infrastructure
ļ· Experience must include a combination of:
o Experience working with configuration management tools ā Ansible, Chef, Puppet, SaltStack (expertise in at least one tool is a must)
o Ability to write and maintain code in at least one scripting language (Python preferred)
o Practical knowledge of shell scripting
o Cloud knowledge ā AWS, VMware vSphere o Good understanding and familiarity with Linux
o Networking knowledge ā Firewalls, VPNs, Load Balancers
o Web/Application servers, Nginx, JVM environments
o Virtualization and containers - Xen, KVM, Qemu, Docker, Kubernetes, etc.
o Familiarity with logging systems - Logstash, Elasticsearch, Kibana
o Git, Jenkins, Jira
ApnaComplex is one of Indiaās largest and fastest-growing PropTech disruptors within the Society & Apartment Management business.Ā The SaaS based B2C platform is headquartered out of Indiaās tech start-up hub, Bangalore, with branches in 6 other cities. It currently empowers 3,600 Societies, managing over 6 Lakh Households in over 80 Indian cities to effortlessly manage all aspects of running large complexes seamlessly.
ApnaComplex is part of ANAROCK Group. ANAROCK Group is India's leading specialized real estate services company having diversified interests across the real estate value chain.
If it excites you to - drive innovation, create industry-first solutions, build new capabilities ground-up, and work with multiple new technologies, ApnaComplex is the place for you.
Ā
Must have-
Ā
- Knowledge of Docker
- Knowledge of Terraforms
- Knowledge of AWS
Ā
Good to have -
- Kubernetes
- Scripting language: PHP/Go Lang and Python
- Webserver knowledge
- Logging and monitoring experience
- Test, build, design, deployment, and ability to maintain continuous integration and continuous delivery process using tools like Jenkins, maven Git, etc.Ā
- Build and maintain highly available production systems.
- Must know how to choose the best tools and technologies which best fits the business needs.Ā
- Develop software to integrate with internal back-end systems.
- Investigate and resolve technical issues.
- Problem-solving attitude.
- Ability to automate test and deploy the code and monitor.Ā
- Work in close coordination with the development and operations team such that the application is in line with performance according to the customer's expectation.
- Lead and guide the team in identifying and implementing new technologies.
Ā
Ā
Skills that will help you build a success story with us
Ā
- An ability to quickly understand and solve new problems
- Strong interpersonal skills
- Excellent data interpretation
- Context-switching
- Intrinsically motivated
- A tactical and strategic track record for delivering research-driven results
Ā
Quick Glances:
Ā
- https://www.apnacomplex.com/why-apnacomplex">What to look for at ApnaComplex
- https://www.linkedin.com/company/1070467/admin/">Who are we A glimpse of ApnaComplex, know us better
- https://www.apnacomplex.com/media-buzz">ApnaComplex - Media ā Visit our media page Ā
Ā
ANAROCK Ethos - Values Over Value:
Our assurance of consistent ethical dealing with clients and partners reflects our motto - Values Over Value.
We value diversity within ANAROCK Group and are committed to offering equal opportunities in employment. We do not discriminate against any team member or applicant for employment based on nationality, race, color, religion, caste, gender identity / expression, sexual orientation, disability, social origin and status, indigenous status, political opinion, age, marital status or any other personal characteristics or status. ANAROCK Group values all talent and will do its utmost to hire, nurture and grow them.
Requirements:
ā Knowledge of building micro-services.
ā Experience in managing cloud infrastructure with disaster recovery and security in
mind (AWS, GCP, Azure).
ā Experience with High Availability clusters setup.
ā Experience in creating alerting and monitoring strategies.
ā Strong debugging skills.
ā Experience with 0 downtime Continuous Delivery setup (Jenkins, AWS Code
Deploy, Team City, Go CD etc).
ā Experience with Infrastructure as Code & Automation tools (Bash, Ansible,
Puppet, Chef, Terraform etc).
ā Master of *nix systems, including working with docker, process & network
monitoring tools.
ā Knowledge of monitoring tools like New Relic, App Dynamics etc.
ā Experience with Messaging systems (RMQ, Kafka etc. ).
ā Knowledge of DevOps Intelligence.
ā Experience in setting up & driving DevOps initiatives in side the org Excellen.
ā Good team player.
ā Good to have experience in Kubernetes cluster management.

- AWS Cloud, CICD, Serverless setups, Monitoring Setup
- Performance setup, scalability in hands experience, Linux expertise, DevOps Operations
- AWS Cloud, CICD, Serverless setups, Monitoring Setup, Performance setup, scalability in hands experience, Linux expertise, DevOps Operations.
We are looking for people with programming skills in Python, SQL, Cloud Computing. Candidate should have experience in at least one of the major cloud-computing platforms - AWS/Azure/GCP. He should professioanl experience in handling applications and databases in the cloud using VMs and Docker images. He should have ability to design and develop applications for the cloud.
Ā
You will be responsible for
- Leading the DevOps strategy and development of SAAS Product Deployments
- Leading and mentoring other computer programmers.
- Evaluating student work and providing guidance in the online courses in programming and cloud computing.
Ā
Desired experience/skills
Qualifications: Graduate degree in Computer Science or related field, or equivalent experience.
Ā
Skills:
- Strong programming skills in Python, SQL,
- Cloud Computing
Ā
Experience:
2+ years of programming experience including Python, SQL, and Cloud Computing. Familiarity with command line working environment.
Ā
Note:Ā A strong programming background, in any language and cloud computing platform is required. We are flexible about the degree of familiarity needed for the specific environments Python, SQL. If you have extensive experience in one of the cloud computing platforms and less in others you should still, consider applying.
Ā
Soft Skills:
- Good interpersonal, written, and verbal communication skills; including the ability to explain the concepts to others.
- A strong understanding of algorithms and data structures, and their performance characteristics.
- Awareness of and sensitivity to the educational goals of a multicultural population would also be desirable.
- Detail oriented and well organized.Ā Ā Ā
- You have experience of 2-4 years in building high-performance consumer-facing mobile applications at Product companies of a decent scale.
- You can write code preferably in Golang and Python.
- You have experience with debugging production issues and writing RCAs.
- You have demonstrable stories of being on-call and how outages have been handled.
- You have experience developing products on Kubernetes and cloud providers like GCP and AWS.
- You have worked with Cloud Native (CNCF) technologies.
- You have experience automating CI/CD pipelines.
- You are an excellent collaborator & communicator. You know that start-ups are a team sport.
- You listen to others, arenāt afraid to speak your mind and always try to ask the right questions.
- You are excited by the prospect of working in a distributed team and company
Ā
Ā
We are looking for an experienced software engineer with a strong background in DevOps and handling traffic & infrastructure at scale.
Responsibilities :
Work closely with product engineers to implement scalable and highly reliable systems.
Scale existing backend systems to handle ever-increasing amounts of traffic and new product requirements.
Collaborate with other developers to understand & setup tooling needed for - Continuous Integration/Delivery/
Build & operate infrastructure to support website, backend cluster, ML projects in the organization.
Monitor and track performance and reliability of our services and software to meet promised SLA
2+ years of experience working on distributed systems and shipping high-quality product features on schedule
Intimate knowledge of the whole web stack (Front end, APIs, database, networks etc.)
Ability to build highly scalable, robust, and fault-tolerant services and stay up-to-date with the latest architectural trends
Experience with container based deployment, microservices, in-memory caches, relational databases, key-value stores
Hands-on experience with cloud infrastructure provisioning, deployment, monitoring (we are on AWS and use ECS, RDS, ELB, EC2, Elasticache, Elasticsearch, S3, CloudWatch)
DevOps Engineer Skills Building a scalable and highly available infrastructure for data science Knows data science project workflows Hands-on with deployment patterns for online/offline predictions (server/serverless)
Experience with either terraform or Kubernetes
Experience of ML deployment frameworks like Kubeflow, MLflow, SageMaker Working knowledge of Jenkins or similar tool Responsibilities Owns all the ML cloud infrastructure (AWS) Help builds out an entirely CI/CD ecosystem with auto-scaling Work with a testing engineer to design testing methodologies for ML APIs Ability to research & implement new technologies Help with cost optimizations of infrastructure.
Knowledge sharing Nice to Have Develop APIs for machine learning Can write Python servers for ML systems with API frameworks Understanding of task queue frameworks like Celery








