
Bachelor’s degree in computer science or other technical/scientific discipline
● Solid experience with cloud environments like Azure, GCP, AWS
● Proficient with CI/CD tools such as CircleCI, Travis, Jenkins, Ansible, Terraform, Puppet, Chef
● Strong expertise in Serverless architecture, Kubernetes cluster, and service mesh.
● Strong hands-on experience with monitoring tools such as Prometheus
● Ability to program with one or more high-level languages, such as Python, Java, Ruby, and JavaScript
● Familiar with Git
● Experience with performance monitoring and optimization. Ideally worked on large scale e-commerce projects

Similar jobs
- Bachelor's degree in Computer Science, Engineering, or a related field (or equivalent experience)
- Proven 2 to 12 years of professional experience in Node.js development
- Strong proficiency in JavaScript, including ES6+ syntax
- Expertise in designing and developing RESTful APIs and microservices architecture
- Experience with popular Node.js frameworks such as Express.js, Nest.js, or Hapi.js
- Experience in modern front-end technologies (e.g., HTML, CSS, JavaScript frameworks like React, Angular, or Vue.js)
- Solid understanding of asynchronous programming, event-driven architecture, and design patterns
- Extensive knowledge of database systems, both SQL and NoSQL (e.g., MongoDB, PostgreSQL, MySQL)
- Experience with cloud platforms such as AWS, Azure, or Google Cloud Platform
- Familiarity with containerization technologies (e.g., Docker, Kubernetes)

About us
Astra is a cyber security SaaS company that makes otherwise chaotic pentests a breeze with its one of a kind Pentest Platform. Astra's continuous vulnerability scanner emulates hacker behavior to scan applications for 8300+ security tests. CTOs & CISOs love Astra because it helps them fix vulnerabilities in record time and move from DevOps to DevSecOps with Astra's CI/CD integrations.
Astra is loved by 650+ companies across the globe. In 2023 Astra uncovered 2 million+ vulnerabilities for its customers, saving customers $69M+ in potential losses due to security vulnerabilities.
We've been awarded by the President of France Mr. François Hollande at the La French Tech program and Prime Minister of India Shri Narendra Modi at the Global Conference on Cyber Security. Loom, MamaEarth, Muthoot Finance, Canara Robeco, ScripBox etc. are a few of Astra’s customers.
Role Overview
As an SDE 2 Back-end Engineer at Astra, you will play a crucial role in the development of a new vulnerability scanner from scratch. You will be architecting & engineering a scalable technical solution from the ground-up.
You will have the opportunity to work alongside talented individuals, collaborating to deliver innovative solutions and pushing the boundaries of what's possible in vulnerability scanning. The role requires deep collaboration with the founders, product, engineering & security teams.
Join our team and contribute to the development of a cutting-edge SaaS security platform, where high-quality engineering and continuous learning are at the core of everything we do.
Roles & Responsibilities:
- You will be joining our Vulnerability Scanner team which builds a security engine to identify vulnerabilities in technical infrastructure.
- You will be the technical product owner of the scanner, which would involve managing a lean team of backend engineers to ensure smooth implementation of the technical product roadmap.
- Research about security vulnerabilities, CVEs, and zero-days affecting cloud/web/API infrastructure.
- Work in an agile environment of engineers to architect, design, develop and build our microservice infrastructure.
- You will research, design, code, troubleshoot and support (on-call). What you create is also what you own.
- Writing secure, high quality, modular, testable & well documented code for features outlined in every sprint.
- Design and implement APIs in support of other services with a highly scalable, flexible, and secure backend using GoLang
- Hands-on experience with creating production-ready code & optimizing it by identifying and correcting bottlenecks.
- Driving strict code review standards among the team.
- Ensuring timely delivery of the features/products
- Working with product managers to ensure product delivery status is transparent & the end product always looks like how it was imagined
- Work closely with Security & Product teams in writing vulnerability detection rules, APIs etc.
Required Qualifications & Skills:
- Strong 2-4 years relevant development experience in GoLang
- Experience in building a technical product from idea to production.
- Design and build highly scalable and maintainable systems in Golang
- Expertise in Goroutines and Channels to write efficient code utilizing multi-core CPU optimally
- Must have hands-on experience with managing AWS/Google Cloud infrastructure
- Hands on experience in creating low latency high throughput REST APIs
- Write test suites and maintain code coverage above 80%
- Working knowledge of PostgreSQL, Redis, Kafka
- Good to have experience in Docker, Kubernetes, Kafka
- Good understanding of Data Structures, Algorithms and Operating Systems.
- Understanding of cloud/web security concepts would be an added advantage
What We Offer:
- Adrenalin rush of being a part of a fast-growing company
- Fully remote & agile working environment
- A wholesome opportunity in a fast-paced environment where you get to build things from scratch, improve and influence product design decisions
- Holistic understanding of SaaS and enterprise security business
- Opportunity to engage and collaborate with developers globally
- Experience with security side of things
- Annual trips to beaches or mountains (last one was Chikmangaluru)
- Open and supportive culture
Responsibilities:
Develop and maintain high-quality, scalable, and efficient Java codebase for our ad-serving platform.
Collaborate with cross-functional teams including product managers, designers, and other developers to
understand requirements and translate them into technical solutions.
Design and implement new features and functionalities in the ad-serving system, focusing on performance
optimization and reliability.
Troubleshoot and debug complex issues in the ad server environment, providing timely resolutions to ensure
uninterrupted service.
Conduct code reviews, provide constructive feedback, and enforce coding best practices to maintain code quality
and consistency across the platform.
Stay updated with emerging technologies and industry trends in ad serving and digital advertising, and integrate
relevant innovations into our platform.
Work closely with DevOps and infrastructure teams to deploy and maintain the ad-serving platform in a cloud- based environment.
Collaborate with stakeholders to gather requirements, define technical specifications, and estimate development
efforts for new projects and features.
Mentor junior developers, sharing knowledge and best practices to foster a culture of continuous learning and
improvement within the development team.
Participate in on-call rotations and provide support for production issues as needed, ensuring maximum uptime
and reliability of the ad-serving platform.


💥 What will you do?
As a Lead Backend Engineer, you will
- Build out and help scale our Django (+ReactJS / VueJS) based web application as we add new features and customers
-
Help us transition from monolithic system architecture to microservices and serverless system architecture to help meet our future scaling requirements
- Collaborate on a daily basis with a small, nimble team of product managers, engineers, and UX designers to understand business requirements and user experience goals and pain points.
- Build out features to enable multi-channel customer acquisition include partner channels, enterprise channels, and government channels
- Integrate third-party plugins to enhance customer experience and internal tooling build internal tooling to improve internal team efficiencies.
- Actively participate in code reviews
- Work with our quality assurance team to improve coverage on our automated testing suites
🙋 What are we looking for?
While we do not have a strict list of requirements for candidates interested in this role, some indicators that you would fit this role and our engineering culture are
-
Prior experience of a few years (3+ years) with python based frameworks as Django or Flask would be essential to work on our application stack.
-
You actively promote a culture of engineering excellence some of which could be writing efficient code, use of elegant design patterns, and styling your code through code-linting policies.
-
You enjoy designing software architecture by collaborating with engineering managers, architects, and other lead engineers to explore existing systems, determining areas of complexity, potential risks to successful implementation.
- You enjoy coaching folks to achieve outcomes through nudges.
📢 Other information you may want to consider
- We will be flexible for the rest of the pandemic and work remotely; however, we are not a remote-first company, and the work location would be Bangalore when things settle.
- Our backend tech stack includes Django, FastAPI, Postgres, Redis, Clickhouse, and TigerGraph. Our environments are managed through Docker, Kubernetes, and Terraform.
· Core responsibilities to include analyze business requirements and designs for accuracy and completeness. Develops and maintains relevant product.
· BlueYonder is seeking a Senior/Principal Architect in the Data Services department (under Luminate Platform ) to act as one of key technology leaders to build and manage BlueYonder’ s technology assets in the Data Platform and Services.
· This individual will act as a trusted technical advisor and strategic thought leader to the Data Services department. The successful candidate will have the opportunity to lead, participate, guide, and mentor other people in the team on architecture and design in a hands-on manner. You are responsible for technical direction of Data Platform. This position reports to the Global Head, Data Services and will be based in Bangalore, India.
· Core responsibilities to include Architecting and designing (along with counterparts and distinguished Architects) a ground up cloud native (we use Azure) SaaS product in Order management and micro-fulfillment
· The team currently comprises of 60+ global associates across US, India (COE) and UK and is expected to grow rapidly. The incumbent will need to have leadership qualities to also mentor junior and mid-level software associates in our team. This person will lead the Data platform architecture – Streaming, Bulk with Snowflake/Elastic Search/other tools
Our current technical environment:
· Software: Java, Springboot, Gradle, GIT, Hibernate, Rest API, OAuth , Snowflake
· • Application Architecture: Scalable, Resilient, event driven, secure multi-tenant Microservices architecture
· • Cloud Architecture: MS Azure (ARM templates, AKS, HD insight, Application gateway, Virtue Networks, Event Hub, Azure AD)
· Frameworks/Others: Kubernetes, Kafka, Elasticsearch, Spark, NOSQL, RDBMS, Springboot, Gradle GIT, Ignite

A Fintech start up company that bridges the gap between the digital payments
and the micro-retailing industry of the market. A nexus for the B2B sector providing multi-utility services via an Android app, enabling the micro-retailers across India to revolutionize their business and thus providing them with an opportunity to generate more revenue by offering multiple micropayment and digital services at their offline store. We intend to leverage the mode of smartphone technology in the rural and urban markets across India, thus empowering micro-retailers and creating a sustainable self-employment for this segment.
Overall Competencies:
1. Tech Depth/Breadth/Problem Solving.
2. Design & Architecture (Code/Design review).
3. Team Management.
Leadership Competency:
- Managing & mentoring a team.
- Ability to deliver projects/sprints on time.
- Obsessed with quality: Design reviews, code reviews, testing (system/performance).
- Spending at least 50% time with hands-on coding while doing the above Programming.
Tech Competency:
- Hands-on with Python/Django/Flask.
- Knowledge of PHP & Laravel is a plus.
- Experience with any JavaScript framework (Angular / React ).
- Prior experience in micro-services based architecture.
- Good understanding of SQL (MySQL, Postgres, or similar).
- Good understanding of one no-SQL (MongoDB / DynamoDB etc).
- Comfortable producing and refactoring code without assistance.
- Can reason about the complexity of algorithms (time and space).
- Basic CI/CD and DevOps.
- Experience in Docker & Kubernetes.
- Design, QA, and Infrastructure.
- Knowledge of design patterns and cloud scaling.
- AWS & GCP experience. AWS is mandatory.
- Understands BDD/TDD/Performance/Security/Smoke testing/Unit Testing
- Understands REST principles and Microservices
- Understands version control & branching strategies
- Understands application security fundamentals and potential security risks
Education: Bachelors in Computer Science|
Worked on Containerized microservices solution., node/express, docker/Kubernetes
Additional Skill:-, Experience with API Gateways: authentication and transformations (preferably with Kong),
Having worked on Event Broker Governance, Platform Communication and Integration Strategies, Event Management between Enterprise Systems
Roles – Lead ( upto 6 years of experience) and Architect ( 7+ years of experience) profiles
-
Bachelor’s or master’s degree in Computer Engineering, Computer Science, Computer Applications, Mathematics, Statistics, or related technical field. Relevant experience of at least 3 years in lieu of above if from a different stream of education.
-
Well-versed in and 3+ hands-on demonstrable experience with: ▪ Stream & Batch Big Data Pipeline Processing using Apache Spark and/or Apache Flink.
▪ Distributed Cloud Native Computing including Server less Functions
▪ Relational, Object Store, Document, Graph, etc. Database Design & Implementation
▪ Micro services Architecture, API Modeling, Design, & Programming -
3+ years of hands-on development experience in Apache Spark using Scala and/or Java.
-
Ability to write executable code for Services using Spark RDD, Spark SQL, Structured Streaming, Spark MLLib, etc. with deep technical understanding of Spark Processing Framework.
-
In-depth knowledge of standard programming languages such as Scala and/or Java.
-
3+ years of hands-on development experience in one or more libraries & frameworks such as Apache Kafka, Akka, Apache Storm, Apache Nifi, Zookeeper, Hadoop ecosystem (i.e., HDFS, YARN, MapReduce, Oozie & Hive), etc.; extra points if you can demonstrate your knowledge with working examples.
-
3+ years of hands-on development experience in one or more Relational and NoSQL datastores such as PostgreSQL, Cassandra, HBase, MongoDB, DynamoDB, Elastic Search, Neo4J, etc.
-
Practical knowledge of distributed systems involving partitioning, bucketing, CAP theorem, replication, horizontal scaling, etc.
-
Passion for distilling large volumes of data, analyze performance, scalability, and capacity performance issues in Big Data Platforms.
-
Ability to clearly distinguish system and Spark Job performances and perform spark performance tuning and resource optimization.
-
Perform benchmarking/stress tests and document the best practices for different applications.
-
Proactively work with tenants on improving the overall performance and ensure the system is resilient, and scalable.
-
Good understanding of Virtualization & Containerization; must demonstrate experience in technologies such as Kubernetes, Istio, Docker, OpenShift, Anthos, Oracle VirtualBox, Vagrant, etc.
-
Well-versed with demonstrable working experience with API Management, API Gateway, Service Mesh, Identity & Access Management, Data Protection & Encryption.
Hands-on experience with demonstrable working experience with DevOps tools and platforms viz., Jira, GIT, Jenkins, Code Quality & Security Plugins, Maven, Artifactory, Terraform, Ansible/Chef/Puppet, Spinnaker, etc.
-
Well-versed in AWS and/or Azure or and/or Google Cloud; must demonstrate experience in at least FIVE (5) services offered under AWS and/or Azure or and/or Google Cloud in any categories: Compute or Storage, Database, Networking & Content Delivery, Management & Governance, Analytics, Security, Identity, & Compliance (or) equivalent demonstrable Cloud Platform experience.
-
Good understanding of Storage, Networks and Storage Networking basics which will enable you to work in a Cloud environment.
-
Good understanding of Network, Data, and Application Security basics which will enable you to work in a Cloud as well as Business Applications / API services environment.
Containerization and CICD environment experience.
Strong analytical skills and problem solving aptitude attention to details


