
What You'll Do :
You will be a part of our backend team working on keeping our REST API and GraphQL API up and running and making sure that our users get the right data at the right time.
You will have ownership of developing and maintaining our backend services including users, courses, and operations tools that manage our product and logistics.
You will work on architecting and scaling highly-available RESTful services and back-end systems from scratch.
This is a position for an experienced Node programmer with at least 2 years under the belt, but you don't have to be a rock star, a ninja, or a superhero to apply.
What You'll Need :
You will fit well in the backend team if you are passionate about technology and have experience programming in Node. Knowledge of technologies like Cassandra, Elasticsearch, PostgreSQL, REST and JSON will help you get going from day one.
As an experienced Node developer, you should be intimately familiar with the platform, with
JavaScript, and with the stables of full-stack web development : HTTP, JavaScript, CSS, HTML, SQL.
It's a bonus if you're broadly familiar with other languages as well- we write some services in Go and have data pipelines written in Python- but your main work will be JavaScript through and through.
You will be a perfect match with our team if you love collaborating with people from all disciplines to solve complex problems, always want to learn new skills and take ownership of your work.

Similar jobs
We are looking for a skilled and motivated Data Engineer with strong experience in Python programming and Google Cloud Platform (GCP) to join our data engineering team. The ideal candidate will be responsible for designing, developing, and maintaining robust and scalable ETL (Extract, Transform, Load) data pipelines. The role involves working with various GCP services, implementing data ingestion and transformation logic, and ensuring data quality and consistency across systems.
Key Responsibilities:
- Design, develop, test, and maintain scalable ETL data pipelines using Python.
- Work extensively on Google Cloud Platform (GCP) services such as:
- Dataflow for real-time and batch data processing
- Cloud Functions for lightweight serverless compute
- BigQuery for data warehousing and analytics
- Cloud Composer for orchestration of data workflows (based on Apache Airflow)
- Google Cloud Storage (GCS) for managing data at scale
- IAM for access control and security
- Cloud Run for containerized applications
- Perform data ingestion from various sources and apply transformation and cleansing logic to ensure high-quality data delivery.
- Implement and enforce data quality checks, validation rules, and monitoring.
- Collaborate with data scientists, analysts, and other engineering teams to understand data needs and deliver efficient data solutions.
- Manage version control using GitHub and participate in CI/CD pipeline deployments for data projects.
- Write complex SQL queries for data extraction and validation from relational databases such as SQL Server, Oracle, or PostgreSQL.
- Document pipeline designs, data flow diagrams, and operational support procedures.
Required Skills:
- 4–8 years of hands-on experience in Python for backend or data engineering projects.
- Strong understanding and working experience with GCP cloud services (especially Dataflow, BigQuery, Cloud Functions, Cloud Composer, etc.).
- Solid understanding of data pipeline architecture, data integration, and transformation techniques.
- Experience in working with version control systems like GitHub and knowledge of CI/CD practices.
- Strong experience in SQL with at least one enterprise database (SQL Server, Oracle, PostgreSQL, etc.).
- Job Description:
You will be working with a team of highly-motivated, young & dynamic teams of engineers & reporting to the Engineering Lead.
Position Responsibilities
- Be honest, reliable & consistent
- Write efficient & clean code
- Have a strong sense of ownership
- Be a part of development & maintenance of Pluang web app, Operations dashboard and other 3rd party products we own
- Contribute to improving the quality of engineering process & engineering culture
Position Requirements
- Strong in data structure and algorithms
- Experience in Node.js, Express, API Design & DOM
- Understanding of component based design or other design patterns
- Experience with unit testing, integration testing & continuous integration
- RDBMS and NoSQL databases preferably PostgreSQL, MongoDB
- Good to have passion for investing
Pluang Offers
- Attractive compensation package - competitive salary, flexible bonus scheme.
- We are always looking for ways to promote and inspire innovation.
- Individual career path - management and technical career growth, enhanced by learning and development program, regular performance assessment, teams of multi-national IT professionals.
- Healthy work environment - company-sponsored medical program, food, and beverage program, open communication.
- Friendly policies to support Work-life balance, team building, and celebrations.
The Sr. Analytics Engineer would provide technical expertise in needs identification, data modeling, data movement, and transformation mapping (source to target), automation and testing strategies, translating business needs into technical solutions with adherence to established data guidelines and approaches from a business unit or project perspective.
Understands and leverages best-fit technologies (e.g., traditional star schema structures, cloud, Hadoop, NoSQL, etc.) and approaches to address business and environmental challenges.
Provides data understanding and coordinates data-related activities with other data management groups such as master data management, data governance, and metadata management.
Actively participates with other consultants in problem-solving and approach development.
Responsibilities :
Provide a consultative approach with business users, asking questions to understand the business need and deriving the data flow, conceptual, logical, and physical data models based on those needs.
Perform data analysis to validate data models and to confirm the ability to meet business needs.
Assist with and support setting the data architecture direction, ensuring data architecture deliverables are developed, ensuring compliance to standards and guidelines, implementing the data architecture, and supporting technical developers at a project or business unit level.
Coordinate and consult with the Data Architect, project manager, client business staff, client technical staff and project developers in data architecture best practices and anything else that is data related at the project or business unit levels.
Work closely with Business Analysts and Solution Architects to design the data model satisfying the business needs and adhering to Enterprise Architecture.
Coordinate with Data Architects, Program Managers and participate in recurring meetings.
Help and mentor team members to understand the data model and subject areas.
Ensure that the team adheres to best practices and guidelines.
Requirements :
- Strong working knowledge of at least 3 years of Spark, Java/Scala/Pyspark, Kafka, Git, Unix / Linux, and ETL pipeline designing.
- Experience with Spark optimization/tuning/resource allocations
- Excellent understanding of IN memory distributed computing frameworks like Spark and its parameter tuning, writing optimized workflow sequences.
- Experience of relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., Redshift, Bigquery, Cassandra, etc).
- Familiarity with Docker, Kubernetes, Azure Data Lake/Blob storage, AWS S3, Google Cloud storage, etc.
- Have a deep understanding of the various stacks and components of the Big Data ecosystem.
- Hands-on experience with Python is a huge plus
Designation: Python/Django Developer (Backend)
Department: Software Engineering Team
Location: Whitefield (Bangalore)
Qualification: Any Computer (Graduate or Post Graduate)
Experience: 3+ years of relevant experience in building functional and efficient server-side applications using Python/Django with 2+ years of relevant experience in working with RDBMS.
Responsibilities:
- Help design and implement functional requirements
- Build efficient back-end features in Python
- Integrate front-end components into applications
- Manage testing and bug fixes
- Prepare technical documentation
- Collaborate with UX/UI designers to implement design into the code
- Coach junior team members
- Implement software enhancements and suggest improvements
Skills Required:
- Experience with Python frameworks (e.g. Django, Flask, FastAPI)
- Familiarity with Amazon Web Services (AWS) cloud technologies
- Excellent understanding of REST API Development and HTTP Methods
- Understanding of databases and SQL
- Good experience in working with RDBMS
- Good understanding of computer science fundamentals - data structures, algorithms, memory management, performance.
- Knowledge of web stack - javascript, html, web frameworks like React.js is a plus
- Good understanding of Good Troubleshooting experience and exposure
- Ability to work with Jr engineers, mentor, guide and help them
- Enjoys coding and problem solving
- Good communication skills. Ability to express ideas clearly and concisely
- Familiarity with code versioning tools (such as Git, SVN, and Mercurial)
Salary: Upto 21LPA (Negotiable)
IMMEDIATE JOINING PREFERRED!
Who are we?
We are a venture capital-backed software development company headquartered in Canada. We develop in-house products to disrupt one industry at a time and partner as a technology service provider to selected startups.
Who are you?
Experience in writing applications using Nodejs including Express or similar.
Must be good in MySQL or one of the databases such as Mongo.
Proficient in Javascript and good experience and knowledge of open source tools, frameworks, broader cutting-edge technologies around server-side development.
Excellent data structure, algorithm, and problem-solving skills.
Created and consumed various APIs in the past.
Should be an active contributor to developer communities like Stack Overflow, GitHub, Google Developer Groups (GDGs).
Customer-focused, react well to changes, work with teams, and able to multi-task.
Must be a proven performer and team player that enjoys challenging assignments in a high-energy, fast-growing, and start-up workplace.
Must be a self-starter who can work well with minimal guidance and in a fluid environment.
Some of the technologies we use are:
NodeJS
ExpressJS
Angular 9
AWS
Github
PSQL, MongoDB
Selenium
ABOUT THE JOB
We are looking for a Senior Software Engineer to join our team. We believe in giving engineers responsibility, not tasks. Our goal is to motivate and challenge people to do their best work. To do that, we have a very fluid structure and give people flexibility to work on projects that they enjoy the most. This develops more capable engineers, and keeps everyone engaged and happy.
Responsibilities
- Design, develop, test, deploy, maintain and improve the software.
- Manage individual projects priorities, deadlines and deliverables with your technical expertise.
- Identify and solve for bottlenecks within our software stack.
ABOUT YOU
Rubrik Software Engineers are self-starters, driven, and can manage themselves. Bottom line, if you have a limitless drive and like to win, we want to talk to you - come make history!
- Bachelor’s or Master’s degree or equivalent in computer science or related field
- 8+ years of relevant work experience,
- Proficiency in one or more general purpose programming languages like Java, C/C++, Scala, Python
- Experience with Google Cloud Platform/AWS/Azure or other public cloud technologies is a plus
- Experience working with two or more from the following: Unix/Linux environments, Windows environments, distributed systems, networking, developing large software systems, file systems, storage systems, hypervisors, databases and/or security software development.
ABOUT THE TEAM
Galactus team owns the end to end development of Rubrik’s data management suite for commercial public clouds - AWS, Azure and GCP. Our objective is to bring the simplicity of Rubrik’s on-prem data protection and management offerings to our customers in the cloud through a solution designed from ground up to be highly scalable, available & secure and yet optimized to minimize our customer’s cloud costs. We achieve this by taking a cloud-first approach to design - leveraging technologies built for the scale, elasticity and automation needs of the cloud; and deploying on our brand new SaaS platform called Polaris.
Recently we have :-
- Built policy based backup and recovery for Virtual Machines in AWS, Azure and GCP and managed databases in AWS.
- Built features like granular file recovery leveraging managed Kubernetes Service in AWS for elastic compute
ABOUT RUBRIK
Rubrik is one of the fastest-growing companies in Silicon Valley, revolutionising data protection, and management in the emerging multi-cloud world. We are the leader in cloud data management, delivering a single platform to manage and protect data in the cloud, at the edge, and on-premises. Enterprises choose Rubrik to simplify backup and recovery, accelerate cloud adoption, enable automation at scale, and secure against cyberthreats. We’ve been recognized as a Forbes Cloud 100 Company two years in a row and as a LinkedIn Top 10 startup.
Rubrik provides equal employment opportunities (EEO) to all employees and applicants for employment without regard to race, color, religion, sex, national origin, age, disability or genetics. In addition to federal law requirements, Rubrik complies with applicable state and local laws governing nondiscrimination in employment in every location in which the company has facilities. This policy applies to all terms and conditions of employment, including recruiting, hiring, placement, promotion, termination, layoff, recall, transfer, leaves of absence, compensation and training.









