50+ AWS (Amazon Web Services) Jobs in Bangalore (Bengaluru) | AWS (Amazon Web Services) Job openings in Bangalore (Bengaluru)
Apply to 50+ AWS (Amazon Web Services) Jobs in Bangalore (Bengaluru) on CutShort.io. Explore the latest AWS (Amazon Web Services) Job opportunities across top companies like Google, Amazon & Adobe.
Key Responsibilities
- Design and implement Dremio lakehouse architecture on cloud (AWS/Azure/Snowflake/Databricks ecosystem).
- Define data ingestion, curation, and semantic modeling strategies to support analytics and AI workloads.
- Optimize Dremio reflections, caching, and query performance for diverse data consumption patterns.
- Collaborate with data engineering teams to integrate data sources via APIs, JDBC, Delta/Parquet, and object storage layers (S3/ADLS).
- Establish best practices for data security, lineage, and access control aligned with enterprise governance policies.
- Support self-service analytics by enabling governed data products and semantic layers.
- Develop reusable design patterns, documentation, and standards for Dremio deployment, monitoring, and scaling.
- Work closely with BI and data science teams to ensure fast, reliable, and well-modeled access to enterprise data.
Qualifications
- Bachelor’s or Master’s in Computer Science, Information Systems, or related field.
- 10+ years in data architecture and engineering, with 3+ years in Dremio or modern lakehouse platforms.
- Strong expertise in SQL optimization, data modeling, and performance tuning within Dremio or similar query engines (Presto, Trino, Athena).
- Hands-on experience with cloud storage (S3, ADLS, GCS), Parquet/Delta/Iceberg formats, and distributed query planning.
- Knowledge of data integration tools and pipelines (Airflow, DBT, Kafka, Spark, etc.).
- Familiarity with enterprise data governance, metadata management, and role-based access control (RBAC).
- Excellent problem-solving, documentation, and stakeholder communication skills.
Preferred:
- Experience integrating Dremio with BI tools (Tableau, Power BI, Looker) and data catalogs (Collibra, Alation, Purview).
- Exposure to Snowflake, Databricks, or BigQuery environments.
- Experience in high-tech, manufacturing, or enterprise data modernization programs.
Our Mission
To make video as accessible to machines as text and voice are today.
At lookup, we believe the world's most valuable asset is trapped. Video is everywhere, but it's unsearchable—a black box of insight that no one can open or atleast open affordably. We’re changing that. We're building the search engine for the visual world, so anyone can find or do anything with video just by asking.
Text is queryable. Voice is transcribed. Video, the largest and richest data source of all, is still a black box. A computer can't understand it, and so its value remains trapped.
Our mission at lookup is to fix this.
About the Role
We are looking for founding Backend Engineers to build a highly performant, reliable, and scalable API platform that makes enterprise video knowledge readily available for video search, summarization, and natural‑language Q&A. You will partner closely with our ML team working on vision‑language models to productionize research and deliver fast, trustworthy APIs for customers.
Examples of technical challenges you will work on include: distributed video storage, a unified application framework and data model for indexing large video libraries, low‑latency clip retrieval, vector search at scale, and end‑to‑end build, test, deploy, and observability in cloud environments.
What You’ll Do:
- Design and build robust backend services and APIs (REST, gRPC) for vector search, video summarization, and video Q&A.
- Own API performance and reliability, including low‑latency retrieval, pagination, rate limiting, and backwards‑compatible versioning.
- Design schemas and tune queries in Postgres, and integrate with unstructured storage.
- Implement observability across metrics, logs, and traces. Set error budgets and SLOs.
- Write clear design docs and ship high‑quality, well‑tested code.
- Collaborate with ML engineers to integrate and productionize VLMs and retrieval pipelines.
- Take ownership of architecture from inception to production launch.
Who You Are:
- 3+ years of professional experience in backend development.
- Proven experience building and scaling polished WebSocket, gRPC, and REST APIs.
- Exposure to distributed systems and container orchestration (Docker and Kubernetes).
- Hands‑on experience with AWS.
- Strong knowledge of SQL (Postgres) and NoSQL (e.g., Cassandra), including schema design, query optimization, and scaling.
- Familiarity with our stack is a plus, but not mandatory: Python (FastAPI), Celery, Kafka, Postgres, Redis, Weaviate, React.
- Ability to diagnose complex issues, identify root causes, and implement effective fixes.
- Comfortable working in a fast‑paced startup environment.
Nice to have:
- Hands-on work with LLM agents, vector embeddings, or RAG applications.
- Building video streaming pipelines and storage systems at scale (FFmpeg, RTSP, WebRTC).
- Proficiency with modern frontend frameworks (React, TypeScript, Tailwind CSS) and responsive UI design.
Location & Culture
- Full-time, in-office role in Bangalore (we’re building fast and hands-on).
- Must be comfortable with a high-paced environment and collaboration across PST time zones for our US customers and investors.
- Expect startup speed — daily founder syncs, rapid design-to-prototype cycles, and a culture of deep ownership.
Why You’ll Love This Role
- Work on the frontier of video understanding and real-world AI — products that can redefine trust and automation.
- Build core APIs that make video queryable and power real customer use.
- Own systems end to end: performance, reliability, and developer experience.
- Work closely with founders and collaborate in person in Bangalore.
- Competitive salary with meaningful early equity.
You will be responsible for building a highly-scalable and extensible robust application. This position reports to the Engineering Manager.
Responsibilities:
- Align Sigmoid with key Client initiatives
- Interface daily with customers across leading Fortune 500 companies to understand strategic requirements
- Ability to understand business requirements and tie them to technology solutions
- Open to work from client location as per the demand of the project / customer.
- Facilitate in Technical Aspects
- Develop and evolve highly scalable and fault-tolerant distributed components using Java technologies.
- Excellent experience in Application development and support, integration development and quality assurance.
- Provide technical leadership and manage it day to day basis
- Interface daily with customers across leading Fortune 500 companies to understand strategic requirements
- Stay up-to-date on the latest technology to ensure the greatest ROI for customer & Sigmoid
- Hands on coder with good understanding on enterprise level code.
- Design and implement APIs, abstractions and integration patterns to solve challenging distributed computing problems
- Experience in defining technical requirements, data extraction, data transformation, automating jobs, productionizing jobs, and exploring new big data technologies within a Parallel Processing environment
- Culture
- Must be a strategic thinker with the ability to think unconventional / out:of:box.
- Analytical and solution driven orientation.
- Raw intellect, talent and energy are critical.
- Entrepreneurial and Agile : understands the demands of a private, high growth company.
- Ability to be both a leader and hands on "doer".
Qualifications: -
- 3-5 year track record of relevant work experience and a computer Science or a related technical discipline is required
- Experience in development of Enterprise scale applications and capable in developing framework, design patterns etc. Should be able to understand and tackle technical challenges, and propose comprehensive solutions.
- Experience with functional and object-oriented programming, Java (Preferred) or Python is a must.
- Hand-On knowledge in Map Reduce, Hadoop, PySpark, Hbase and ElasticSearch.
- Development and support experience in Big Data domain
- Experience with database modelling and development, data mining and warehousing.
- Unit, Integration and User Acceptance Testing.
- Effective communication skills (both written and verbal)
- Ability to collaborate with a diverse set of engineers, data scientists and product managers
- Comfort in a fast-paced start-up environment.
Preferred Qualification:
- Experience in Agile methodology.
- Proficient with SQL and its variation among popular databases.
- Experience working with large, complex data sets from a variety of sources.
Role: SDE 3- Lead Software Engineer (Backend)
Salary: INR 30 to INR 40L per annum
Performance Bonus: Up to 10% of the base salary can be added
Location: Hulimavu, Bangalore, India
Experience: 6-10 years
About AbleCredit:
AbleCredit has built a foundational AI platform to help BFSI enterprises reduce OPEX by up to 70% by powering workflows for onboarding, claims, credit, and collections. Our GenAI model achieves over 95% accuracy in understanding Indian dialects and excels in financial analysis.
The company was founded in June 2023 by Utkarsh Apoorva (IIT Delhi, built Reshamandi, Guitarstreet, Edulabs); Harshad Saykhedkar (IITB, ex-AI Lead at Slack); and Ashwini Prabhu (IIML, co-founder of Mythiksha, ex-Product Head at Reshamandi, HandyTrain).
What Work You’ll Do
- Build best-in-class AI systems - that enterprises can trust, where reliability and explainability are not optional.
- Operate in founder mode — build, patch, or fork, whatever it takes to ship today, not next week.
- Work at the frontier of AI x Systems — making AI models behave predictably to solve real, enterprise-grade problems.
- Own end-to-end feature delivery — from requirement scoping to design, development, testing, deployment, and post-release optimization.
- Design and implement complex, distributed systems that support large-scale workflows and integrations for enterprise clients.
- Operate with full technical ownership — make architectural decisions, review code, and mentor junior engineers to maintain quality and velocity.
- Build scalable, event-driven services leveraging AWS Lambda, SQS/SNS, and modern asynchronous patterns.
- Work with cross-functional teams to design robust notification systems, third-party integrations, and data pipelines that meet enterprise reliability and security standards.
The Skills You Have..
- Strong background as an Individual Contributor — capable of owning systems from concept to production without heavy oversight.
- Expertise in system design, scalability, and fault-tolerant architecture.
- Proficiency in Node.js (bonus) or another backend language such as Go, Java, or Python.
- Deep understanding of SQL (PostgreSQL/MySQL) and NoSQL (MongoDB/DynamoDB) systems.
- Hands-on experience with AWS services — Lambda, API Gateway, S3, CloudWatch, ECS/EKS, and event-based systems.
- Experience in designing and scaling notification systems and third-party API integrations.
- Proficiency in event-driven architectures and multi-threading/concurrency models.
- Strong understanding of data modeling, security practices, and performance optimization.
- Familiarity with CI/CD pipelines, automated testing, and monitoring tools.
- Strong debugging, performance tuning, and code review skills.
What You Should Have Done in the Past
- Delivered multiple complex backend systems or microservices from scratch in a production environment.
- Led system design discussions and guided teams on performance, reliability, and scalability trade-offs.
- Mentored SDE-1 and SDE-2 engineers, enabling them to deliver features independently.
- Owned incident response and root cause analysis for production systems.
- (Bonus) Built or contributed to serverless systems using AWS Lambda, with clear metrics on uptime, throughput, and cost-efficiency.
Highlights:
- PTO & Holidays
- Opportunity to work with a core Gen AI startup.
- Flexible hours and an extremely positive work environment
Domain - Credit risk / Fintech
Roles and Responsibilities:
1. Development, validation and monitoring of Application and Behaviour score cards
for Retail loan portfolio
2. Improvement of collection efficiency through advanced analytics
3. Development and deployment of fraud scorecard
4. Upsell / Cross-sell strategy implementation using analytics
5. Create modern data pipelines and processing using AWS PAAS components (Glue,
Sagemaker studio, etc.)
6. Deploying software using CI/CD tools such as Azure DevOps, Jenkins, etc.
7. Experience with API tools such as REST, Swagger, and Postman
8. Model deployment in AWS and management of production environment
9. Team player who can work with cross-functional teams to gather data and derive
insights
Mandatory Technical skill set :
1. Previous experience in scorecard development and credit risk strategy development
2. Python and Jenkins
3. Logistic regression, Scorecard, ML and neural networks
4. Statistical analysis and A/B testing
5. AWS Sagemaker, S3 , Ec2, Dockers
6. REST API, Swagger and Postman
7. Excel
8. SQL
9. Visualisation tools such as Redash / Grafana
10. Bitbucket, Githubs and versioning tools

Global digital transformation solutions provider.
Job Description
Java or Python, Angular Front end, basic AWS (current project)
Nice to Haves
Experience with other modern JavaScript frameworks like Vue.js.
Exposure to containerization and orchestration technologies such as Docker and Kubernetes.
Knowledge of CI/CD best practices and tools (e.g., Jenkins, GitLab CI, CircleCI).
Familiarity with cloud-native architecture and microservices design.
Strong problem-solving skills and a collaborative mindset.
Experience working in fast-paced, dynamic startup or product-driven environments.
Must-Haves
Minimum of 5+ years of professional experience in the below skills:
Java or Python, Angular Front end, basic AWS (this skills should be visible in his current project)
Notice period - 0 to 15days only
Work locations: Bangalore, Chennai, Gurgaon, Hyderabad, Kochi, Noida, Thiruvananthapuram , India
(Virtual interview mode)
Interview Date: 15th Nov
Additional Guidelines
Interview process: - 2 Technical round + 1 Client round
3 days in office, Hybrid model.
Senior Python Django Developer
Experience: Back-end development: 6 years (Required)
Location: Bangalore/ Bhopal
Job Description:
We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the startup environment.
This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.
Responsibilities:
- Design and develop scalable, secure, and high-performance applications using Python (Django framework).
- Architect system components, define database schemas, and optimize backend services for speed and efficiency.
- Lead and implement design patterns and software architecture best practices.
- Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
- Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
- Drive performance improvements, monitor system health, and troubleshoot production issues.
- Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
- Contribute to technical decision-making and mentor junior developers.
Requirements:
- 6 to 10 years of professional backend development experience with Python and Django.
- Strong background in payments/financial systems or FinTech applications.
- Proven experience in designing software architecture in a microservices or modular monolith environment.
- Experience working in fast-paced startup environments with agile practices.
- Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
- Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
- Hands-on experience with test-driven development (TDD) and frameworks like pytest, unittest, or factory_boy.
- Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).
Preferred Skills:
- Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
- Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
- Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
- Contributions to open-source or personal finance-related projects.
Job Types: Full-time, Permanent
Schedule:
- Day shift
Supplemental Pay:
- Performance bonus
- Yearly bonus
Ability to commute/relocate:
- JP Nagar, 5th Phase, Bangalore, Karnataka or Indrapuri, Bhopal, Madhya Pradesh: Reliably commute or willing to relocate with an employer-provided relocation package (Preferred)
Read less
Role Overview
An opportunity to work with the BI WORLDWIDE Technology team as a .Net Full Stack Developer.
Responsibilities
- Ability to convert requirements to design and deliver code
- Be able to take complete ownership of the delivery of a single/multiple module on time and schedule
- Quick learner to be able to pick new technologies to navigate through changes that happen time to time
Key Skills – Must Have
- Proven experience in .NET Development (.Net Core or .Net Framework Version above 4.0)
- Strong hands-on experience in Angular / AngularJS (14.0; above 8 plus can be considered)
- Should be expertise in Entity Framework (CF and EF Core)
- Strong Hands-on C# and OOPS Concepts and Design Pattern
- Expertise in MVC and WebAPI
- Hands-on Frontend technologies
- Familiarity with architecture styles/APIs (WebAPI, REST, RPC)
- Understanding of Agile methodologies
- SQL Server 2017
Key Skills – Good to Have
- Postgres SQL
- AWS
- GitLab
- Azure
- Redis
- LINQ Framework
Educational Qualification: BE/BTech, MTech / MSc
About BI Worldwide
Introduction
BI WORLDWIDE’s mission is to put our engagement expertise to work for you. We understand how to engage employees, channel partners and customers better than anyone else and we have the tools and resources to help your business tap into that knowledge to drive and sustain measurable results.
Our origins are both humble and ingenious. (We like to think that describes us pretty well). It started with founder Guy Scheidecker discovering innovative ways to provide consumer loyalty programs and incentives to various industries. That idea caught on, big time. It grew into what is now a multinational company serving Global 2000 corporations in over 125 countries in more than 20 languages.
Our U.S. headquarters is located in Minneapolis, Minnesota. We have more than 25 sales offices throughout the U.S. and headquarters around the world in Australia, Canada, China, India, Latin America, Singapore and the United Kingdom.
BI WORLDWIDE INDIA
In India, our headquarters is located in Bengaluru, Karnataka and other offices in Chennai, Delhi and Mumbai. In Chennai, we have a fully equipped infrastructure to develop and deploy advanced technology solutions for your engagement, rewards and loyalty programs.
With over 220+ associates and 40+ clients including marquee names such as Genpact, Wipro, Dr Reddy’s, Hewlett Packard, IBM, Michelin, Tata Motors, Amdocs etc, BIW India has established itself as a premium service provider and achieved a No.1 ranking in the employee recognition marketplace within a span of 8 years.
What makes us different
How did we get to be the world’s leading expert on engagement? Experience is part of it but it also stems from our unique approach. We design and deliver innovative solutions grounded in the science of behavioral economics.
In other words, we look at how thought and emotion combine to drive decisions and behavior. We don’t have to guess what might motivate teams or customers; we actually use research and data to figure out what will motivate them. Then we act on it.
What you’ll love about us?
For starters, we take these things to heart: Work hard, play hard, celebrate. Make work meaningful. Magnify successes. True to our EVP of Grow & have fun in a culture of innovation and trust, some initiatives we offer all our Associates.
For further information please visit our website www.biworldwide.com
Junior DevOps Engineer
Experience: 2–3 years
About Us
We are a fast-growing fintech/trading company focused on building scalable, high-performance systems for financial markets. Our technology stack powers real-time trading, risk management, and analytics platforms. We are looking for a motivated Junior DevOps Engineer to join our dynamic team and help us maintain and improve our infrastructure.
Key Responsibilities
- Support deployment, monitoring, and maintenance of trading and fintech applications.
- Automate infrastructure provisioning and deployment pipelines using tools like Ansible, Terraform, or similar.
- Collaborate with development and operations teams to ensure high availability, reliability, and security of systems.
- Troubleshoot and resolve production issues in a fast-paced environment.
- Implement and maintain CI/CD pipelines for continuous integration and delivery.
- Monitor system performance and optimize infrastructure for scalability and cost-efficiency.
- Assist in maintaining compliance with financial industry standards and security best practices.
Required Skills
- 2–3 years of hands-on experience in DevOps or related roles.
- Proficiency in Linux/Unix environments.
- Experience with containerization (Docker) and orchestration (Kubernetes).
- Familiarity with cloud platforms (AWS, GCP, or Azure).
- Working knowledge of scripting languages (Bash, Python).
- Experience with configuration management tools (Ansible, Puppet, Chef).
- Understanding of networking concepts and security practices.
- Exposure to monitoring tools (Prometheus, Grafana, ELK stack).
- Basic understanding of CI/CD tools (Jenkins, GitLab CI, GitHub Actions).
Preferred Skills
- Experience in fintech, trading, or financial services.
- Knowledge of high-frequency trading systems or low-latency environments.
- Familiarity with financial data protocols and APIs.
- Understanding of regulatory requirements in financial technology.
What We Offer
- Opportunity to work on cutting-edge fintech/trading platforms.
- Collaborative and learning-focused environment.
- Competitive salary and benefits.
- Career growth in a rapidly expanding domain.
Job Summary :
We are looking for a proactive and skilled Senior DevOps Engineer to join our team and play a key role in building, managing, and scaling infrastructure for high-performance systems. The ideal candidate will have hands-on experience with Kubernetes, Docker, Python scripting, cloud platforms, and DevOps practices around CI/CD, monitoring, and incident response.
Key Responsibilities :
- Design, build, and maintain scalable, reliable, and secure infrastructure on cloud platforms (AWS, GCP, or Azure).
- Implement Infrastructure as Code (IaC) using tools like Terraform, Cloud Formation, or similar.
- Manage Kubernetes clusters, configure namespaces, services, deployments, and auto scaling.
CI/CD & Release Management :
- Build and optimize CI/CD pipelines for automated testing, building, and deployment of services.
- Collaborate with developers to ensure smooth and frequent deployments to production.
- Manage versioning and rollback strategies for critical deployments.
Containerization & Orchestration using Kubernetes :
- Containerize applications using Docker, and manage them using Kubernetes.
- Write automation scripts using Python or Shell for infrastructure tasks, monitoring, and deployment flows.
- Develop utilities and tools to enhance operational efficiency and reliability.
Monitoring & Incident Management :
- Analyze system performance and implement infrastructure scaling strategies based on load and usage trends.
- Optimize application and system performance through proactive monitoring and configuration tuning.
Desired Skills and Experience :
- Experience Required - 8+ yrs.
- Hands-on experience on cloud services like AWS, EKS etc.
- Ability to design a good cloud solution.
- Strong Linux troubleshooting, Shell Scripting, Kubernetes, Docker, Ansible, Jenkins Skills.
- Design and implement the CI/CD pipeline following the best industry practices using open-source tools.
- Use knowledge and research to constantly modernize our applications and infrastructure stacks.
- Be a team player and strong problem-solver to work with a diverse team.
- Having good communication skills.
Job Description.
1. Cloud experience (Any cloud is fine although AWS is preferred. If non-AWS cloud, then the experience should reflect familiarity with the cloud's common services)
2. Good grasp of Scripting (in Linux for sure ie bash/sh/zsh etc, Windows : nice to have)
3. Python or Java or JS basic knowledge (Python Preferred)
4. Monitoring tools
5. Alerting tools
6. Logging tools
7. CICD
8. Docker/containers/(k8s/terraform nice to have)
9. Experience working on distributed applications with multiple services
10. Incident management
11. DB experience in terms of basic queries
12. Understanding of performance analysis of applications
13. Idea about data pipelines would be nice to have
14. Snowflake querying knowledge: nice to have
The person should be able to :
Monitor system issues
Create strategies to detect and address issues
Implement automated systems to troubleshoot and resolve issues.
Write and review post-mortems
Manage infrastructure for multiple product teams
Collaborate with product engineering teams to ensure best practices are being followed
About Borderless Access
Borderless Access is a company that believes in fostering a culture of innovation and collaboration to build and deliver digital-first products for market research methodologies. This enables our customers to stay ahead of their competition.
We are committed to becoming the global leader in providing innovative digital offerings for consumers backed by advanced analytics, AI, ML, and cutting-edge technological capabilities.
Our Borderless Product Innovation and Operations team is dedicated to creating a top-tier market research platform that will drive our organization's growth. To achieve this, we're embracing modern technologies and a cutting-edge tech stack for faster, higher-quality product development.
The Product Development team is the core of our strategy, fostering collaboration and efficiency. If you're passionate about innovation and eager to contribute to our rapidly evolving market research domain, we invite you to join our team.
Key Responsibilities
- Lead, mentor, and grow a cross-functional team of engineers specializing.
- Foster a culture of collaboration, accountability, and continuous learning.
- Oversee the design and development of robust platform architecture with a focus on scalability, security, and maintainability.
- Establish and enforce engineering best practices including code reviews, unit testing, and CI/CD pipelines.
- Promote clean, maintainable, and well-documented code across the team.
- Lead architectural discussions and technical decision-making, with clear and concise documentation for software components and systems.
- Collaborate with Product, Design, and other stakeholders to define and prioritize platform features.
- Track and report on key performance indicators (KPIs) such as velocity, code quality, deployment frequency, and incident response times.
- Ensure timely delivery of high-quality software aligned with business goals.
- Work closely with DevOps to ensure platform reliability, scalability, and observability.
- Conduct regular 1:1s, performance reviews, and career development planning.
- Conduct code reviews and provide constructive feedback to ensure code quality and maintainability.
- Participate in the entire software development lifecycle, from requirements gathering to deployment and maintenance.
Added Responsibilities
- Defining and adhering to the development process.
- Taking part in regular external audits and maintaining artifacts.
- Identify opportunities for automation to reduce repetitive tasks.
- Mentor and coach team members in the teams.
- Continuously optimize application performance and scalability.
- Collaborate with the Marketing team to understand different user journeys.
Growth and Development
The following are some of the growth and development activities that you can look forward to at Borderless Access as an Engineering Manager:
- Develop leadership skills – Enhance your leadership abilities through workshops or coaching from Senior Leadership and Executive Leadership.
- Foster innovation – Become part of a culture of innovation and experimentation within the product development and operations team.
- Drive business objectives – Become part of defining and taking actions to meet the business objectives.
About You
- Bachelor's degree in Computer Science, Engineering, or a related field.
- 8+ years of experience in software development.
- Experience with microservices architecture and container orchestration.
- Excellent problem-solving and analytical skills.
- Strong communication and collaboration skills.
- Solid understanding of data structures, algorithms, and software design patterns.
- Solid understanding of enterprise system architecture patterns.
- Experience in managing a small to medium-sized team with varied experiences.
- Strong proficiency in back-end development, including programming languages like Python, Java, or Node.js, and frameworks like Spring or Express.
- Strong proficiency in front-end development, including HTML, CSS, JavaScript, and popular frameworks like React or Angular.
- Experience with databases (e.g., MySQL, PostgreSQL, MongoDB).
- Experience with cloud platforms AWS, Azure, or GCP (preferred is Azure).
- Knowledge of containerization technologies Docker and Kubernetes.
Python Backend Developer
We are seeking a skilled Python Backend Developer responsible for managing the interchange of data between the server and the users. Your primary focus will be on developing server-side logic to ensure high performance and responsiveness to requests from the front end. You will also be responsible for integrating front-end elements built by your coworkers into the application, as well as managing AWS resources.
Roles & Responsibilities
- Develop and maintain scalable, secure, and robust backend services using Python
- Design and implement RESTful APIs and/or GraphQL endpoints
- Integrate user-facing elements developed by front-end developers with server-side logic
- Write reusable, testable, and efficient code
- Optimize components for maximum performance and scalability
- Collaborate with front-end developers, DevOps engineers, and other team members
- Troubleshoot and debug applications
- Implement data storage solutions (e.g., PostgreSQL, MySQL, MongoDB)
- Ensure security and data protection
Mandatory Technical Skill Set
- Implementing optimal data storage (e.g., PostgreSQL, MySQL, MongoDB, S3)
- Python backend development experience
- Design, implement, and maintain CI/CD pipelines using tools such as Jenkins, GitLab CI/CD, or GitHub Actions
- Implemented and managed containerization platforms such as Docker and orchestration tools like Kubernetes
- Previous hands-on experience in:
- EC2, S3, ECS, EMR, VPC, Subnets, SQS, CloudWatch, CloudTrail, Lambda, SageMaker, RDS, SES, SNS, IAM, S3, Backup, AWS WAF
- SQL
Job Title: Senior AI Engineer
Location: Bengaluru, India – (Hybrid)
About Reltio
At Reltio®, we believe data should fuel business success. Reltio’s AI-powered data unification and management capabilities—encompassing entity resolution, multi-domain master data management (MDM), and data products—transform siloed data from disparate sources into unified, trusted, and interoperable data.
Reltio Data Cloud™ delivers interoperable data where and when it's needed, empowering data and analytics leaders with unparalleled business responsiveness. Leading enterprise brands across multiple industries around the globe rely on our award-winning data unification and cloud-native MDM capabilities to improve efficiency, manage risk, and drive growth.
At Reltio, our values guide everything we do. With an unyielding commitment to prioritizing our “Customer First”, we strive to ensure their success. We embrace our differences and are “Better Together” as One Reltio. We “Simplify and Share” our knowledge to remove obstacles for each other. We “Own It”, holding ourselves accountable for our actions and outcomes. Every day, we innovate and evolve so that today is “Always Better Than Yesterday.”
If you share and embody these values, we invite you to join our team at Reltio and contribute to our mission of excellence.
Reltio has earned numerous awards and top rankings for our technology, our culture, and our people. Founded on a distributed workforce, Reltio offers flexible work arrangements to help our people manage their personal and professional lives. If you’re ready to work on unrivaled technology as part of a collaborative team on a mission to enable digital transformation with connected data, let’s talk!
Job Summary
As a Senior AI Engineer at Reltio, you will be a core part of the team responsible for building intelligent systems that enhance data quality, automate decision-making, and drive entity resolution at scale.
You will work with cross-functional teams to design and deploy advanced AI/ML solutions that are production-ready, scalable, and embedded into our flagship data platform.
This is a high-impact engineering role with exposure to cutting-edge problems in entity resolution, deduplication, identity stitching, record linking, and metadata enrichment.
Job Duties and Responsibilities
- Design, implement, and optimize state-of-the-art AI/ML models for solving real-world data management challenges such as entity resolution, classification, similarity matching, and anomaly detection.
- Work with structured, semi-structured, and unstructured data to extract signals and engineer intelligent features for large-scale ML pipelines.
- Develop scalable ML workflows using Spark, MLlib, PyTorch, TensorFlow, or MLFlow, with seamless integration into production systems.
- Translate business needs into technical design and collaborate with data scientists, product managers, and platform engineers to operationalize models.
- Continuously monitor and improve model performance using feedback loops, A/B testing, drift detection, and retraining strategies.
- Conduct deep dives into customer data challenges and apply innovative machine learning algorithms to address accuracy, speed, and bias.
- Actively contribute to research and experimentation efforts, staying updated with the latest AI trends in graph learning, NLP, probabilistic modeling, etc.
- Document designs and present outcomes to both technical and non-technical stakeholders, fostering transparency and knowledge sharing.
Skills You Must Have
- Bachelor’s or Master’s degree in Computer Science, Machine Learning, Artificial Intelligence, or related field. PhD is a plus.
- 4+ years of hands-on experience in developing and deploying machine learning models in production environments.
- Proficiency in Python (NumPy, scikit-learn, pandas, PyTorch/TensorFlow) and experience with large-scale data processing tools (Spark, Kafka, Airflow).
- Strong understanding of ML fundamentals, including classification, clustering, feature selection, hyperparameter tuning, and evaluation metrics.
- Demonstrated experience working with entity resolution, identity graphs, or data deduplication.
- Familiarity with containerized environments (Docker, Kubernetes) and cloud platforms (AWS, GCP, Azure).
- Strong debugging, analytical, and communication skills with a focus on delivery and impact.
- Attention to detail, ability to work independently, and a passion for staying updated with the latest advancements in the field of data science.
Skills Good to Have
- Experience with knowledge graphs, graph-based ML, or embedding techniques.
- Exposure to deep learning applications in data quality, record matching, or information retrieval.
- Experience building explainable AI solutions in regulated domains.
- Prior work in SaaS, B2B enterprise platforms, or data infrastructure companies.
Company Description
Appiness Interactive Pvt. Ltd. is a Bangalore-based product development and UX firm that specializes in digital services for startups to fortune-500s. We work closely with our clients to
create a comprehensive soul for their brand in the online world, engaged through multiple
platforms of digital media. Our team is young, passionate, and aggressive, not afraid to think out
of the box or tread the un-trodden path in order to deliver the best results for our clients. We
pride ourselves on Practical Creativity where the idea is only as good as the returns it fetches for
our clients.
Role Overview
We are hiring a Founding Backend Engineer to architect and build the core backend
infrastructure for our enterprise AI chat platform. This role involves creating everything from
secure chat APIs and data pipelines to document embeddings, vector search, and RAG
(Retrieval-Augmented Generation) workflows. You will work directly with the CTO and play a
pivotal role in shaping the platform’s architecture, performance, and scalability as we onboard
enterprise customers. This is a high-ownership role where you’ll influence product direction, tech
decisions, and long-term engineering culture.
Key Responsibilities
● Architect, develop, and scale backend systems and APIs powering AI chat and knowledge
retrieval.
● Build data ingestion & processing pipelines for structured and unstructured enterprise
data.
● Implement multi-tenant security, user access control (RBAC), encryption, and
compliance-friendly design.
● Integrate and orchestrate LLMs (OpenAI, Anthropic, etc.) with vector databases
(Pinecone, Qdrant, OpenSearch) to support advanced AI and RAG workflows.
● Ensure platform reliability, performance, and fault tolerance from day one.
● Own end-to-end CI/CD, observability, and deployment pipelines.
● Collaborate directly with leadership on product strategy, architecture, and scaling
roadmap.
Required Skills
● Strong hands-on experience in Python (Django/FastAPI) or Node.js (TypeScript) — Python
preferred.
● Deep understanding of PostgreSQL, Redis, Docker, and modern API design patterns.
● Experience with LLM integration, RAG pipelines, and vector search technologies.
● Strong exposure to cloud platforms (AWS or GCP), CI/CD, and microservice architecture.
● Solid foundation in security best practices — authentication, RBAC, encryption, data
isolation.
● Ability to independently design and deliver high-performance distributed systems.
Designation: Senior Python Django Developer
Position: Senior Python Developer
Job Types: Full-time, Permanent
Pay: Up to ₹800,000.00 per year
Schedule: Day shift
Ability to commute/relocate: Bhopal Indrapuri (MP) And Bangalore JP Nagar
Experience: Back-end development: 4 years (Required)
Job Description:
We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the startup environment.
This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.
Responsibilities:
- Design and develop scalable, secure, and high-performance applications using Python (Django framework).
- Architect system components, define database schemas, and optimize backend services for speed and efficiency.
- Lead and implement design patterns and software architecture best practices.
- Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
- Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
- Drive performance improvements, monitor system health, and troubleshoot production issues.
- Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
- Contribute to technical decision-making and mentor junior developers.
Requirements:
- 4 to 10 years of professional backend development experience with Python and Django.
- Strong background in payments/financial systems or FinTech applications.
- Proven experience in designing software architecture in a microservices or modular monolith environment.
- Experience working in fast-paced startup environments with agile practices.
- Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
- Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
- Hands-on experience with test-driven development (TDD) and frameworks like pytest, unit test, or factory boy.
- Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).
Preferred Skills:
- Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
- Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
- Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
- Contributions to open-source or personal finance-related projects.
Job Details
- Job Title: Lead II - Software Engineering- AI, NLP, Python, Data science
- Industry: Technology
- Domain - Information technology (IT)
- Experience Required: 7-9 years
- Employment Type: Full Time
- Job Location: Bangalore
- CTC Range: Best in Industry
Job Description:
Role Proficiency:
Act creatively to develop applications by selecting appropriate technical options optimizing application development maintenance and performance by employing design patterns and reusing proven solutions. Account for others' developmental activities; assisting Project Manager in day-to-day project execution.
Additional Comments:
Mandatory Skills Data Science Skill to Evaluate AI, Gen AI, RAG, Data Science
Experience 8 to 10 Years
Location Bengaluru
Job Description
Job Title AI Engineer Mandatory Skills Artificial Intelligence, Natural Language Processing, python, data science Position AI Engineer – LLM & RAG Specialization Company Name: Sony India Software Centre About the role: We are seeking a highly skilled AI Engineer with 8-10 years of experience to join our innovation-driven team. This role focuses on the design, development, and deployment of advanced enterprise-scale Large Language Models (eLLM) and Retrieval Augmented Generation (RAG) solutions. You will work on end-to-end AI pipelines, from data processing to cloud deployment, delivering impactful solutions that enhance Sony’s products and services. Key Responsibilities: Design, implement, and optimize LLM-powered applications, ensuring high performance and scalability for enterprise use cases. Develop and maintain RAG pipelines, including vector database integration (e.g., Pinecone, Weaviate, FAISS) and embedding model optimization. Deploy, monitor, and maintain AI/ML models in production, ensuring reliability, security, and compliance. Collaborate with product, research, and engineering teams to integrate AI solutions into existing applications and workflows. Research and evaluate the latest LLM and AI advancements, recommending tools and architectures for continuous improvement. Preprocess, clean, and engineer features from large datasets to improve model accuracy and efficiency. Conduct code reviews and enforce AI/ML engineering best practices. Document architecture, pipelines, and results; present findings to both technical and business stakeholders. Job Description: 8-10 years of professional experience in AI/ML engineering, with at least 4+ years in LLM development and deployment. Proven expertise in RAG architectures, vector databases, and embedding models. Strong proficiency in Python; familiarity with Java, R, or other relevant languages is a plus. Experience with AI/ML frameworks (PyTorch, TensorFlow, etc.) and relevant deployment tools. Hands-on experience with cloud-based AI platforms such as AWS SageMaker, AWS Q Business, AWS Bedrock or Azure Machine Learning. Experience in designing, developing, and deploying Agentic AI systems, with a focus on creating autonomous agents that can reason, plan, and execute tasks to achieve specific goals. Understanding of security concepts in AI systems, including vulnerabilities and mitigation strategies. Solid knowledge of data processing, feature engineering, and working with large-scale datasets. Experience in designing and implementing AI-native applications and agentic workflows using the Model Context Protocol (MCP) is nice to have. Strong problem-solving skills, analytical thinking, and attention to detail. Excellent communication skills with the ability to explain complex AI concepts to diverse audiences. Day-to-day responsibilities: Design and deploy AI-driven solutions to address specific security challenges, such as threat detection, vulnerability prioritization, and security automation. Optimize LLM-based models for various security use cases, including chatbot development for security awareness or automated incident response. Implement and manage RAG pipelines for enhanced LLM performance. Integrate AI models with existing security tools, including Endpoint Detection and Response (EDR), Threat and Vulnerability Management (TVM) platforms, and Data Science/Analytics platforms. This will involve working with APIs and understanding data flows. Develop and implement metrics to evaluate the performance of AI models. Monitor deployed models for accuracy and performance and retrain as needed. Adhere to security best practices and ensure that all AI solutions are developed and deployed securely. Consider data privacy and compliance requirements. Work closely with other team members to understand security requirements and translate them into AI-driven solutions. Communicate effectively with stakeholders, including senior management, to present project updates and findings. Stay up to date with the latest advancements in AI/ML and security and identify opportunities to leverage new technologies to improve our security posture. Maintain thorough documentation of AI models, code, and processes. What We Offer Opportunity to work on cutting-edge LLM and RAG projects with global impact. A collaborative environment fostering innovation, research, and skill growth. Competitive salary, comprehensive benefits, and flexible work arrangements. The chance to shape AI-powered features in Sony’s next-generation products. Be able to function in an environment where the team is virtual and geographically dispersed
Education Qualification: Graduate
Skills: AI, NLP, Python, Data science
Must-Haves
Skills
AI, NLP, Python, Data science
NP: Immediate – 30 Days
Job Details
- Job Title: Lead I - Software Engineering - Java, J2Ee, Spring
- Industry: Technology
- Domain - Information technology (IT)
- Experience Required: 6-12 years
- Employment Type: Full Time
- Job Location: Bangalore
- CTC Range: Best in Industry
Job Description:
Role Summary:
We are looking for an experienced Senior Java Developer with expertise in building robust, scalable web applications using Java/J2EE, Spring Boot, REST APIs, and modern microservices architectures. The ideal candidate will be skilled in both back-end and middleware technologies, with strong experience in cloud platforms (AWS), and capable of mentoring junior developers while contributing to high-impact enterprise projects.
The developer will be responsible for full-cycle application development: from interpreting specifications and writing clean, reusable code, to testing, integration, and deployment. You will also work closely with customers and project teams to understand requirements and deliver solutions that optimize cost, performance, and maintainability.
Key Responsibilities:
Application Development & Delivery
- Design, code, debug, test, and document Java-based web applications aligned with design specifications.
- Build scalable and secure microservices using Spring Boot and RESTful APIs.
- Optimize application performance, maintainability, and reusability by using proven design patterns.
- Handle complex data structures and develop multi-threaded, performance-optimized applications.
- Ensure code quality through TDD (JUnit) and best practices.
Cloud & DevOps
- Develop and deploy applications on AWS Cloud Services: EC2, S3, DynamoDB, SNS, SES, etc.
- Leverage containerization tools like Docker and orchestration using Kubernetes.
Integration & Configuration
- Integrate with various databases (PostgreSQL, MySQL, Oracle, NoSQL).
- Configure development environments and CI/CD pipelines as per project needs.
- Follow configuration management processes and ensure compliance.
Testing & Quality Assurance
- Review and create unit test cases, scenarios, and support UAT phases.
- Perform defect root cause analysis (RCA) and proactively implement quality improvements.
Documentation
- Create and review technical documents: HLD, LLD, SAD, user stories, design docs, test cases, and release notes.
- Contribute to project knowledge bases and code repositories.
Team & Project Management
- Mentor team members; conduct code and design reviews.
- Assist Project Manager in effort estimation, planning, and task allocation.
- Set and review FAST goals for yourself and your team; provide regular performance feedback.
Customer Interaction
- Engage with customers to clarify requirements and present technical solutions.
- Conduct product demos and design walkthroughs.
- Interface with customer architects for design finalization.
Key Skills & Tools:
Core Technologies:
- Java/J2EE, Spring Boot, REST APIs
- Object-Oriented Programming (OOP), Design Patterns, Domain-Driven Design (DDD)
- Multithreading, Data Structures, TDD using JUnit
Web & Data Technologies:
- JSON, XML, AJAX, Web Services
- Database Technologies: PostgreSQL, MySQL, Oracle, NoSQL (e.g., DynamoDB)
- Persistence Frameworks: Hibernate, JPA
Cloud & DevOps:
- AWS: S3, EC2, DynamoDB, SNS, SES
- Version Control & Containerization: GitHub, Docker, Kubernetes
Agile & Development Practices:
- Agile methodologies: Scrum or Kanban
- CI/CD concepts
- IDEs: Eclipse, IntelliJ, or equivalent
Expected Outcomes:
- Timely delivery of high-quality code and application components
- Improved performance, cost-efficiency, and maintainability of applications
- High customer satisfaction through accurate requirement translation and delivery
- Team productivity through effective mentoring and collaboration
- Minimal post-production defects and technical issues
Performance Indicators:
- Adherence to coding standards and engineering practices
- On-time project delivery and milestone completion
- Reduction in defect count and issue recurrence
- Knowledge contributions to project and organizational repositories
- Completion of mandatory compliance and technology/domain certifications
Preferred Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Engineering, or related field
- Relevant certifications (e.g., AWS Certified Developer, Oracle Certified, Scrum Master)
Soft Skills:
- Strong analytical and problem-solving mindset
- Excellent communication and presentation skills
- Team leadership and mentorship abilities
- High accountability and ability to work under pressure
- Positive team dynamics and proactive collaboration
Skills
Java, J2Ee, Spring
Must-Haves
Java, J2Ee, Spring
Machine Learning + Aws+ (EKS OR ECS OR Kubernetes) + (Redshift AND Glue) + Sagemaker
NP: Immediate – 30 Days
About Corridor Platforms
Corridor Platforms is a leader in next-generation risk decisioning and responsible AI governance, empowering banks and lenders to build transparent, compliant, and data-driven solutions. Our platforms combine advanced analytics, real-time data integration, and GenAI to support complex financial decision workflows for regulated industries.
Role Overview
As a Backend Engineer at Corridor Platforms, you will:
- Architect, develop, and maintain backend components for our Risk Decisioning Platform.
- Build and orchestrate scalable backend services that automate, optimize, and monitor high-value credit and risk decisions in real time.
- Integrate with ORM layers – such as SQLAlchemy – and multi RDBMS solutions (Postgres, MySQL, Oracle, MSSQL, etc) to ensure data integrity, scalability, and compliance.
- Collaborate closely with Product Team, Data Scientists, QA Teams to create extensible APIs, workflow automation, and AI governance features.
- Architect workflows for privacy, auditability, versioned traceability, and role-based access control, ensuring adherence to regulatory frameworks.
- Take ownership from requirements to deployment, seeing your code deliver real impact in the lives of customers and end users.
Technical Skills
- Languages: Python 3.9+, SQL, JavaScript/TypeScript, Angular
- Frameworks: Flask, SQLAlchemy, Celery, Marshmallow, Apache Spark
- Databases: PostgreSQL, Oracle, SQL Server, Redis
- Tools: pytest, Docker, Git, Nx
- Cloud: Experience with AWS, Azure, or GCP preferred
- Monitoring: Familiarity with OpenTelemetry and logging frameworks
Why Join Us?
- Cutting-Edge Tech: Work hands-on with the latest AI, cloud-native workflows, and big data tools—all within a single compliant platform.
- End-to-End Impact: Contribute to mission-critical backend systems, from core data models to live production decision services.
- Innovation at Scale: Engineer solutions that process vast data volumes, helping financial institutions innovate safely and effectively.
- Mission-Driven: Join a passionate team advancing fair, transparent, and compliant risk decisioning at the forefront of fintech and AI governance.
What We’re Looking For
- Proficiency in Python, SQLAlchemy (or similar ORM), and SQL databases.
- Experience developing and maintaining scalable backend services, including API, data orchestration, ML workflows, and workflow automation.
- Solid understanding of data modeling, distributed systems, and backend architecture for regulated environments.
- Curiosity and drive to work at the intersection of AI/ML, fintech, and regulatory technology.
- Experience mentoring and guiding junior developers.
Ready to build backends that shape the future of decision intelligence and responsible AI?
Apply now and become part of the innovation at Corridor Platforms!
MUST-HAVES:
- LLM, AI, Prompt Engineering LLM Integration & Prompt Engineering
- Context & Knowledge Base Design.
- Context & Knowledge Base Design.
- Experience running LLM evals
NOTICE PERIOD: Immediate – 30 Days
SKILLS: LLM, AI, PROMPT ENGINEERING
NICE TO HAVES:
Data Literacy & Modelling Awareness Familiarity with Databricks, AWS, and ChatGPT Environments
ROLE PROFICIENCY:
Role Scope / Deliverables:
- Scope of Role Serve as the link between business intelligence, data engineering, and AI application teams, ensuring the Large Language Model (LLM) interacts effectively with the modeled dataset.
- Define and curate the context and knowledge base that enables GPT to provide accurate, relevant, and compliant business insights.
- Collaborate with Data Analysts and System SMEs to identify, structure, and tag data elements that feed the LLM environment.
- Design, test, and refine prompt strategies and context frameworks that align GPT outputs with business objectives.
- Conduct evaluation and performance testing (evals) to validate LLM responses for accuracy, completeness, and relevance.
- Partner with IT and governance stakeholders to ensure secure, ethical, and controlled AI behavior within enterprise boundaries.
KEY DELIVERABLES:
- LLM Interaction Design Framework: Documentation of how GPT connects to the modeled dataset, including context injection, prompt templates, and retrieval logic.
- Knowledge Base Configuration: Curated and structured domain knowledge to enable precise and useful GPT responses (e.g., commercial definitions, data context, business rules).
- Evaluation Scripts & Test Results: Defined eval sets, scoring criteria, and output analysis to measure GPT accuracy and quality over time.
- Prompt Library & Usage Guidelines: Standardized prompts and design patterns to ensure consistent business interactions and outcomes.
- AI Performance Dashboard / Reporting: Visualizations or reports summarizing GPT response quality, usage trends, and continuous improvement metrics.
- Governance & Compliance Documentation: Inputs to data security, bias prevention, and responsible AI practices in collaboration with IT and compliance teams.
KEY SKILLS:
Technical & Analytical Skills:
- LLM Integration & Prompt Engineering – Understanding of how GPT models interact with structured and unstructured data to generate business-relevant insights.
- Context & Knowledge Base Design – Skilled in curating, structuring, and managing contextual data to optimize GPT accuracy and reliability.
- Evaluation & Testing Methods – Experience running LLM evals, defining scoring criteria, and assessing model quality across use cases.
- Data Literacy & Modeling Awareness – Familiar with relational and analytical data models to ensure alignment between data structures and AI responses.
- Familiarity with Databricks, AWS, and ChatGPT Environments – Capable of working in cloud-based analytics and AI environments for development, testing, and deployment.
- Scripting & Query Skills (e.g., SQL, Python) – Ability to extract, transform, and validate data for model training and evaluation workflows.
- Business & Collaboration Skills Cross-Functional Collaboration – Works effectively with business, data, and IT teams to align GPT capabilities with business objectives.
- Analytical Thinking & Problem Solving – Evaluates LLM outputs critically, identifies improvement opportunities, and translates findings into actionable refinements.
- Commercial Context Awareness – Understands how sales and marketing intelligence data should be represented and leveraged by GPT.
- Governance & Responsible AI Mindset – Applies enterprise AI standards for data security, privacy, and ethical use.
- Communication & Documentation – Clearly articulates AI logic, context structures, and testing results for both technical and non-technical audiences.
We are looking for experienced Data Engineers who can independently build, optimize, and manage scalable data pipelines and platforms.
In this role, you’ll:
- Work closely with clients and internal teams to deliver robust data solutions powering analytics, AI/ML, and operational systems.
- Mentor junior engineers and bring engineering discipline into our data engagements.
Key Responsibilities
- Design, build, and optimize large-scale, distributed data pipelines for both batch and streaming use cases.
- Implement scalable data models, warehouses/lakehouses, and data lakes to support analytics and decision-making.
- Collaborate with stakeholders to translate business requirements into technical solutions.
- Drive performance tuning, monitoring, and reliability of data pipelines.
- Write clean, modular, production-ready code with proper documentation and testing.
- Contribute to architectural discussions, tool evaluations, and platform setup.
- Mentor junior engineers and participate in code/design reviews.
Must-Have Skills
- Strong programming skills in Python and advanced SQL expertise.
- Deep understanding of ETL/ELT, data modeling (OLTP & OLAP), warehousing, and stream processing.
- Hands-on with distributed data processing frameworks (Apache Spark, Flink, or similar).
- Experience with orchestration tools like Airflow (or similar).
- Familiarity with CI/CD pipelines and Git.
- Ability to debug, optimize, and scale data pipelines in production.
Good to Have
- Experience with cloud platforms (AWS preferred; GCP/Azure also welcome).
- Exposure to Databricks, dbt, or similar platforms.
- Understanding of data governance, quality frameworks, and observability.
- Certifications (e.g., AWS Data Analytics, Solutions Architect, or Databricks).
Other Expectations
- Comfortable working in fast-paced, client-facing environments.
- Strong analytical and problem-solving skills with attention to detail.
- Ability to adapt across tools, stacks, and business domains.
- Willingness to travel within India for short/medium-term client engagements, as needed.
To design, develop, and maintain highly scalable, secure, and efficient backend systems that
power core business applications. The Senior Engineer – Backend will be responsible for
architecting APIs, optimizing data flow, and ensuring system reliability and performance. This
role will collaborate closely with frontend, DevOps, and product teams to deliver robust
solutions that enable seamless user experiences and support organizational growth through
clean, maintainable, and well-tested code.
Responsibilities:
• Design, develop, and maintain robust and scalable backend services using Node.js.
• Collaborate with front-end developers and product managers to define and implement
API specifications.
• Optimize application performance and scalability by identifying bottlenecks and
proposing solutions.
• Write clean, maintainable, and efficient code, and conduct code reviews to ensure
quality standards.
• Develop unit tests and maintain code coverage to ensure high quality.
• Document architectural solutions and system designs to ensure clarity and
maintainability.
• Troubleshoot and resolve issues in development, testing, and production environments.
• Stay up to date with emerging technologies and industry trends to continuously improve
our tech stack.
• Mentor and guide junior engineers, fostering a culture of learning and growth.
Key Skills and Qualifications:
• Bachelor’s degree in Computer Science, Engineering, or a related field (or equivalent
experience).
• 7+ years of experience in backend development with a focus on Node.js and Javascript.
• Strong understanding of RESTful APIs and microservices architecture.
• Proficiency in database technologies (SQL and NoSQL, such as DynamoDB, MongoDB,
PostgreSQL, etc.).
• Familiarity with containerization and orchestration technologies (Docker, Kubernetes).
• Knowledge of cloud platform (AWS) and deployment best practices.
• Excellent problem-solving skills and the ability to work in a fast-paced environment.
• Strong communication and teamwork skills.
Good to have:
• Experience with front-end frameworks (e.g. Angular, React, Vue.js).
• Understanding of HTML, CSS, and JavaScript.
• Familiarity with responsive design and user experience principles.
Who we are: My AI Client is building the foundational platform for the "agentic economy," moving beyond simple chatbots to create an ecosystem for autonomous AI agents and they aim to provide tools for developers to launch, manage, and monetize AI agents as "digital coworkers."
The Challenge
The current AI stack is fragmented, leading to issues with multimodal data, silent webhook failures, unpredictable token usage, and nascent agent-to-agent collaboration. My AI Client is building a unified, robust backend to resolve these issues for the developer community.
Your Mission
As a foundational member of the backend team, you will architect core systems, focusing on:
- Agent Nervous System: Designing agent-to-agent messaging, lifecycle management, and high-concurrency, low-latency communication.
- Multimodal Chaos Taming: Engineering systems to process and understand real-time images, audio, video, and text.
- Bulletproof Systems: Developing secure, observable webhook systems with robust billing, metering, and real-time payment pipelines.
What You'll Bring
- My AI Client seeks an experienced engineer comfortable with complex systems and ambiguity.
Core Experience:
● Typically 3 to 5 years of experience in backend engineering roles.
● Expertise in Python, especially with async frameworks like FastAPI.
● Strong command of Docker and cloud deployment (AWS, Cloud Run, or similar).
● Proven experience designing and building microservice or agent-based architectures.
Specialized Experience (Ideal):
- Real-Time Systems: Experience with real-time media transmission like WebRTC, WebSockets and ways to process them.
- Scalable Systems: Experience in building scalable, fault-tolerant systems with a strong understanding of observability, monitoring, and alerting best practices.
- Reliable Webhooks: Knowledge of scalable webhook infrastructure with retry logic, backoffs, and security.
- Data Processing: Experience with multimodal data (e.g., OCR, audio transcription, video chunking with FFmpeg/OpenCV).
- Payments & Metering: Familiarity with usage-based billing systems or token-based ledgers.
Your Impact
- The systems designed by this role will form the foundation for:
- Thousands of AI agents for major partners across chat, video, and APIs.
- A new creator economy enabling developers to earn revenue through agents.
- The overall speed, security, and scalability of my client’s AI platform.
Why Join Us?
- Opportunity to solve hard problems with clean, scalable code.
- Small, fast-paced team with high ownership and zero micromanagement.
- Belief in platform engineering as a craft and care for developer experience.
- Conviction that AI agents are the future, and a desire to build their powering platform.
- Dynamic, collaborative in-office work environment in Bengaluru in a Hybrid setup (weekly 2 days from office)
- Meaningful equity in a growing, well-backed company.
- Direct work with founders and engineers from top AI companies.
- A real voice in architectural and product decisions.
- Opportunity to solve cutting-edge problems with no legacy code.
Ready to Build the Future?
My AI Client is building the core platform for the next software paradigm. Interested candidates are encouraged to apply with their GitHub, resume, or anything that showcases their thinking.
Like us, you'll be deeply committed to delivering impactful outcomes for customers.
- 7+ years of demonstrated ability to develop resilient, high-performance, and scalable code tailored to application usage demands.
- Ability to lead by example with hands-on development while managing project timelines and deliverables. Experience in agile methodologies and practices, including sprint planning and execution, to drive team performance and project success.
- Deep expertise in Node.js, with experience in building and maintaining complex, production-grade RESTful APIs and backend services.
- Experience writing batch/cron jobs using Python and Shell scripting.
- Experience in web application development using JavaScript and JavaScript libraries.
- Have a basic understanding of Typescript, JavaScript, HTML, CSS, JSON and REST based applications.
- Experience/Familiarity with RDBMS and NoSQL Database technologies like MySQL, MongoDB, Redis, ElasticSearch and other similar databases.
- Understanding of code versioning tools such as Git.
- Understanding of building applications deployed on the cloud using Google cloud platform(GCP)or Amazon Web Services (AWS)
- Experienced in JS-based build/Package tools like Grunt, Gulp, Bower, Webpack.
Key Responsibilities:
- Application Development: Design and implement both client-side and server-side architecture using JavaScript frameworks and back-end technologies like Golang.
- Database Management: Develop and maintain relational and non-relational databases (MySQL, PostgreSQL, MongoDB) and optimize database queries and schema design.
- API Development: Build and maintain RESTfuI APIs and/or GraphQL services to integrate with front-end applications and third-party services.
- Code Quality & Performance: Write clean, maintainable code and implement best practices for scalability, performance, and security.
- Testing & Debugging: Perform testing and debugging to ensure the stability and reliability of applications across different environments and devices.
- Collaboration: Work closely with product managers, designers, and DevOps engineers to deliver features aligned with business goals.
- Documentation: Create and maintain documentation for code, systems, and application architecture to ensure knowledge transfer and team alignment.
Requirements:
- Experience: 1+ years in backend development in micro-services ecosystem, with proven experience in front-end and back-end frameworks.
- 1+ years experience Golang is mandatory
- Problem-Solving & DSA: Strong analytical skills and attention to detail.
- Front-End Skills: Proficiency in JavaScript and modern front-end frameworks (React, Angular, Vue.js) and familiarity with HTML/CSS.
- Back-End Skills: Experience with server-side languages and frameworks like Node.js, Express, Python or GoLang.
- Database Knowledge: Strong knowledge of relational databases (MySQL, PostgreSQL) and NoSQL databases (MongoDB).
- API Development: Hands-on experience with RESTfuI API design and integration, with a plus for GraphQL.
- DevOps Understanding: Familiarity with cloud platforms (AWS, Azure, GCP) and containerization (Docker, Kubernetes) is a bonus.
- Soft Skills: Excellent problem-solving skills, teamwork, and strong communication abilities.
Nice-to-Have:
- UI/UX Sensibility: Understanding of responsive design and user experience principles.
- CI/CD Knowledge: Familiarity with CI/CD tools and workflows (Jenkins, GitLab CI).
- Security Awareness: Basic understanding of web security standards and best practices.
Job Role
The role is for an SAP UI5 Consultant responsible for developing and enhancing web applications using SAP UI5 and related technologies.
Responsibilities
- Develop Fiori-like web applications based on SAPUI5.
- Implement REST Web services and enhancements to SAP Fiori apps.
- Work with Web Technologies including HTML5, CSS3, and JavaScript.
- Participate in knowledge transfers and evaluate new technologies.
- Understand and implement software architecture for enterprise applications.
Qualifications
The candidate should have a BA/BE/BTech qualification with excellent verbal and written communication skills in English. Ability to work flexible hours is necessary.
🚀 Hiring: QA Engineer (Manual + Automation)
⭐ Experience: 3+ Years
📍 Location: Bangalore
⭐ Work Mode:- Hybrid
⏱️ Notice Period: Immediate Joiners
(Only immediate joiners & candidates serving notice period)
💫 About the Role:
We’re looking for a skilled QA Engineer You’ll ensure product quality through manual and automated testing across web, mobile, and APIs — working with tools and technologies like Postman, Playwright, Appium, Rest Assured, GCP/AWS, and React/Next.js.
Key Responsibilities:
✅ Develop & maintain automated tests using Cucumber, Playwright, Pytest, etc.
✅ Perform API testing using Postman.
✅ Work on cloud platforms (GCP/AWS) and CI/CD (Jenkins).
✅ Test web & mobile apps (Appium, BrowserStack, LambdaTest).
✅ Collaborate with developers to ensure seamless releases.
Must-Have Skills:
✅ API Testing (Postman)
✅ Cloud (GCP / AWS)
✅ Frontend understanding (React / Next.js)
✅ Strong SQL & Git skills
✅ Familiarity with OpenAI APIs
Role: Full-Time
Work Location: Bangalore (Client Location – LeadSquared)
Address: 2nd & 3rd Floor, Omega, Embassy Tech Square, Marathahalli - Sarjapur Outer Ring Rd, Kaverappa Layout, Kadubeesanahalli, Bellandur, Bengaluru, Karnataka – 560103
Interview Process: Test and Technical Discussion (Face-to-Face at Client Location)
Work Mode: 4 Days Work from Office
Preference: Local Bangalore Candidates
Responsibilities & Skills Required
- 4–6 years of experience in building Web Applications & APIs
- Proficient in C#
- Web Framework: React.js
- API Framework: .NET Core
- Database: MySQL or SQL Server
- Strong knowledge of multi-threading and asynchronous programming
- Experience with Cloud Platforms: AWS, GCP, or Azure
- Strong SQL programming skills with experience in handling large datasets (millions of records)
- Ability to write clean, maintainable, and scalable code
- Sound understanding of scalable web application design principles
MUST-HAVES:
- LLM, AI, Prompt Engineering LLM Integration & Prompt Engineering
- Context & Knowledge Base Design.
- Context & Knowledge Base Design.
- Experience running LLM evals
NOTICE PERIOD: Immediate – 30 Days
SKILLS: LLM, AI, PROMPT ENGINEERING
NICE TO HAVES:
Data Literacy & Modelling Awareness Familiarity with Databricks, AWS, and ChatGPT Environments
ROLE PROFICIENCY:
Role Scope / Deliverables:
- Scope of Role Serve as the link between business intelligence, data engineering, and AI application teams, ensuring the Large Language Model (LLM) interacts effectively with the modeled dataset.
- Define and curate the context and knowledge base that enables GPT to provide accurate, relevant, and compliant business insights.
- Collaborate with Data Analysts and System SMEs to identify, structure, and tag data elements that feed the LLM environment.
- Design, test, and refine prompt strategies and context frameworks that align GPT outputs with business objectives.
- Conduct evaluation and performance testing (evals) to validate LLM responses for accuracy, completeness, and relevance.
- Partner with IT and governance stakeholders to ensure secure, ethical, and controlled AI behavior within enterprise boundaries.
KEY DELIVERABLES:
- LLM Interaction Design Framework: Documentation of how GPT connects to the modeled dataset, including context injection, prompt templates, and retrieval logic.
- Knowledge Base Configuration: Curated and structured domain knowledge to enable precise and useful GPT responses (e.g., commercial definitions, data context, business rules).
- Evaluation Scripts & Test Results: Defined eval sets, scoring criteria, and output analysis to measure GPT accuracy and quality over time.
- Prompt Library & Usage Guidelines: Standardized prompts and design patterns to ensure consistent business interactions and outcomes.
- AI Performance Dashboard / Reporting: Visualizations or reports summarizing GPT response quality, usage trends, and continuous improvement metrics.
- Governance & Compliance Documentation: Inputs to data security, bias prevention, and responsible AI practices in collaboration with IT and compliance teams.
KEY SKILLS:
Technical & Analytical Skills:
- LLM Integration & Prompt Engineering – Understanding of how GPT models interact with structured and unstructured data to generate business-relevant insights.
- Context & Knowledge Base Design – Skilled in curating, structuring, and managing contextual data to optimize GPT accuracy and reliability.
- Evaluation & Testing Methods – Experience running LLM evals, defining scoring criteria, and assessing model quality across use cases.
- Data Literacy & Modeling Awareness – Familiar with relational and analytical data models to ensure alignment between data structures and AI responses.
- Familiarity with Databricks, AWS, and ChatGPT Environments – Capable of working in cloud-based analytics and AI environments for development, testing, and deployment.
- Scripting & Query Skills (e.g., SQL, Python) – Ability to extract, transform, and validate data for model training and evaluation workflows.
- Business & Collaboration Skills Cross-Functional Collaboration – Works effectively with business, data, and IT teams to align GPT capabilities with business objectives.
- Analytical Thinking & Problem Solving – Evaluates LLM outputs critically, identifies improvement opportunities, and translates findings into actionable refinements.
- Commercial Context Awareness – Understands how sales and marketing intelligence data should be represented and leveraged by GPT.
- Governance & Responsible AI Mindset – Applies enterprise AI standards for data security, privacy, and ethical use.
- Communication & Documentation – Clearly articulates AI logic, context structures, and testing results for both technical and non-technical audiences.
🚀 Your Role at a Glance
We are hiring a Senior Staff Backend Engineer – Site Reliability for our client (Code Name: SORIN), a global leader building high-scale observability platforms. In this high-impact leadership role, you’ll architect, scale, and optimise the systems that drive how enterprises monitor their distributed applications. You’ll collaborate across teams, mentor engineers, and shape the technical direction of mission-critical backend services in a modern cloud-native environment.
This role is ideal for experienced backend engineers who thrive in distributed systems, care deeply about system performance and observability, and want to influence large-scale technical decisions that span multiple teams.
🏢 About the Client (Code Name: SORIN)
Our client is a global enterprise software company focused on simplifying IT management through powerful, secure, and scalable platforms. With a strong commitment to innovation and customer-centricity, they help organisations accelerate digital transformation across observability, incident response, and performance monitoring. You’ll join a team of passionate engineers working on critical systems trusted by Fortune 500 companies and growing SaaS-first businesses alike.
🔧 Key Responsibilities
● Collaborate with software engineering teams to define infrastructure requirements, drive best practices in reliability, monitoring, incident response, and automation, ensuring seamless integration and optimal
performance of applications and systems
● Lead and mentor a team of SREs, providing technical guidance and
support to ensure the ongoing reliability and performance of our systems
● Play a key role in driving the automation, tools, and observability initiatives, assuming ownership of designing and implementing scalable and efficient solutions.
● Leading the response to production incidents, conducting comprehensive learning reviews, driving continuous improvement initiatives, and actively participating in an on-call rotation, fostering a culture of learning, resilience, and ongoing enhancement within our systems.
● Establish and drive operations performance through SLOs
● Demonstrate proficiency in technical skills, exhibit an expert-level understanding of relevant technologies and tools, and use this knowledge to mentor and support team members, helping them improve their skills and succeed in their roles.
✅ Required Skills & Experience
● At least 10 + years of experience designing, building & maintaining SAAS environments
● 8+ years of experience designing, building & maintaining AWS/AZURE infrastructure with Terraform
● 5+ years of experience building and running Kubernetes clusters
● Strong experience working with data platform infrastructure.
● Experience with coding in a high-level programming language like Python, Go (Golang). Knowledge of writing shell scripts, SQL queries.
● Experience with observability (monitoring – logging, tracing, metrics)
● Experience with SQL/NoSQL database technologies.
● Experience with GitOps and CI/CD processes is a plus
● Experience with security operations – security policies, infrastructure, key management, setup of encryption at rest and transport.
● Experience in mentoring and fostering the professional development of team members, promoting a culture of continuous learning and collaboration.
🌟 Nice to Have
● Strong customer orientation
● Excellent interpersonal and organisational skills
● Attention to detail and focus on quality
● Strong communication skills to effectively liaise with both technical and non-technical staff
● Ability to act decisively and work well under pressure
● Must be a collaborative problem solver
● Strong bias for ownership and action
💡 Why Join This Role
● Architect foundational systems for one of the world’s most trusted observability platforms
● Join a fast-paced, mission-driven team with global reach
● Collaborate with senior leaders and contribute to technology direction across the company
● Drive innovation in how modern engineering teams monitor and respond to incidents
● Competitive compensation, strong engineering culture, and direct influence on product direction
Apply Here: https://airtable.com/app0U8lKNDT4KiEnD/pagKI67YRVCs8Yrrn/form?prefill_Applied+Position=recOhfqwuEbUIAsCb
Join the MachineHack - AIM Media House Team
At MachineHack, we craft AI-powered platforms and products that push the boundaries of
learning and innovation. From MachineHack (the hackathon hub) to CloudLab (hands-on AI
playground) to Datalyze (data-driven insights), plus countless experimental AI projects, we’re
always building, always shipping.
We’re not just coding software, we’re creating experiences where people can learn, compete, and create. If you’re passionate about solving problems, experimenting with AI, and vibing with a team that learns every day, you’ll feel right at home here
Why Work With Us?
● Be part of a fast-moving AI-first team where new ideas turn into products quickly.
● Work with modern tech stacks across AI, backend, frontend, and DevOps.
● Collaborate with engineers who are learners first,humble, supportive, and curious.
● Opportunity to contribute to high-impact platforms used by thousands in the AI/ML community.
● A culture that values learning > titles and team vibe > ego.
Our Tech Playground
We don’t shy away from complexity. We thrive in it. Here’s what we love working with:
● Backend & AI: Node.js, Python, TypeScript, Low-Level System Design, DSA, Prompt Engineering
● Frontend: React, Next.js, TypeScript
● DevOps & Cloud: AWS (EC2, Lambda, Terraform, CI/CD), Docker, Kubernetes, Helm Charts
● Testing & Quality: Playwright, Selenium, Unit Testing frameworks
● Infra & Systems: Container orchestration, Virtual Machines, AWS Infra Deployment
● Bonus Points: ML/AI model experience, guiding juniors, proven UX understanding
What Makes You a Good Fit
● 2–6 years of experience in software development
● Strong grounding in JS, Python, Node, and cloud-native tools
● Comfortable with system design, backend architecture, and team collaboration
● Can mentor juniors and handle multiple projects at once
● Fluent in English and able to communicate ideas clearly
● Humble, supportive, approachable,like a teammate and a friend
● Excited to learn, build, and experiment in AI projects
● Bonus: Some ML exposure (but not mandatory)
Position: Lead Software Engineer
Practice: Development
Reporting To: Project Manager
Experience: 8 - 10 Years
Job Summary
We are seeking a highly skilled Lead Software Engineer to lead a team of software developers in designing, developing, testing, and maintaining high-quality software applications. The ideal candidate will provide technical guidance, collaborate with cross-functional teams, and ensure that software solutions align with business objectives and user requirements
Key Responsibilities
Technical Leadership
- Lead, mentor, and guide a team of developers, providing technical direction, coaching, and performance feedback.
- Define and implement architectural designs for scalable, robust, and maintainable applications.
- Ensure adherence to coding standards, quality assurance practices, and performance optimization.
Full-Stack Development
- Design and develop responsive and user-friendly interfaces using React.js and reusable component architecture.
- Build and maintain backend services using Node.js , including RESTful APIs, business logic, and data integrations.
- Implement real-time features and integrate third-party APIs for enhanced functionality.
Project Management & Collaboration
- Collaborate with product managers, designers, and stakeholders to translate requirements into effective technical solutions.
- Participate in design and technical discussions, evaluating alternatives and mitigating potential risks.
- Oversee the end-to-end software development lifecycle, from requirement analysis to deployment and post-release support.
Quality Assurance & DevOps
- Conduct regular code reviews to ensure clean, maintainable, and well-tested code.
- Identify and resolve complex technical issues and performance bottlenecks.
- Contribute to cloud deployment strategies, CI/CD pipelines, and containerization practices.
Innovation & Continuous Improvement
- Stay updated with emerging technologies and frameworks in full-stack development.
- Recommend process improvements and technical upgrades to enhance system performance and team efficiency.
Mandatory Skills
Frontend: JavaScript, React.js, Redux, HTML5 CSS3
Backend: Node.js, Express.js, REST API design
Database: PostgreSQL, MongoDB, or other relational/non-relational databases
DevOps & Cloud: Familiarity with AWS, Azure, GCP, CI/CD pipelines
Version Control: Git and branching strategies
Architecture: Scalable design principles, microservices architecture
Desired Skills
- Experience with Docker, Kubernetes, and containerized deployments.
- Exposure to Agile/Scrum development methodologies.
- Familiarity with performance monitoring and application observability tools.
- Strong analytical and problem-solving capabilities.
- Excellent communication and leadership skills to collaborate effectively with stakeholders.
Qualifications
- Bachelor's degree in Computer Science, Information Technology, or related field.
- Proven experience as a full-stack developer, with significant hands-on experience in React.js and Node.js .
- Demonstrated experience in leading and managing software development teams.
- Solid understanding of software development methodologies and best practices.
- Passion for innovation, continuous learning, and delivering high-quality solutions.
Company Name – Wissen Technology
Location : Pune / Bangalore / Mumbai (Based on candidate preference)
Work mode: Hybrid
Experience: 5+ years
Job Description
Wissen Technology is seeking an experienced C# .NET Developer to build and maintain applications related to streaming market data. This role involves developing message-based C#/.NET applications to process, normalize, and summarize large volumes of market data efficiently. The candidate should have a strong foundation in Microsoft .NET technologies and experience working with message-driven, event-based architecture. Knowledge of capital markets and equity market data is highly desirable.
Responsibilities
- Design, develop, and maintain message-based C#/.NET applications for processing real-time and batch market data feeds.
- Build robust routines to download and process data from AWS S3 buckets on a frequent schedule.
- Implement daily data summarization and data normalization routines.
- Collaborate with business analysts, data providers, and other developers to deliver high-quality, scalable market data solutions.
- Troubleshoot and optimize market data pipelines to ensure low latency and high reliability.
- Contribute to documentation, code reviews, and team knowledge sharing.
Required Skills and Experience
- 5+ years of professional experience programming in C# and Microsoft .NET framework.
- Strong understanding of message-based and real-time programming architectures.
- Experience working with AWS services, specifically S3, for data retrieval and processing.
- Experience with SQL and Microsoft SQL Server.
- Familiarity with Equity market data, FX, Futures & Options, and capital markets concepts.
- Excellent interpersonal and communication skills.
- Highly motivated, curious, and analytical mindset with the ability to work well both independently and in a team environment.
Education
- Bachelor’s degree in Computer Science, Engineering, or a related technical field.
Who We Are:
Increff is the most preferred retail SaaS partner, solving complex inventory management and supply chain challenges for retailers seeking to revolutionize their supply chains both technologically and operationally
.
What We Offer:
Our core focus is on providing innovative retail tech solutions, including merchandising and omni channel inventory management. These solutions are meticulously designed to cater to the specific needs of brands and retailers, empowering them to thrive in the dynamic marketplace Innovative technology, comprehensive support, and a dynamic environment for career growth.
Our Vision:
To be the most admired retail technology company.
Role Overview:
We’re looking for a passionate Software Engineer to design, develop and install software solutions. Ideally, the candidate should be able to build high-quality, innovative and fully performing software in compliance with coding standards and technical design. Software engineers must be skilled in development, writing code, and documenting functionality.
Responsibilities:
- Executing full life-cycle software development
- Writing well designed, testable, efficient code
- Producing specifications and determining operational feasibility
- Integrating software components into fully functional software systems
- Developing software verification plans and quality assurance procedures
- Documenting and maintaining software functionality
- Tailoring and deploying software tools, processes and metrics
- Serving as a subject matter expert
- Complying with project plans and industry standards
Requirements
- Bachelor’s degree in Engineering, Computer Science, or a related field.
- Proven work experience as a software engineer or software developer.
- Proficiency in Python and Java for backend development, with experience in object-oriented programming concepts.
- Strong hands-on skills in SQL and working with relational databases, including query optimization and ORM frameworks (e.g., Hibernate, JPA2).
- Advanced knowledge of data structures, algorithms, and core system design fundamentals.
- Demonstrated experience with at least one web application framework such as Flask or Spring.
- Sound understanding of software engineering best practices, with familiarity in test-driven development and software development methodologies.
- Experience with cloud platforms and Big Data concepts is a plus (e.g., Apache Spark, Trino, Azure, Azure Synapse Analytics).
- Good knowledge of statistical concepts and data analysis is advantageous.
- Strong problem-solving skills, aptitude, and the ability to take ownership and accountability of deliverables.
- Excellent communication skills and the ability to document requirements, designs, and specifications effectively.
Our Culture:
At Increff, we take great pride in fostering an open, flexible, and collaborative workplace. Our culture empowers employees to innovate, build fulfilling careers, and enjoy their work. Moreover, we strongly encourage the development of leadership skills from within the organization. Our commitment to transparency ensures that at every level, individuals have the autonomy to initiate, take ownership of projects, and successfully execute them.

One of the reputed Client in India
Our Client is looking to hire Databricks Amin immediatly.
This is PAN-INDIA Bulk hiring
Minimum of 6-8+ years with Databricks, Pyspark/Python and AWS.
Must have AWS
Notice 15-30 days is preferred.
Share profiles at hr at etpspl dot com
Please refer/share our email to your friends/colleagues who are looking for job.
Job Description:
We’re looking for a Software Engineer who’s passionate about building scalable, high-performance applications using Java and Kotlin on AWS. You’ll collaborate closely with cross-functional teams — including DevOps, QA, and Product — to design, develop, and deliver robust software solutions.
Our culture values collaboration, ownership, and continuous learning. We embrace modern technologies to solve real-world problems and continuously evolve our platform to meet changing business needs.
Key Responsibilities
- Develop and enhance features to collect, process, and deliver user-generated content.
- Collaborate with engineers, product managers, and designers to build end-to-end solutions.
- Write clean, maintainable, and efficient code following best practices.
- Participate in design discussions, code reviews, and technical brainstorming sessions.
- Identify and fix performance bottlenecks and technical issues.
- Contribute to CI/CD pipelines, infrastructure automation, and developer tooling initiatives.
- Maintain and improve application reliability, scalability, and security.
Required Skills & Experience
- 4+ years of full-stack development experience using Java/Kotlin and AWS.
- Hands-on experience in API development for front-end applications.
- Strong understanding of relational databases (PostgreSQL, MySQL) and handling large datasets.
- Experience working with CI/CD tools (CircleCI, GitHub Actions, Drone).
- Experience with Infrastructure as Code using Terraform.
- Exposure to event-driven architectures (SQS, SNS, Kafka, Kinesis, Pub/Sub) and idempotent patterns.
Nice-to-Have Skills
- Familiarity with automated database migration tools (Liquibase, Flyway).
- Experience with cloud storage (AWS S3, GCP Cloud Storage, Azure Blob) or document data stores (DynamoDB, MongoDB).
- Experience containerizing applications using Docker and deploying to ECS or Kubernetes.
- Proficiency with Git and collaborative development workflows.
Job Title: Mid-Level .NET Developer (Agile/SCRUM)
Location: Mohali, PTP or anywhere else)
Night Shift from 6:30 pm to 3:30 am IST
Experience:
5 Years
Job Summary:
We are seeking a proactive and detail-oriented Mid-Level .NET Developer to join our dynamic team. The ideal candidate will be responsible for designing, developing, and maintaining high-quality applications using Microsoft technologies with a strong emphasis on .NET Core, C#, Web API, and modern front-end frameworks. You will collaborate with cross-functional teams in an Agile/SCRUM environment and participate in the full software development lifecycle—from requirements gathering to deployment—while ensuring adherence to best coding and delivery practices.
Key Responsibilities:
- Design, develop, and maintain applications using C#, .NET, .NET Core, MVC, and databases such as SQL Server, PostgreSQL, and MongoDB.
- Create responsive and interactive user interfaces using JavaScript, TypeScript, Angular, HTML, and CSS.
- Develop and integrate RESTful APIs for multi-tier, distributed systems.
- Participate actively in Agile/SCRUM ceremonies, including sprint planning, daily stand-ups, and retrospectives.
- Write clean, efficient, and maintainable code following industry best practices.
- Conduct code reviews to ensure high-quality and consistent deliverables.
- Assist in configuring and maintaining CI/CD pipelines (Jenkins or similar tools).
- Troubleshoot, debug, and resolve application issues effectively.
- Collaborate with QA and product teams to validate requirements and ensure smooth delivery.
- Support release planning and deployment activities.
Required Skills & Qualifications:
- 4–6 years of professional experience in .NET development.
- Strong proficiency in C#, .NET Core, MVC, and relational databases such as SQL Server.
- Working knowledge of NoSQL databases like MongoDB.
- Solid understanding of JavaScript/TypeScript and the Angular framework.
- Experience in developing and integrating RESTful APIs.
- Familiarity with Agile/SCRUM methodologies.
- Basic knowledge of CI/CD pipelines and Git version control.
- Hands-on experience with AWS cloud services.
- Strong analytical, problem-solving, and debugging skills.
- Excellent communication and collaboration skills.
Preferred / Nice-to-Have Skills:
- Advanced experience with AWS services.
- Knowledge of Kubernetes or other container orchestration platforms.
- Familiarity with IIS web server configuration and management.
- Experience in the healthcare domain.
- Exposure to AI-assisted code development tools (e.g., GitHub Copilot, ChatGPT).
- Experience with application security and code quality tools such as Snyk or SonarQube.
- Strong understanding of SOLID principles and clean architecture patterns.
Technical Proficiencies:
- ASP.NET Core, ASP.NET MVC
- C#, Entity Framework, Razor Pages
- SQL Server, MongoDB
- REST API, jQuery, AJAX
- HTML, CSS, JavaScript, TypeScript, Angular
- Azure Services, Azure Functions, AWS
- Visual Studio
- CI/CD, Git
We are seeking a mid-to-senior level Full-Stack Developer with a foundational understanding of software development, cloud services, and database management. In this role, you will contribute to both the front-end and back-end of our application. focusing on creating a seamless user experience, supported by robust and scalable cloud infrastructure.
Key Responsibilities
● Develop and maintain user-facing features using React.js and TypeScript.
● Write clean, efficient, and well-documented JavaScript/TypeScript code.
● Assist in managing and provisioning cloud infrastructure on AWS using Infrastructure as Code (IaC) principles.
● Contribute to the design, implementation, and maintenance of our databases.
● Collaborate with senior developers and product managers to deliver high-quality software.
● Troubleshoot and debug issues across the full stack.
● Participate in code reviews to maintain code quality and share knowledge.
Qualifications
● Bachelor's degree in Computer Science, a related technical field, or equivalent practical experience.
● 5+ years of professional experience in web development.
● Proficiency in JavaScript and/or TypeScript.
● Proficiency in Golang and Python.
● Hands-on experience with the React.js library for building user interfaces.
● Familiarity with Infrastructure as Code (IaC) tools and concepts (e.g.(AWS CDK, Terraform, or CloudFormation).
● Basic understanding of AWS and its core services (e.g., S3, EC2, Lambda, DynamoDB).
● Experience with database management, including relational (e.g., PostgreSQL) or NoSQL (e.g., DynamoDB, MongoDB) databases.
● Strong problem-solving skills and a willingness to learn.
● Familiarity with modern front-end build pipelines and tools like Vite and Tailwind CSS.
● Knowledge of CI/CD pipelines and automated testing.
Role: Data Scientist (Python + R Expertise)
Exp: 8 -12 Years
CTC: up to 30 LPA
Required Skills & Qualifications:
- 8–12 years of hands-on experience as a Data Scientist or in a similar analytical role.
- Strong expertise in Python and R for data analysis, modeling, and visualization.
- Proficiency in machine learning frameworks (scikit-learn, TensorFlow, PyTorch, caret, etc.).
- Strong understanding of statistical modeling, hypothesis testing, regression, and classification techniques.
- Experience with SQL and working with large-scale structured and unstructured data.
- Familiarity with cloud platforms (AWS, Azure, or GCP) and deployment practices (Docker, MLflow).
- Excellent analytical, problem-solving, and communication skills.
Preferred Skills:
- Experience with NLP, time series forecasting, or deep learning projects.
- Exposure to data visualization tools (Tableau, Power BI, or R Shiny).
- Experience working in product or data-driven organizations.
- Knowledge of MLOps and model lifecycle management is a plus.
If interested kindly share your updated resume on 82008 31681
Job Summary:
Deqode is looking for a highly motivated and experienced Python + AWS Developer to join our growing technology team. This role demands hands-on experience in backend development, cloud infrastructure (AWS), containerization, automation, and client communication. The ideal candidate should be a self-starter with a strong technical foundation and a passion for delivering high-quality, scalable solutions in a client-facing environment.
Key Responsibilities:
- Design, develop, and deploy backend services and APIs using Python.
- Build and maintain scalable infrastructure on AWS (EC2, S3, Lambda, RDS, etc.).
- Automate deployments and infrastructure with Terraform and Jenkins/GitHub Actions.
- Implement containerized environments using Docker and manage orchestration via Kubernetes.
- Write automation and scripting solutions in Bash/Shell to streamline operations.
- Work with relational databases like MySQL and SQL, including query optimization.
- Collaborate directly with clients to understand requirements and provide technical solutions.
- Ensure system reliability, performance, and scalability across environments.
Required Skills:
- 3.5+ years of hands-on experience in Python development.
- Strong expertise in AWS services such as EC2, Lambda, S3, RDS, IAM, CloudWatch.
- Good understanding of Terraform or other Infrastructure as Code tools.
- Proficient with Docker and container orchestration using Kubernetes.
- Experience with CI/CD tools like Jenkins or GitHub Actions.
- Strong command of SQL/MySQL and scripting with Bash/Shell.
- Experience working with external clients or in client-facing roles.
Preferred Qualifications:
- AWS Certification (e.g., AWS Certified Developer or DevOps Engineer).
- Familiarity with Agile/Scrum methodologies.
- Strong analytical and problem-solving skills.
- Excellent communication and stakeholder management abilities.
At Rocketium, engineers aren’t just writing code. We’re building the engine that powers creativity at scale. Imagine helping a global brand launch hundreds of personalized ads in a day instead of weeks, or designing workflows where AI takes care of the grunt work so humans can focus on storytelling. That’s the kind of impact your code will have here.
As a Software Development Engineer with us, you will:
- Bring ideas to life: Work with product managers and designers to turn messy, real-world creative challenges into elegant software solutions.
- Build for scale: Write reliable, performant code that can handle thousands of assets being generated, automated, and personalized every day.
- Shape the future of AI workflows: Experiment with AI/automation to create features that change how creative teams think about production.
- Make the platform better, every day: Debug tricky issues, optimize performance, and refactor where needed — because we value craft as much as speed.
- Grow with the team: Share openly, learn from peers, and give feedback that sharpens everyone’s work.
Who you are:
- Your toolkit includes JavaScript/TypeScript, React, and Node.js, and you’re quick to pick up new tools as needed.
- You’re comfortable switching hats: debugging tricky issues one moment, brainstorming workflows with designers the next.
- Ambiguity doesn’t scare you; you enjoy finding clarity where none exists.
- Bonus points if you’ve dabbled in AI, workflow automation, or creative tech.
- Most importantly, you embody our prime directives.
Job Title : Senior QA Automation Architect (Cloud & Kubernetes)
Experience : 6+ Years
Location : India (Multiple Offices)
Shift Timings : 12 PM to 9 PM (Noon Shift)
Working Days : 5 Days WFO (NO Hybrid)
About the Role :
We’re looking for a Senior QA Automation Architect with deep expertise in cloud-native systems, Kubernetes, and automation frameworks.
You’ll design scalable test architectures, enhance automation coverage, and ensure product reliability across hybrid-cloud and distributed environments.
Key Responsibilities :
- Architect and maintain test automation frameworks for microservices.
- Integrate automated tests into CI/CD pipelines (Jenkins, GitHub Actions).
- Ensure reliability, scalability, and observability of test systems.
- Work closely with DevOps and Cloud teams to streamline automation infrastructure.
Mandatory Skills :
- Kubernetes, Helm, Docker, Linux
- Cloud Platforms : AWS / Azure / GCP
- CI/CD Tools : Jenkins, GitHub Actions
- Scripting : Python, Pytest, Bash
- Monitoring & Performance : Prometheus, Grafana, Jaeger, K6
- IaC Practices : Terraform / Ansible
Good to Have :
- Experience with Service Mesh (Istio/Linkerd).
- Container Security or DevSecOps exposure.
About the Role
We are seeking an experienced Next.js Developer with minimum 6 years of hands-on experience to join our development team. The ideal candidate will have a strong background in architecting scalable applications, deploying projects, and writing robust test cases. You will play a key role in the full software development lifecycle, from initial concept and architecture through deployment and maintenance.
Key Responsibilities
· Design, develop, and maintain high-performance, scalable web applications using Next.js
· Architect project structure for maintainability, scalability, and performance, following best practices for folder and component organization
· Lea d and participate in project deployments using modern CI/CD pipelines and cloud platforms. AWS is must
· Write comprehensive unit, integration, and end-to-end test cases using frameworks like Jest, React Testing Library, or Cypress
· Collaborate with cross-functional teams (designers, backend developers, product managers) to deliver seamless user experiences and integrate APIs (REST/GraphQL)
· Optimize applications for SEO, server-side rendering (SSR), static site generation (SSG), and performance
· Conduct code reviews, enforce coding standards, and mentor junior developers
· Troubleshoot, debug, and resolve issues across the stack to ensure reliability and security
· Maintain technical documentation, including architecture decisions, deployment guides, and test plans
Required Skills and Qualifications
· Development experience of 6+ years in Reactjs and Nextjs (preferred)
· Proven expertise in Leading, architecting and organizing large-scale Nextjs projects.
· Hands-on experience with project deployment and cloud platform as AWS (preferred)
· Strong proficiency in JavaScript (ES6+), TypeScript, MaterialUI and CSS Frameworks
· Experience writing and maintaining test cases using Jest, React Testing Library, Cypress, or similar tools
· Familiarity with state management libraries (Redux, Context API) and API integration (REST, GraphQL)
· Working knowledge of version control systems (Git) and CI/CD pipelines (preferred)
· Solid understanding of SSR, SSG, and SEO best practices
· Excellent problem-solving, communication, and teamwork skills
· Bachelor’s degree in Computer Science, Information Technology, or a related field (preferred)
Nice to Have
· Experience with CSS frameworks (Tailwind, Bootstrap, Sass)
· Familiarity with containerization (Docker) and monitoring tools.
· Contributions to open-source Next.js projects or technical blogs.
About the Role
We are seeking an experienced Next.js Developer with minimum 6 years of hands-on experience to join our development team. The ideal candidate will have a strong background in architecting scalable applications, deploying projects, and writing robust test cases. You will play a key role in the full software development lifecycle, from initial concept and architecture through deployment and maintenance.
Key Responsibilities
· Design, develop, and maintain high-performance, scalable web applications using Next.js
· Architect project structure for maintainability, scalability, and performance, following best practices for folder and component organization
· Lea d and participate in project deployments using modern CI/CD pipelines and cloud platforms. AWS is must
· Write comprehensive unit, integration, and end-to-end test cases using frameworks like Jest, React Testing Library, or Cypress
· Collaborate with cross-functional teams (designers, backend developers, product managers) to deliver seamless user experiences and integrate APIs (REST/GraphQL)
· Optimize applications for SEO, server-side rendering (SSR), static site generation (SSG), and performance
· Conduct code reviews, enforce coding standards, and mentor junior developers
· Troubleshoot, debug, and resolve issues across the stack to ensure reliability and security
· Maintain technical documentation, including architecture decisions, deployment guides, and test plans
Required Skills and Qualifications
· Development experience of 6+ years in Reactjs and Nextjs (preferred)
· Proven expertise in Leading, architecting and organizing large-scale Nextjs projects.
· Hands-on experience with project deployment and cloud platform as AWS (preferred)
· Strong proficiency in JavaScript (ES6+), TypeScript, MaterialUI and CSS Frameworks
· Experience writing and maintaining test cases using Jest, React Testing Library, Cypress, or similar tools
· Familiarity with state management libraries (Redux, Context API) and API integration (REST, GraphQL)
· Working knowledge of version control systems (Git) and CI/CD pipelines (preferred)
· Solid understanding of SSR, SSG, and SEO best practices
· Excellent problem-solving, communication, and teamwork skills
· Bachelor’s degree in Computer Science, Information Technology, or a related field (preferred)
Nice to Have
· Experience with CSS frameworks (Tailwind, Bootstrap, Sass)
· Familiarity with containerization (Docker) and monitoring tools.
· Contributions to open-source Next.js projects or technical blogs.
We’re building a suite of SaaS products for WordPress professionals—each with a clear product-market fit and the potential to become a $100M+ business. As we grow, we need engineers who go beyond feature delivery. We’re looking for someone who wants to build enduring systems, make practical decisions, and help us ship great products with high velocity.
What You’ll Do
- Work with product, design, and support teams to turn real customer problems into thoughtful, scalable solutions.
- Design and build robust backend systems, services, and APIs that prioritize long-term maintainability and performance.
- Use AI-assisted tooling (where appropriate) to explore solution trees, accelerate development, and reduce toil.
- Improve velocity across the team by building reusable tools, abstractions, and internal workflows—not just shipping isolated features.
- Dig into problems deeply—whether it's debugging a performance issue, streamlining a process, or questioning a product assumption.
- Document your decisions clearly and communicate trade-offs with both technical and non-technical stakeholders.
What Makes You a Strong Fit
- You’ve built and maintained real-world software systems, ideally at meaningful scale or complexity.
- You think in systems and second-order effects—not just in ticket-by-ticket outputs.
- You prefer well-reasoned defaults over overengineering.
- You take ownership—not just of code, but of the outcomes it enables.
- You work cleanly, write clear code, and make life easier for those who come after you.
- You’re curious about the why, not just the what—and you’re comfortable contributing to product discussions.
Bonus if You Have Experience With
- Building tools or workflows that accelerate other developers.
- Working with AI coding tools and integrating them meaningfully into your workflow.
- Building for SaaS products, especially those with large user bases or self-serve motions.
- Working in small, fast-moving product teams with a high bar for ownership.
Why Join Us
- A small team that values craftsmanship, curiosity, and momentum.
- A product-driven culture where engineering decisions are informed by customer outcomes.
- A chance to work on multiple zero-to-one opportunities with strong PMF.
- No vanity perks—just meaningful work with people who care.
SENIOR DATA ENGINEER:
ROLE SUMMARY:
Own the design and delivery of petabyte-scale data platforms and pipelines across AWS and modern Lakehouse stacks. You’ll architect, code, test, optimize, and operate ingestion, transformation, storage, and serving layers. This role requires autonomy, strong engineering judgment, and partnership with project managers, infrastructure teams, testers, and customer architects to land secure, cost-efficient, and high-performing solutions.
RESPONSIBILITIES:
- Architecture and design: Create HLD/LLD/SAD, source–target mappings, data contracts, and optimal designs aligned to requirements.
- Pipeline development: Build and test robust ETL/ELT for batch, micro-batch, and streaming across RDBMS, flat files, APIs, and event sources.
- Performance and cost tuning: Profile and optimize jobs, right-size infrastructure, and model license/compute/storage costs.
- Data modeling and storage: Design schemas and SCD strategies; manage relational, NoSQL, data lakes, Delta Lakes, and Lakehouse tables.
- DevOps and release: Establish coding standards, templates, CI/CD, configuration management, and monitored release processes.
- Quality and reliability: Define DQ rules and lineage; implement SLA tracking, failure detection, RCA, and proactive defect mitigation.
- Security and governance: Enforce IAM best practices, retention, audit/compliance; implement PII detection and masking.
- Orchestration: Schedule and govern pipelines with Airflow and serverless event-driven patterns.
- Stakeholder collaboration: Clarify requirements, present design options, conduct demos, and finalize architectures with customer teams.
- Leadership: Mentor engineers, set FAST goals, drive upskilling and certifications, and support module delivery and sprint planning.
REQUIRED QUALIFICATIONS:
- Experience: 15+ years designing distributed systems at petabyte scale; 10+ years building data lakes and multi-source ingestion.
- Cloud (AWS): IAM, VPC, EC2, EKS/ECS, S3, RDS, DMS, Lambda, CloudWatch, CloudFormation, CloudTrail.
- Programming: Python (preferred), PySpark, SQL for analytics, window functions, and performance tuning.
- ETL tools: AWS Glue, Informatica, Databricks, GCP DataProc; orchestration with Airflow.
- Lakehouse/warehousing: Snowflake, BigQuery, Delta Lake/Lakehouse; schema design, partitioning, clustering, performance optimization.
- DevOps/IaC: Terraform with 15+ years of practice; CI/CD (GitHub Actions, Jenkins) with 10+ years; config governance and release management.
- Serverless and events: Design event-driven distributed systems on AWS.
- NoSQL: 2–3 years with DocumentDB including data modeling and performance considerations.
- AI services: AWS Entity Resolution, AWS Comprehend; run custom LLMs on Amazon SageMaker; use LLMs for PII classification.
NICE-TO-HAVE QUALIFICATIONS:
- Data governance automation: 10+ years defining audit, compliance, retention standards and automating governance workflows.
- Table and file formats: Apache Parquet; Apache Iceberg as analytical table format.
- Advanced LLM workflows: RAG and agentic patterns over proprietary data; re-ranking with index/vector store results.
- Multi-cloud exposure: Azure ADF/ADLS, GCP Dataflow/DataProc; FinOps practices for cross-cloud cost control.
OUTCOMES AND MEASURES:
- Engineering excellence: Adherence to processes, standards, and SLAs; reduced defects and non-compliance; fewer recurring issues.
- Efficiency: Faster run times and lower resource consumption with documented cost models and performance baselines.
- Operational reliability: Faster detection, response, and resolution of failures; quick turnaround on production bugs; strong release success.
- Data quality and security: High DQ pass rates, robust lineage, minimal security incidents, and audit readiness.
- Team and customer impact: On-time milestones, clear communication, effective demos, improved satisfaction, and completed certifications/training.
LOCATION AND SCHEDULE:
● Location: Outside US (OUS).
● Schedule: Minimum 6 hours of overlap with US time zones.
Senior Python Django Developer
Experience: Back-end development: 6 years (Required)
Location: Bangalore/ Bhopal
Job Description:
We are looking for a highly skilled Senior Python Django Developer with extensive experience in building and scaling financial or payments-based applications. The ideal candidate has a deep understanding of system design, architecture patterns, and testing best practices, along with a strong grasp of the startup environment.
This role requires a balance of hands-on coding, architectural design, and collaboration across teams to deliver robust and scalable financial products.
Responsibilities:
- Design and develop scalable, secure, and high-performance applications using Python (Django framework).
- Architect system components, define database schemas, and optimize backend services for speed and efficiency.
- Lead and implement design patterns and software architecture best practices.
- Ensure code quality through comprehensive unit testing, integration testing, and participation in code reviews.
- Collaborate closely with Product, DevOps, QA, and Frontend teams to build seamless end-to-end solutions.
- Drive performance improvements, monitor system health, and troubleshoot production issues.
- Apply domain knowledge in payments and finance, including transaction processing, reconciliation, settlements, wallets, UPI, etc.
- Contribute to technical decision-making and mentor junior developers.
Requirements:
- 6 to 10 years of professional backend development experience with Python and Django.
- Strong background in payments/financial systems or FinTech applications.
- Proven experience in designing software architecture in a microservices or modular monolith environment.
- Experience working in fast-paced startup environments with agile practices.
- Proficiency in RESTful APIs, SQL (PostgreSQL/MySQL), NoSQL (MongoDB/Redis).
- Solid understanding of Docker, CI/CD pipelines, and cloud platforms (AWS/GCP/Azure).
- Hands-on experience with test-driven development (TDD) and frameworks like pytest, unittest, or factory_boy.
- Familiarity with security best practices in financial applications (PCI compliance, data encryption, etc.).
Preferred Skills:
- Exposure to event-driven architecture (Celery, Kafka, RabbitMQ).
- Experience integrating with third-party payment gateways, banking APIs, or financial instruments.
- Understanding of DevOps and monitoring tools (Prometheus, ELK, Grafana).
- Contributions to open-source or personal finance-related projects.
Job Types: Full-time, Permanent
Schedule:
- Day shift
Supplemental Pay:
- Performance bonus
- Yearly bonus
Ability to commute/relocate:
- JP Nagar, 5th Phase, Bangalore, Karnataka or Indrapuri, Bhopal, Madhya Pradesh: Reliably commute or willing to relocate with an employer-provided relocation package (Preferred)
Position Overview
We are seeking an experienced Solutions Architect to lead the technical design and implementation strategy for our finance automation platform. This role sits at the intersection of business requirements, technical architecture, and implementation excellence. You will be responsible for translating complex Statement of Work (SOW) requirements into comprehensive technical designs while mentoring implementation engineers and driving platform evolution.
Key Responsibilities
Solution Design & Architecture
1. Translate SOW requirements into detailed C4 architecture models and Business Process Canvas documentation
2. Design end-to-end solutions for complex finance automation workflows including reconciliations, book closure, and financial reporting
3. Create comprehensive technical specifications for custom development initiatives
4. Establish architectural standards and best practices for finance domain solutions
Technical Leadership & Mentorship
1. Mentor Implementation Engineers on solution design, technical approaches, and best practices
2. Lead technical reviews and ensure solution quality across all implementations
3. Provide guidance on complex technical challenges and architectural decisions
4. Foster knowledge sharing and technical excellence within the solutions team
Platform Strategy & Development
1. Make strategic decisions on when to push feature development to the Platform Team vs. custom implementation
2. Interface with Implementation Support team to assess platform gaps and enhancement opportunities
3. Collaborate with Program Managers to track and prioritize new platform feature development
4. Contribute to product roadmap discussions based on client requirements and market trends
Client Engagement & Delivery
1. Lead technical discussions with enterprise clients during pre-sales and implementation phases
2. Design scalable solutions that align with client's existing technology stack and future roadmap
3. Ensure solutions comply with financial regulations (Ind AS/IFRS/GAAP) and industry standards
4. Drive technical aspects of complex implementations from design through go-live
Required Qualifications
Technical Expertise
● 8+ years of experience in solution architecture, preferably in fintech or enterprise software
● Strong expertise in system integration, API design, and microservices architecture
● Proficiency in C4 modeling and architectural documentation standards
● Experience with Business Process Management (BPM) and workflow design
● Advanced knowledge of data architecture, ETL pipelines, and real-time data processing
● Strong programming skills in Python, Java, or similar languages
● Experience with cloud platforms (AWS, Azure, GCP) and containerization technologies.
Financial Domain Knowledge
● Deep understanding of finance and accounting principles (Ind AS/IFRS/GAAP)
● Experience with financial systems integration (ERP, GL, AP/AR systems)
● Knowledge of financial reconciliation processes and automation strategies
● Understanding of regulatory compliance requirements in financial services
Leadership & Communication
● Proven experience mentoring technical teams and driving technical excellence
● Strong stakeholder management skills with ability to communicate with C-level executives
● Experience working in agile environments with cross-functional teams
● Excellent technical documentation and presentation skills
Preferred Qualifications
● Master's degree in Computer Science, Engineering, or related technical field
● Experience with finance automation platforms (Blackline, Trintech, Anaplan, etc.)
● Certification in enterprise architecture frameworks (TOGAF, Zachman)
● Experience with data visualization tools (Power BI, Tableau, Looker)
● Background in SaaS platform development and multi-tenant architectures
● Experience with DevOps practices and CI/CD pipeline design
● Knowledge of machine learning applications in finance automation.
Skills & Competencies
Technical Skills
● Solution architecture and system design
● C4 modeling and architectural documentation
● API design and integration patterns
● Cloud-native architecture and microservices
● Data architecture and pipeline design
● Programming and scripting languages
Financial & Business Skills
● Financial process automation
● Business process modeling and optimization
● Regulatory compliance and risk management
● Enterprise software implementation
● Change management and digital transformation
Leadership Skills
● Technical mentorship and team development
● Strategic thinking and decision making
● Cross-functional collaboration
● Client relationship management
● Project and program management
Soft Skills
● Critical thinking and problem-solving
● Cross-functional collaboration
● Task and project management
● Stakeholder management
● Team leadership
● Technical documentation
● Communication with technical and non-technical stakeholders
Mandatory Criteria:
● Looking for candidates who are Solution Architects in Finance from Product Companies.
● The candidate should have worked in Fintech for at least 4–5 years.
● Candidate should have Strong Technical and Architecture skills with Finance Exposure.
● Candidate should be from Product companies.
● Candidate should have 8+ years’ experience in solution architecture, preferably in fintech or enterprise software.
● Candidate should have Proficiency in Python, Java (or similar languages) and hands-on with cloud platforms (AWS/Azure/GCP) & containerization (Docker/Kubernetes).
● Candidate should have Deep knowledge of finance & accounting principles (Ind AS/IFRS/GAAP) and financial system integrations (ERP, GL, AP/AR).
● Candidate should have Expertise in system integration, API design, microservices, and C4 modeling.
● Candidate should have Experience in financial reconciliations, automation strategies, and regulatory compliance.
● Candidate should be Strong in problem-solving, cross-functional collaboration, project management, documentation, and communication.
● Candidate should have Proven experience in mentoring technical teams and driving excellence.
Job Title: PySpark/Scala Developer
Functional Skills: Experience in Credit Risk/Regulatory risk domain
Technical Skills: Spark ,PySpark, Python, Hive, Scala, MapReduce, Unix shell scripting
Good to Have Skills: Exposure to Machine Learning Techniques
Job Description:
5+ Years of experience with Developing/Fine tuning and implementing programs/applications
Using Python/PySpark/Scala on Big Data/Hadoop Platform.
Roles and Responsibilities:
a) Work with a Leading Bank’s Risk Management team on specific projects/requirements pertaining to risk Models in
consumer and wholesale banking
b) Enhance Machine Learning Models using PySpark or Scala
c) Work with Data Scientists to Build ML Models based on Business Requirements and Follow ML Cycle to Deploy them all
the way to Production Environment
d) Participate Feature Engineering, Training Models, Scoring and retraining
e) Architect Data Pipeline and Automate Data Ingestion and Model Jobs
Skills and competencies:
Required:
· Strong analytical skills in conducting sophisticated statistical analysis using bureau/vendor data, customer performance
Data and macro-economic data to solve business problems.
· Working experience in languages PySpark & Scala to develop code to validate and implement models and codes in
Credit Risk/Banking
· Experience with distributed systems such as Hadoop/MapReduce, Spark, streaming data processing, cloud architecture.
- Familiarity with machine learning frameworks and libraries (like scikit-learn, SparkML, tensorflow, pytorch etc.
- Experience in systems integration, web services, batch processing
- Experience in migrating codes to PySpark/Scala is big Plus
- The ability to act as liaison conveying information needs of the business to IT and data constraints to the business
applies equal conveyance regarding business strategy and IT strategy, business processes and work flow
· Flexibility in approach and thought process
· Attitude to learn and comprehend the periodical changes in the regulatory requirement as per FED
About the Role
We are looking for a hands-on and solution-oriented Senior Data Scientist – Generative AI to join our growing AI practice. This role is ideal for someone who thrives in designing and deploying Gen AI solutions on AWS, enjoys working with customers directly, and can lead end-to-end implementations. You will play a key role in architecting AI solutions, driving project delivery, and guiding junior team members.
Key Responsibilities
- Design and implement end-to-end Generative AI solutions for customers on AWS.
- Work closely with customers to understand business challenges and translate them into Gen AI use-cases.
- Own technical delivery, including data preparation, model integration, prompt engineering, deployment, and performance monitoring.
- Lead project execution – ensure timelines, manage stakeholder communications, and collaborate across internal teams.
- Provide technical guidance and mentorship to junior data scientists and engineers.
- Develop reusable components and reference architectures to accelerate delivery.
- Stay updated with latest developments in Gen AI, particularly AWS offerings like Bedrock, SageMaker, LangChain integrations, etc.
Required Skills & Experience
- 4–8 years of hands-on experience in Data Science/AI/ML, with at least 2–3 years in Generative AI projects.
- Proficient in building solutions using AWS AI/ML services (e.g., SageMaker, Amazon Bedrock, Lambda, API Gateway, S3, etc.).
- Experience with LLMs, prompt engineering, RAG pipelines, and deployment best practices.
- Solid programming experience in Python, with exposure to libraries such as Hugging Face, LangChain, etc.
- Strong problem-solving skills and ability to work independently in customer-facing roles.
- Experience in collaborating with Systems Integrators (SIs) or working with startups in India is a major plus.
Soft Skills
- Strong verbal and written communication for effective customer engagement.
- Ability to lead discussions, manage project milestones, and coordinate across stakeholders.
- Team-oriented with a proactive attitude and strong ownership mindset.
What We Offer
- Opportunity to work on cutting-edge Generative AI projects across industries.
- Collaborative, startup-like work environment with flexibility and ownership.
- Exposure to full-stack AI/ML project lifecycle and client-facing roles.
- Competitive compensation and learning opportunities in the AWS AI ecosystem.
About Oneture Technologies
Founded in 2016, Oneture is a cloud-first, full-service digital solutions company, helping clients harness the power of Digital Technologies and Data to drive transformations and turning ideas into business realities. Our team is full of curious, full-stack, innovative thought leaders who are dedicated to providing outstanding customer experiences and building authentic relationships. We are compelled by our core values to drive transformational results from Ideas to Reality for clients across all company sizes, geographies, and industries. The Oneture team delivers full lifecycle solutions— from ideation, project inception, planning through deployment to ongoing support and maintenance.
Our core competencies and technical expertise includes Cloud powered: Product Engineering, Big Data and AI ML. Our deep commitment to value creation for our clients and partners and “Startups-like agility with Enterprises-like maturity” philosophy has helped us establish long-term relationships with our clients and enabled us to build and manage mission-critical platforms for them.


























