
What you’ll be doing
Weare much more than our job descriptions, but here is where you will begin:
As a Senior Software Engineer Data & ML You’ll Be:
● Architect, design, test, implement, deploy, monitor and maintain end-to-end backend
services. You build it, you own it.
● Work with people from other teams and departments on a day to day basis to ensure
efficient project execution with a focus on delivering value to our members.
● Regularly aligning your team’s vision and roadmap with the target architecture within your
domain and to ensure the success of complex multi domain initiatives.
● Integrate already trained ML and GenAI models (preferably GCP in services.
ROLE:
Whatyou’ll need,
Like us, you’ll be deeply committed to delivering impactful outcomes for customers.
What Makes You a Great Fit
● 5 years of proven work experience as a Backend Python Engineer
● Understanding of software engineering fundamentals OOPS, SOLID, etc.)
● Hands-on experience with Python libraries like Pandas, NumPy, Scikit-learn,
Lang chain/LLamaIndex etc.
● Experience with machine learning frameworks such as PyTorch or TensorFlow, Keras, being
proficient in Python
● Hands-on Experience with frameworks such as Django or FastAPI or Flask
● Hands-on experience with MySQL, MongoDB, Redis and BigQuery (or equivalents)
● Extensive experience integrating with or creating REST APIs
● Experience with creating and maintaining CI/CD pipelines- we use GitHub Actions.
● Experience with event-driven architectures like Kafka, RabbitMq or equivalents.
● Knowledge about:
o LLMs
o Vector stores/databases
o PromptEngineering
o Embeddings and their implementations
● Somehands-onexperience in implementations of the above ML/AI will be preferred
● Experience with GCP/AWS services.
● You are curious about and motivated by the future trends in data, AI/ML, analytics

About Recro
About
Recro is a developer-focused platform that was founded with the aim of seamlessly matching individual expertise with the right opportunities.
We empower talented developers by providing them with relevant experience at fast-growing startups based on technical competencies and aspirations. These opportunities have a significant impact on their career success and help them become their best self.
On the other hand, startups get instant access to top-quality developers with guaranteed productivity from the very beginning. We help them to scale up/down based on their needs, thus ensuring an efficient and high-yielding workforce.
Developers solve real-time complex problems and get exposure to the uplifting and challenging work culture at start-ups like Flipkart, Dunzo, Swiggy, and Zivame among many others. At Recro, we ensure continuous support from our strong community to accelerate careers for developers and strive to create optimal business outcomes for high-growth startups.
Company video


Photos
Connect with the team
Similar jobs
Job description
Job Overview:
The position requires an experienced and ambitious candidate who is passionate about technology and is self-driven. We have a challenging workplace where we welcome innovative ideas and offers growth opportunities and positive environment for accomplishing goals. Our purpose is to create abundance for everyone we touch.
Job Description:
- Experience on Open Source Platforms in designing/ developing Web-based applications.
- Require Strong knowledge in Python with application/package/module development, tuning, and debugging tools.
- Proficient understanding of Python Identifiers, Reserved Words, Basic Operators, Variable Types, and User defined exception handling with their usage.
- In-depth knowledge of Python Dictionary and default modules included in Python (String, DateTime, Numbers, and other required functions usage).
- Proficient understanding of Sequences and its differences (Tuples/Lists).
- File manipulation-using Python (Create, Edit, and Delete view file).
- Ability to organize code logically and with the understandable procedure that is well defined, documented, and testable.
Skills Required:
- Strong understanding of OOPS and Design Patterns (Code design skills in Python-based object-oriented programming and functional programming).
- Developing Web Applications with Python - Hands on experience using MVC frameworks like Django.
- Exposure to Code Versioning Systems such as Atlassian Bitbucket.
- In-depth Knowledge in JavaScript and jQuery.
- The candidate must display excellent written and verbal skills with demonstrated interpersonal and organizational abilities.

Role: Python Developer (Immediate Joiner)
Location: Gurugram, India (Onsite)
Experience: 4+ years
Company: Mizzle Cloud Pvt Ltd
Working Days: 6 Days( 5 days- WFO, Sat- WFH)
Job Summary
We are seeking a skilled Python Django Developer with expertise in building robust, scalable, and efficient web applications. Must have 3+ years of core work experience. The ideal candidate will have hands-on experience with RabbitMQ, Redis, Celery, and PostgreSQL to ensure seamless background task management, caching, and database performance.
Key Responsibilities
- Develop, maintain, and enhance Django-based web applications and APIs.
- Design and implement message broker solutions using RabbitMQ to manage asynchronous communication.
- Integrate Redis for caching and session storage to optimize performance.
- Implement and manage task queues using Celery for background processes.
- Work with PostgreSQL for database design, optimization, and query tuning.
- Collaborate with front-end developers, DevOps engineers, and stakeholders to deliver high-quality software solutions.
- Write clean, modular, and well-documented code following best practices and standards.
- Debug, troubleshoot, and resolve issues across the application stack.
- Participate in code reviews, system design discussions, and team meetings.
- Ensure scalability, reliability, and security of applications.
Technical Skills:
- Must have minimum 4+ years of relevant work experience.
- Strong proficiency in Python and Django framework.
- Experience with message brokers, particularly RabbitMQ.
- Familiarity with Redis for caching and session management.
- Hands-on experience with Celery for distributed task queues.
- Proficiency in PostgreSQL, including database design and optimization.
- Knowledge of RESTful API design and development.
- Understanding of Docker and containerized applications.
Preferred Skills:
- Experience with CI/CD pipelines.
- Familiarity with cloud platforms (AWS, GCP).
- Knowledge of Django ORM and its optimizations.
- Basic understanding of front-end technologies (HTML, CSS, JavaScript).
Soft Skills:
- Strong problem-solving skills and attention to detail.
- Excellent communication and teamwork abilities.
- Ability to work in an agile environment and adapt to changing requirements.
Educational Requirements
- Bachelor’s degree in Computer Science, Software Engineering, or a related field.
Job Description:
Candidate must have 2 to 5-years of experience in various phase of development of a python-based application or API development.
Requirements:
Must Have: Strong expertise in PYTHON and its built-in data structures, developing API using flask or fastapi, data wrangling using standard python frameworks like pandas, NumPy etc., integration of various applications (third party or inhouse) with python.
Good to Have: Deployment using Nginx, GUnicorn, IIS, docker and Kubernetes etc. Good understanding of optimization solutions like differential evolution. Hands on experience with python packages like Scipy. Tkinter to make python-based applications.
Roles and Responsibilities:
- Understanding/gathering the requirements from stakeholder, formulating the problem, and leveraging AI/ML to solve the business problems.
- Integration of third party or in house application with python solution.
- Working on Deployment and optimization of various engineering problems using Numerical/Constraint optimization frameworks such as Differential Evolution and deploy a web-based API using Flask and Nginx/IIS.
- Unit testing of various python modules developed and testing of the API.
Job Title: AI/ML Engineer – Voice (2–3 Years)
Location: Bengaluru (On-site)
Employment Type: Full-time
About Impacto Digifin Technologies
Impacto Digifin Technologies enables enterprises to adopt digital transformation through intelligent, AI-powered solutions. Our platforms reduce manual work, improve accuracy, automate complex workflows, and ensure compliance—empowering organizations to operate with speed, clarity, and confidence.
We combine automation where it’s fastest with human oversight where it matters most. This hybrid approach ensures trust, reliability, and measurable efficiency across fintech and enterprise operations.
Role Overview
We are looking for an AI Engineer Voice with strong applied experience in machine learning, deep learning, NLP, GenAI, and full-stack voice AI systems.
This role requires someone who can design, build, deploy, and optimize end-to-end voice AI pipelines, including speech-to-text, text-to-speech, real-time streaming voice interactions, voice-enabled AI applications, and voice-to-LLM integrations.
You will work across core ML/DL systems, voice models, predictive analytics, banking-domain AI applications, and emerging AGI-aligned frameworks. The ideal candidate is an applied engineer with strong fundamentals, the ability to prototype quickly, and the maturity to contribute to R&D when needed.
This role is collaborative, cross-functional, and hands-on.
Key Responsibilities
Voice AI Engineering
- Build end-to-end voice AI systems, including STT, TTS, VAD, audio processing, and conversational voice pipelines.
- Implement real-time voice pipelines involving streaming interactions with LLMs and AI agents.
- Design and integrate voice calling workflows, bi-directional audio streaming, and voice-based user interactions.
- Develop voice-enabled applications, voice chat systems, and voice-to-AI integrations for enterprise workflows.
- Build and optimize audio preprocessing layers (noise reduction, segmentation, normalization)
- Implement voice understanding modules, speech intent extraction, and context tracking.
Machine Learning & Deep Learning
- Build, deploy, and optimize ML and DL models for prediction, classification, and automation use cases.
- Train and fine-tune neural networks for text, speech, and multimodal tasks.
- Build traditional ML systems where needed (statistical, rule-based, hybrid systems).
- Perform feature engineering, model evaluation, retraining, and continuous learning cycles.
NLP, LLMs & GenAI
- Implement NLP pipelines including tokenization, NER, intent, embeddings, and semantic classification.
- Work with LLM architectures for text + voice workflows
- Build GenAI-based workflows and integrate models into production systems.
- Implement RAG pipelines and agent-based systems for complex automation.
Fintech & Banking AI
- Work on AI-driven features related to banking, financial risk, compliance automation, fraud patterns, and customer intelligence.
- Understand fintech data structures and constraints while designing AI models.
Engineering, Deployment & Collaboration
- Deploy models on cloud or on-prem (AWS / Azure / GCP / internal infra).
- Build robust APIs and services for voice and ML-based functionalities.
- Collaborate with data engineers, backend developers, and business teams to deliver end-to-end AI solutions.
- Document systems and contribute to internal knowledge bases and R&D.
Security & Compliance
- Follow fundamental best practices for AI security, access control, and safe data handling.
- Awareness of financial compliance standards (plus, not mandatory).
- Follow internal guidelines on PII, audio data, and model privacy.
Primary Skills (Must-Have)
Core AI
- Machine Learning fundamentals
- Deep Learning architectures
- NLP pipelines and transformers
- LLM usage and integration
- GenAI development
- Voice AI (STT, TTS, VAD, real-time pipelines)
- Audio processing fundamentals
- Model building, tuning, and retraining
- RAG systems
- AI Agents (orchestration, multi-step reasoning)
Voice Engineering
- End-to-end voice application development
- Voice calling & telephony integration (framework-agnostic)
- Realtime STT ↔ LLM ↔ TTS interactive flows
- Voice chat system development
- Voice-to-AI model integration for automation
Fintech/Banking Awareness
- High-level understanding of fintech and banking AI use cases
- Data patterns in core banking analytics (advantageous)
Programming & Engineering
- Python (strong competency)
- Cloud deployment understanding (AWS/Azure/GCP)
- API development
- Data processing & pipeline creation
Secondary Skills (Good to Have)
- MLOps & CI/CD for ML systems
- Vector databases
- Prompt engineering
- Model monitoring & evaluation frameworks
- Microservices experience
- Basic UI integration understanding for voice/chat
- Research reading & benchmarking ability
Qualifications
- 2–3 years of practical experience in AI/ML/DL engineering.
- Bachelor’s/Master’s degree in CS, AI, Data Science, or related fields.
- Proven hands-on experience building ML/DL/voice pipelines.
- Experience in fintech or data-intensive domains preferred.
Soft Skills
- Clear communication and requirement understanding
- Curiosity and research mindset
- Self-driven problem solving
- Ability to collaborate cross-functionally
- Strong ownership and delivery discipline
- Ability to explain complex AI concepts simply
SDE 2 / SDE 3 – AI Infrastructure & LLM Systems Engineer
Location: Pune / Bangalore (India)
Experience: 4–8 years
Compensation: no bar for the right candidate
Bonus: Up to 10% of base
About the Company
AbleCredit builds production-grade AI systems for BFSI enterprises, reducing OPEX by up to 70% across onboarding, credit, collections, and claims.
We run our own LLMs on GPUs, operate high-concurrency inference systems, and build AI workflows that must scale reliably under real enterprise traffic.
Role Summary (What We’re Really Hiring For)
We are looking for a strong backend / systems engineer who can:
- Deploy AI models on GPUs
- Expose them via APIs
- Scale inference under high parallel load using async systems and queues
This is not a prompt-engineering or UI-AI role.
Core Responsibilities
- Deploy and operate LLMs on GPU infrastructure (cloud or on-prem).
- Run inference servers such as vLLM / TGI / SGLang / Triton or equivalents.
- Build FastAPI / gRPC APIs on top of AI models.
- Design async, queue-based execution for AI workflows (fan-out, retries, backpressure).
- Plan and reason about capacity & scaling:
- GPU count vs RPS
- batching vs latency
- cost vs throughput
- Add observability around latency, GPU usage, queue depth, failures.
- Work closely with AI researchers to productionize models safely.
Must-Have Skills
- Strong backend engineering fundamentals (distributed systems, async workflows).
- Hands-on experience running GPU workloads in production.
- Proficiency in Python (Golang acceptable).
- Experience with Docker + Kubernetes (or equivalent).
- Practical knowledge of queues / workers (Redis, Kafka, SQS, Celery, Temporal, etc.).
- Ability to reason quantitatively about performance, reliability, and cost.
Strong Signals (Recruiter Screening Clues)
Look for candidates who have:
- Personally deployed models on GPUs
- Debugged GPU memory / latency / throughput issues
- Scaled compute-heavy backends under load
- Designed async systems instead of blocking APIs
Nice to Have
- Familiarity with LangChain / LlamaIndex (as infra layers, not just usage).
- Experience with vector DBs (Qdrant, Pinecone, Weaviate).
- Prior work on multi-tenant enterprise systems.
Not a Fit If
- Only experience is calling OpenAI / Anthropic APIs.
- Primarily a prompt engineer or frontend-focused AI dev.
- No hands-on ownership of infra, scaling, or production reliability.
What we want to accomplish and why we need you?
Jio Haptik is an AI leader having pioneered AI-powered innovation since 2013. Reliance Jio Digital Services acquired Haptik in April 2019. Haptik currently leads India’s AI market having become the first to process 15 billion+ two-way conversations across 10+ channels and in 135 languages. Haptik is also a Category Leader across platforms including Gartner, G2, Opus Research & more. Recently Haptik won the award for “Tech Startup of the Year” in the AI category at Entrepreneur India Awards 2023, and gold medal for “Best Chat & Conversational Bot” at Martequity Awards 2023. Haptik has a headcount of 200+ employees with offices in Mumbai, Delhi, and Bangalore.
What will you do everyday?
As a backend engineer you will be responsible for building the Haptik platform which is used by people across the globe. You will be responsible for developing, architecting and scaling the systems that support all the functions of the Haptik platform. While you know how to work hard, you also know how to have fun at work and make friends with your colleagues.
Ok, you're sold, but what are we looking for in the perfect candidate?
Develop and maintain expertise in backend systems and API development, ensuring seamless integrations and scalable solutions, including:
- Strong expertise in backend systems, including design principles and adherence to good coding practices.
- Proven ability to enhance or develop complex tools at scale with a thorough understanding of system architecture.
- Capability to work cross-functionally with all teams, ensuring seamless implementation of APIs and solutioning for various tools.
- Skilled in high-level task estimation, scoping, and breaking down complex projects into actionable tasks.
- Proficiency in modeling and optimizing database architecture for enhanced performance and scalability.
- Experience collaborating with product teams to build innovative Proof of Concepts (POCs).
- Ability to respond to data requests and generate reports to support data-driven decision-making.
- Active participation in code reviews, automated testing, and quality assurance processes.
- Experience working in a scrum-based agile development environment.
- Commitment to staying updated with technology standards, emerging trends, and software development best practices.
- Strong verbal and written communication skills to facilitate collaboration and clarity.
Requirements*:
- A minimum of 5 years of experience in developing scalable products and applications.
- Must Have Bachelor's degree in Computer Engineering or related field.
- Proficiency in Python and expertise in at least one backend framework, with a preference for Django.
- Hands-on experience designing normalized database schemas for large-scale applications using technologies such as MySQL, MongoDB, or Elasticsearch.
- Practical knowledge of in-memory data stores like Redis or Memcached.
- Familiarity with working in agile environments and exposure to tools like Jira is highly desirable.
- Proficiency in using version control systems like Git.
- Strong communication skills and the ability to collaborate effectively in team settings.
- Self-motivated with a strong sense of ownership and commitment to delivering results.
- Additional knowledge of RabbitMQ, AWS/Azure services, Docker, MQTT, Lambda functions, Cron jobs, Kibana, and Logstash is an added advantage.
- Knowledge of web servers like Nginx/Apache is considered a valuable asset.
* Requirements is such a strong word. We don’t necessarily expect to find a candidate that has done everything listed, but you should be able to make a credible case that you’ve done most of it and are ready for the challenge of adding some new things to your resume.
Tell me more about Haptik
- On a roll: Announced major strategic partnership with Jio.
- Great team: You will be working with great leaders who have been listed in Business World 40 Under 40, Forbes 30 Under 30 and MIT 35 Under 35 Innovators.
- Great culture: The freedom to think and innovate is something that defines the culture of Haptik. Every person is approachable. While we are working hard, it is also important to take breaks to not get too worked up.
- Huge market: Disrupting a massive, growing chatbot market. The global market is projected to attain a valuation of US $0.94 bn by the end of 2024 progressing from US $0.11 bn earned in 2015.
- Great customers: Businesses across industries - Samsung, HDFCLife, Times of India are some that have relied on Haptik's Conversational AI solutions to engage, acquire, service and understand customers.
- Impact: A fun and exciting start-up culture that empowers its people to make a huge impact.
Working hard for things that we don't care about is stress, but working hard for something we love is called passion! At Haptik we passionately solve problems in order to be able to move faster and don't shy away from breaking things!
Role
You will develop and maintain the key backend code and infrastructure of the company stack. You will implement AI solutions like LLMs for various tasks such as voice-based interactive systems, chatbots, and AI web apps. Ability to see projects through from start to finish with good organizational skills and attention to detail. This is a perfect role for someone who likes to build state-of-the-art AI products and work with cutting-edge AI technologies like GPT, LLAMA, etc
Qualifications
- BS or MS in Computer Science or relevant field.
- 4+ years experience in backend software development
- Be able to design high-throughput scalable backend systems
- Eagerness to learn applied AI technologies like LLMs, prompt engineering, etc
- Proficiency in Python.
- Experience with cloud computing platforms (AWS, GCP) and technologies like Docker
- Knowledge of Rest APIs, databases (mysql, mongo, vectorDB)
We are having an immediate requirement for a Python web developer.
You have:
- At least 2 years of experience developing web applications with Django/Flask/FastAPI
- Familiarity with Linux
- Experience in both SQL and NoSQL databases.
- Uses Docker and CI/CD
- Writes tests
- Experienced in application deployments and scaling them on AWS or GCP
You are:
- Eager to work independently without being watched
- Easy going.
- Able to handle clients on your own
Location: Remote (in India)
We are looking to hire a Full-time, Remote (India) Backend engineer with a flair for writing full stack code to help create a SAAS product from the ground up in a niche industry.
This will be the v2.0 of an existing legacy platform with paying customers. This position is within the first few hires in the founding team of the revamped company.
We are on a mission to build a truly world class product and deliver at high speeds with high quality.
Read on and apply for the job if you are a doer, likes no BS, think you're competent enough to get things done, wants some quiet time every day, wants to make a difference, values work life balance and enjoy life outside work.
Experience -
At least 3 to 4 years building scalable SAAS applications in the cloud. It would be great to showcase anything progressive. Education from premier institutions, git contributions, tech blogs, volunteering, pet projects, etc.
Responsibilities -
- Architect, write, deploy and own the (SAAS) product from a technical standpoint working closely with other devs on the team
- Create and deliver the product end-to-end all the way to the user
- Write highly performant code and continuously measure performance and make the application better based on benchmarks
- Debug production issues and keep the application quality high always
- Assist with hiring more people when the team grows
- Pick up any appropriate work as in any early-stage company
General Skills Needed -
- Should be able to build and deploy multi-tenant SAAS applications end to end
- Need to have a technology-agnostic mindset. Should be able to easily pick up any new technology based on the use case.
- Should be an expert software craftsman/craftswoman. Must own the code and be proud of what you build.
- Write well thought out, well-tested code with extensive unit tests and integration tests so that the addition of future features is easy.
- Must know cloud deployment concepts and be comfortable creating and maintaining cost-effective cloud deployment strategies. AWS experience highly preferred.
- Must be able to design and architect simple, market-tested yet highly scalable solutions
- Must be able to quickly iterate and produce working software at a high speed. Must not be shy to scrap and rebuild when there is a need
- Must be appreciative of documentation and write well-documented code and technical documentation alongside
Technical Skills Needed -
- Expert level knowledge in Python and Django. We are a Python/Django/Javascript/ReactJs/AWS shop.
- Must have good knowledge/experience in creating cost-efficient and scalable cloud deployments (AWS preferred)
- Must have good knowledge of industry-standard design patterns and tools
- Must have a good understanding of various frameworks on authentication and authorization, billing and metrics
- Some experience in data analytics and reporting - creating reports based on the data collected and deliver to the frontend
- Expertise with Frameworks
- Django, or similar backend frameworks
- PostgreSQL, MySQL, any NoSQL or other database technology
Job Perks
Perks -
- Opportunity to build a SAAS product from the ground up. Lots of challenges to tackle on
- Work remotely from anywhere within India
- Emphasis on work-life balance and professional development
- Flexible work hours and a lot of autonomy
- Work with a very informal team collaborating over slack. No meaningless meetings. Focus on getting things done and not hours put in
- Opportunity to grow the team with the company
- Market salary and a yearly bonus outside the salary based on company + individual performance
- Generous PTO plan
Role - Backend Developer
Experience:- 2+
Qualification:- BE Computer Engineering/MCA
Programming Language - Python & framework Django
Responsibilities
- Building REST API's & Services In Django Framework
- Building reusable code and libraries for future use
- Optimization of the application for maximum speed and scalability
- Implementation of security and data protection
- Design and implementation of the database schema
- Design and implementation of data storage solutions
- Implementing CI/CD pipeline
- Proficient knowledge of a back-end programming language Python
- Hands-on experience with Python Frameworks like Django
- Proficient knowledge of MySQL, PostgreSQL
- Creating database schemas that represent and support business processes (Relational & NoSQL)
- Understanding of queueing systems like Redis/AWS SQS
- User authentication and authorization between multiple systems, servers, and environments
- Data migration, transformation, and scripting
- Understanding differences between multiple delivery platforms such as mobile vs desktop, and optimizing output to match the specific platform
- Proficient understanding of code versioning tools, such as Git
Brownie Points
- Experience in Docker
- HandsOn Knowledge with implementing CI/CD pipelines
- Experience in managing applications on AWS
- Management of hosting environment, including database administration and scaling application to support load changes
- Product wide application-level thinking on API and data modeling














