
Job Title: AI/ML Engineer – Voice (2–3 Years)
Location: Bengaluru (On-site)
Employment Type: Full-time
About Impacto Digifin Technologies
Impacto Digifin Technologies enables enterprises to adopt digital transformation through intelligent, AI-powered solutions. Our platforms reduce manual work, improve accuracy, automate complex workflows, and ensure compliance—empowering organizations to operate with speed, clarity, and confidence.
We combine automation where it’s fastest with human oversight where it matters most. This hybrid approach ensures trust, reliability, and measurable efficiency across fintech and enterprise operations.
Role Overview
We are looking for an AI Engineer Voice with strong applied experience in machine learning, deep learning, NLP, GenAI, and full-stack voice AI systems.
This role requires someone who can design, build, deploy, and optimize end-to-end voice AI pipelines, including speech-to-text, text-to-speech, real-time streaming voice interactions, voice-enabled AI applications, and voice-to-LLM integrations.
You will work across core ML/DL systems, voice models, predictive analytics, banking-domain AI applications, and emerging AGI-aligned frameworks. The ideal candidate is an applied engineer with strong fundamentals, the ability to prototype quickly, and the maturity to contribute to R&D when needed.
This role is collaborative, cross-functional, and hands-on.
Key Responsibilities
Voice AI Engineering
- Build end-to-end voice AI systems, including STT, TTS, VAD, audio processing, and conversational voice pipelines.
- Implement real-time voice pipelines involving streaming interactions with LLMs and AI agents.
- Design and integrate voice calling workflows, bi-directional audio streaming, and voice-based user interactions.
- Develop voice-enabled applications, voice chat systems, and voice-to-AI integrations for enterprise workflows.
- Build and optimize audio preprocessing layers (noise reduction, segmentation, normalization)
- Implement voice understanding modules, speech intent extraction, and context tracking.
Machine Learning & Deep Learning
- Build, deploy, and optimize ML and DL models for prediction, classification, and automation use cases.
- Train and fine-tune neural networks for text, speech, and multimodal tasks.
- Build traditional ML systems where needed (statistical, rule-based, hybrid systems).
- Perform feature engineering, model evaluation, retraining, and continuous learning cycles.
NLP, LLMs & GenAI
- Implement NLP pipelines including tokenization, NER, intent, embeddings, and semantic classification.
- Work with LLM architectures for text + voice workflows
- Build GenAI-based workflows and integrate models into production systems.
- Implement RAG pipelines and agent-based systems for complex automation.
Fintech & Banking AI
- Work on AI-driven features related to banking, financial risk, compliance automation, fraud patterns, and customer intelligence.
- Understand fintech data structures and constraints while designing AI models.
Engineering, Deployment & Collaboration
- Deploy models on cloud or on-prem (AWS / Azure / GCP / internal infra).
- Build robust APIs and services for voice and ML-based functionalities.
- Collaborate with data engineers, backend developers, and business teams to deliver end-to-end AI solutions.
- Document systems and contribute to internal knowledge bases and R&D.
Security & Compliance
- Follow fundamental best practices for AI security, access control, and safe data handling.
- Awareness of financial compliance standards (plus, not mandatory).
- Follow internal guidelines on PII, audio data, and model privacy.
Primary Skills (Must-Have)
Core AI
- Machine Learning fundamentals
- Deep Learning architectures
- NLP pipelines and transformers
- LLM usage and integration
- GenAI development
- Voice AI (STT, TTS, VAD, real-time pipelines)
- Audio processing fundamentals
- Model building, tuning, and retraining
- RAG systems
- AI Agents (orchestration, multi-step reasoning)
Voice Engineering
- End-to-end voice application development
- Voice calling & telephony integration (framework-agnostic)
- Realtime STT ↔ LLM ↔ TTS interactive flows
- Voice chat system development
- Voice-to-AI model integration for automation
Fintech/Banking Awareness
- High-level understanding of fintech and banking AI use cases
- Data patterns in core banking analytics (advantageous)
Programming & Engineering
- Python (strong competency)
- Cloud deployment understanding (AWS/Azure/GCP)
- API development
- Data processing & pipeline creation
Secondary Skills (Good to Have)
- MLOps & CI/CD for ML systems
- Vector databases
- Prompt engineering
- Model monitoring & evaluation frameworks
- Microservices experience
- Basic UI integration understanding for voice/chat
- Research reading & benchmarking ability
Qualifications
- 2–3 years of practical experience in AI/ML/DL engineering.
- Bachelor’s/Master’s degree in CS, AI, Data Science, or related fields.
- Proven hands-on experience building ML/DL/voice pipelines.
- Experience in fintech or data-intensive domains preferred.
Soft Skills
- Clear communication and requirement understanding
- Curiosity and research mindset
- Self-driven problem solving
- Ability to collaborate cross-functionally
- Strong ownership and delivery discipline
- Ability to explain complex AI concepts simply

About Impacto Digifin Technologies
About
Impacto Digifin Technologies empowers businesses to embrace digital transformation with intelligent, AI-driven solutions. Our platforms simplify document management, data verification, and compliance processes, reducing manual effort, enhancing accuracy, and accelerating results. From fast-growing fintechs to established enterprises, our solutions are designed to adapt to unique operational needs, whether it’s streamlining customer onboarding, automating back-office workflows, or eliminating paperwork bottlenecks.
What sets Impacto Digifin apart is our hybrid approach—leveraging automation for speed while maintaining human oversight where it matters most. This ensures efficiency without compromising trust, enabling organizations to operate with clarity and control. More than just a technology provider, we act as a digital partner, helping teams scale smarter, optimize processes, and transform their operations with confidence.
Candid answers by the company
Impacto Digifin Technologies provides AI-powered solutions that streamline document management, data verification, and compliance for businesses.
Connect with the team
Similar jobs
Designation: Python Developer
Experienced in AI/ML
Location: Turbhe, Navi Mumbai
CTC: 6-12 LPA
Years of Experience: 2-5 years
At Arcitech.ai, we’re redefining the future with AI-powered software solutions across education, recruitment, marketplaces, and beyond. We’re looking for a Python Developer passionate about AI/ML, who’s ready to work on scalable, cloud-native platforms and help build the next generation of intelligent, LLM-driven products.
💼 Your Responsibilities
AI/ML Engineering
- Develop, train, and optimize ML models using PyTorch/TensorFlow/Keras.
- Build end-to-end LLM and RAG (Retrieval-Augmented Generation) pipelines using LangChain.
- Collaborate with data scientists to convert prototypes into production-grade AI applications.
- Integrate NLP, Computer Vision, and Recommendation Systems into scalable products.
- Work with transformer-based architectures (BERT, GPT, LLaMA, etc.) for real-world AI use cases.
Backend & Systems Development
- Design, develop, and maintain robust Python microservices with REST/GraphQL APIs.
- Implement real-time communication with Django Channels/WebSockets.
- Containerize AI services with Docker and deploy on Kubernetes (EKS/GKE/AKS).
- Configure and manage AWS (EC2, S3, RDS, SageMaker, CloudWatch) for AI/ML workloads.
Reliability & Automation
- Develop background task queues with Celery, ensuring smart retries and monitoring.
- Implement CI/CD pipelines for automated model training, testing, and deployment.
- Write automated unit & integration tests (pytest/unittest) with ≥80% coverage.
Collaboration
- Contribute to MLOps best practices and mentor peers in LangChain/AI integration.
- Participate in tech talks, code reviews, and AI learning sessions within the team.
🎓 Required Qualifications
- Bachelor’s or Master’s degree in Computer Science, AI/ML, or related field.
- 2–5 years of experience in Python development with strong AI/ML exposure.
- Hands-on experience with LangChain for building LLM-powered workflows and RAG systems.
- Deep learning experience with PyTorch or TensorFlow.
- Experience deploying ML models and LLM apps into production systems.
- Familiarity with REST/GraphQL APIs and cloud platforms (AWS/Azure/GCP).
- Skilled in Git workflows, automated testing, and CI/CD practices.
🌟 Nice to Have
- Experience with vector databases (Pinecone, Weaviate, FAISS, Milvus) for retrieval pipelines.
- Knowledge of LLM fine-tuning, prompt engineering, and evaluation frameworks.
- Familiarity with Airflow/Prefect/Dagster for data and model pipelines.
- Background in statistics, optimization, or applied mathematics.
- Contributions to AI/ML or LangChain open-source projects.
- Experience with model monitoring and drift detection in production.
🎁 Why Join Us
- Competitive compensation and benefits 💰
- Work on cutting-edge LLM and AI/ML applications 🤖
- A collaborative, innovation-driven work culture 📚
- Opportunities to grow into AI/ML leadership roles 🚀
We're looking for AI/ML enthusiasts who build, not just study. If you've implemented transformers from scratch, fine-tuned LLMs, or created innovative ML solutions, we want to see your work!
What You’ll Do
-Build autonomous AI agents using LangChain, LangGraph, and similar frameworks.
- Develop RAG pipelines with vector DBs like FAISS, Pinecone, or ChromaDB.
- Create FastAPI endpoints to expose agent functionality.
- Implement Model Context Protocol (MCP) for tool-agent integrations.
- Optimize prompts, workflows, and retrieval strategies for real performance.
- Contribute to new agentic AI design patterns and innovations.
Who Should Apply
We’re looking for freshers who are:
-Strong in Python and love experimenting with AI/ML projects.
- Familiar with one or more of these: LangChain/LangGraph, HuggingFace, PyTorch/TensorFlow, RAG pipelines.
- Active on GitHub with 2–3 well-documented projects (clean code + clear README).
- Curious, hands-on builders who want to learn by doing.
Bonus Points if you’ve dabbled with:
- LLM fine-tuning (LoRA, QLoRA), memory systems. AutoGen, CrewAI, MCP, or other agent frameworks.
- Docker, async programming, API integrations.
Education:
- Completed/Pursuing Bachelor's in Computer Science or related field
- Strong foundation in ML theory and practice
Apply if:
- You have done projects using GenAI, Machine Learning, Deep Learning.
- You must have strong Python coding experience.
- Someone who is available immediately to start with us in the office(Hyderabad).
- Someone who has the hunger to learn something new always and aims to step up at a high pace.
We value quality implementations and thorough documentation over quantity. Show us how you think through problems and implement solutions!
Title: Senior Software Engineer – Python (Remote: Africa, India, Portugal)
Experience: 9 to 12 Years
INR : 40 LPA - 50 LPA
Location Requirement: Candidates must be based in Africa, India, or Portugal. Applicants outside these regions will not be considered.
Must-Have Qualifications:
- 8+ years in software development with expertise in Python
- kubernetes is important
- Strong understanding of async frameworks (e.g., asyncio)
- Experience with FastAPI, Flask, or Django for microservices
- Proficiency with Docker and Kubernetes/AWS ECS
- Familiarity with AWS, Azure, or GCP and IaC tools (CDK, Terraform)
- Knowledge of SQL and NoSQL databases (PostgreSQL, Cassandra, DynamoDB)
- Exposure to GenAI tools and LLM APIs (e.g., LangChain)
- CI/CD and DevOps best practices
- Strong communication and mentorship skills
Big companies are like giant boats with a thousand rowers — you can’t feel your pull move the boat. Shoppin isn’t that boat. We’re a 10-person crew rowing like our lives depend on it — each one the best at what they do, each stroke moving the product forward every single day. If you believe small, fast, obsessive teams can beat giants, read on.
What You’ll Do:
Build and optimize Shoppin’s vibe, image, and inspiration search, powering both text and image-based discovery.
Work on vector embeddings, retrieval pipelines, and semantic search using ElasticSearch, Redis caching, and LLM APIs.
Design and ship high-performance Python microservices that move fast and scale beautifully.
Experiment with prompt engineering, ranking models, and multimodal retrieval.
- Collaborate directly with the founder — moving from idea → prototype → production in hours, not weeks.
Tech You’ll Work With
- Languages & Frameworks: Python, FastAPI
- Search & Infra: ElasticSearch, Redis, PostgreSQL
- AI Stack: Vector Databases, Embeddings, LLM APIs (OpenAI, Gemini, etc.)
- Dev Tools: Cursor, Docker, Kubernetes
- Infra: AWS / GCP
What We’re Looking For
- Strong mathematical intuition — you understand cosine similarity, normalization, and ranking functions.
- Experience or deep curiosity in text + image search.
- Comfort with Python, data structures, and system design.
- Speed-obsessed — you optimize for velocity, not bureaucracy.
- Hungry to go all-in, ship hard things, and make a dent.
Bonus Points
- Experience with LLM prompting or orchestration.
- Exposure to recommendation systems, fashion/culture AI, or multimodal embeddings.
- You’ve built or scaled something end-to-end yourself.
SDE 2 / SDE 3 – AI Infrastructure & LLM Systems Engineer
Location: Pune / Bangalore (India)
Experience: 4–8 years
Compensation: no bar for the right candidate
Bonus: Up to 10% of base
About the Company
AbleCredit builds production-grade AI systems for BFSI enterprises, reducing OPEX by up to 70% across onboarding, credit, collections, and claims.
We run our own LLMs on GPUs, operate high-concurrency inference systems, and build AI workflows that must scale reliably under real enterprise traffic.
Role Summary (What We’re Really Hiring For)
We are looking for a strong backend / systems engineer who can:
- Deploy AI models on GPUs
- Expose them via APIs
- Scale inference under high parallel load using async systems and queues
This is not a prompt-engineering or UI-AI role.
Core Responsibilities
- Deploy and operate LLMs on GPU infrastructure (cloud or on-prem).
- Run inference servers such as vLLM / TGI / SGLang / Triton or equivalents.
- Build FastAPI / gRPC APIs on top of AI models.
- Design async, queue-based execution for AI workflows (fan-out, retries, backpressure).
- Plan and reason about capacity & scaling:
- GPU count vs RPS
- batching vs latency
- cost vs throughput
- Add observability around latency, GPU usage, queue depth, failures.
- Work closely with AI researchers to productionize models safely.
Must-Have Skills
- Strong backend engineering fundamentals (distributed systems, async workflows).
- Hands-on experience running GPU workloads in production.
- Proficiency in Python (Golang acceptable).
- Experience with Docker + Kubernetes (or equivalent).
- Practical knowledge of queues / workers (Redis, Kafka, SQS, Celery, Temporal, etc.).
- Ability to reason quantitatively about performance, reliability, and cost.
Strong Signals (Recruiter Screening Clues)
Look for candidates who have:
- Personally deployed models on GPUs
- Debugged GPU memory / latency / throughput issues
- Scaled compute-heavy backends under load
- Designed async systems instead of blocking APIs
Nice to Have
- Familiarity with LangChain / LlamaIndex (as infra layers, not just usage).
- Experience with vector DBs (Qdrant, Pinecone, Weaviate).
- Prior work on multi-tenant enterprise systems.
Not a Fit If
- Only experience is calling OpenAI / Anthropic APIs.
- Primarily a prompt engineer or frontend-focused AI dev.
- No hands-on ownership of infra, scaling, or production reliability.
At Pipaltree, we’re building an AI-enabled platform that helps brands understand how they’re truly perceived — not through surveys or static dashboards, but through real conversations happening across the world.
We’re a small team solving deep technical and product challenges: orchestrating large-scale conversation data, applying reasoning and summarization models, and turning this into insights that businesses can trust.
Requirements:
- Deep understanding of distributed systems and asynchronous programming in Python
- Experience with building scalable applications using LLMs or traditional ML techniques
- Experience with Databases, Cache, and Micro services
- Experience with DevOps is a huge plus
Responsibilities:
- Writing reusable, testable, and efficient code
- Design and implementation of low-latency, high-availability, and performant applications
- Integration of user-facing elements developed by front-end developers with server side logic
- Implementation of security and data protection
- Integration of data storage solutions (may include databases, key-value stores, blob stores, etc.)
- Expert in Python, with knowledge of at least one Python web framework (such as Django, Flask, etc depending on your technology stack)
- Familiarity with some ORM (Object Relational Mapper) libraries
- Able to integrate multiple data sources and databases into one system
- Understanding of the threading limitations of Python, and multi-process architecture
- Good understanding of server-side templating languages (such as Jinja 2, Mako, etc depending on your technology stack)
- Basic understanding of front-end technologies, such as JavaScript, HTML5, and CSS3
- Understanding of accessibility and security compliance (depending on the specific project)
- Knowledge of user authentication and authorization between multiple systems, servers, and environments
- Understanding of fundamental design principles behind a scalable application
- Familiarity with event-driven programming in Python
- Understanding of the differences between multiple delivery platforms, such as mobile vs desktop, and optimizing output to match the specific platform
- Able to create database schemas that represent and support business processes
- Strong unit test and debugging skills
- Basic knowledge of machine learning algorithm and libraries like keras, tensorflow, sklearn.
About Clari
Clari uses AI and automation to drive growth and retention for high-performing revenue teams. Clari’s Revenue Operations platform is currently processing over $300 billion in pipeline, and is used by over 50,000 marketing, sales and customer success professionals across 170 countries. Customers include market leaders like Symantec, Adobe, Alteryx, Workday, Lenovo, Zoom, Medallia, Alteryx and hundreds of others. Clari harvests and analyzes activity signals from dozens of different business systems, including email, calendar, CRM, marketing automation, to shorten sales cycles, increase win rates, and make revenue more predictable.
The result is passionate and frankly humbling customer loyalty. We consistently hear from our customers how we’ve changed their lives - just check out the reviews on https://www.g2crowd.com/products/clari/reviews">G2 Crowd. It never gets old, and we never take it for granted.
Clari is looking for several key experienced engineering who will focus on implementing many different areas of our solution, including but not limited to our overall web architecture, core application features such as data science driven analytics, user management, content management, social graph integration, personalization, emails, collaboration systems and enterprise content repositories as well as on unstructured data analytics, machine learning and our relevance engine.
Join our core applications team where you’ll work with truly remarkable colleagues on highly diverse, complex, and relevant problems while building scalable applications designed to service millions of mobile and web-based information workers. You’ll work closely with product managers, designers, and others in a cross functional environment on multiple projects, from concept phase through testing, launch and ongoing operations.
We work in an open, collaborative environment and seek exceptional developers who enjoy problem solving and straying outside the routine. You will also contribute to the growth of Clari by being a Brand Ambassador and assist in the hiring of great talent.
Qualifications
- 3+ years of professional server development experience using Java or similar object-oriented language
- Strong understanding of web-based architecture - web servers, load balancing, caching, databases etc.
- Basic knowledge of SQL (Postgres, MySQL) and NoSQL databases (MongoDB)
- Experience developing data driven web applications
- Up-to-date knowledge of latest trends in web application development, including Amazon AWS ecosystem
- Experience building and using RESTful APIs
Nice to Have
- Experience with multi-threading, replication etc. concepts in cloud applications
- Familiarity with large scale business intelligence applications
- Familiarity of JavaScript and other web technologies such as React
Why Clari?
Because we have a big mission, a winning product and an amazing fan base of passionate customers.
We’re changing the world and having a lot of fun on the way. Clari is a fun and fast-growing Silicon Valley company. Clari is one of Inc. Magazine’s best places to work in the US and was named as a 2019 Top Bay Area Workplace for the 5th consecutive year. In October 2019, we closed $60M in Series D funding and are growing at 200%. Our product is a winner - we have perennially been given the highest overall rating in G2 Crowd’s Top 20 Sales Analytics Software. We’re backed by top tier investors including Sequoia Capital, Bain Capital, Sapphire Ventures, Madrona Venture Group and Tenaya Capital, and have a superb and supportive board.
Our team is made up of veteran entrepreneurs, brilliant engineers, and tried-and-true sales professionals who have done this before and want to do it again, this time only bigger.
What’s left to add? You.












