
Job Title: AI Engineer - NLP/LLM Data Product Engineer Location: Chennai, TN- Hybrid
Duration: Full time
Job Summary:
About the Role:
We are growing our Data Science and Data Engineering team and are looking for an
experienced AI Engineer specializing in creating GenAI LLM solutions. This position involves collaborating with clients and their teams, discovering gaps for automation using AI, designing customized AI solutions, and implementing technologies to streamline data entry processes within the healthcare sector.
Responsibilities:
· Conduct detailed consultations with clients functional teams to understand client requirements, one use case is related to handwritten medical records.
· Analyze existing data entry workflows and propose automation opportunities.
Design:
· Design tailored AI-driven solutions for the extraction and digitization of information from handwritten medical records.
· Collaborate with clients to define project scopes and objectives.
Technology Selection:
· Evaluate and recommend AI technologies, focusing on NLP, LLM and machine learning.
· Ensure seamless integration with existing systems and workflows.
Prototyping and Testing:
· Develop prototypes and proof-of-concept models to demonstrate the feasibility of proposed solutions.
· Conduct rigorous testing to ensure accuracy and reliability.
Implementation and Integration:
· Work closely with clients and IT teams to integrate AI solutions effectively.
· Provide technical support during the implementation phase.
Training and Documentation:
· Develop training materials for end-users and support staff.
· Create comprehensive documentation for implemented solutions.
Continuous Improvement:
· Monitor and optimize the performance of deployed solutions.
· Identify opportunities for further automation and improvement.
Qualifications:
· Advanced degree in Computer Science, Artificial Intelligence, or related field (Masters or PhD required).
· Proven experience in developing and implementing AI solutions for data entry automation.
· Expertise in NLP, LLM and other machine-learning techniques.
· Strong programming skills, especially in Python.
· Familiarity with healthcare data privacy and regulatory requirements.
Additional Qualifications( great to have):
An ideal candidate will have expertise in the most current LLM/NLP models, particularly in the extraction of data from clinical reports, lab reports, and radiology reports. The ideal candidate should have a deep understanding of EMR/EHR applications and patient-related data.

Similar jobs
SDE 2 / SDE 3 – AI Infrastructure & LLM Systems Engineer
Location: Pune / Bangalore (India)
Experience: 4–8 years
Compensation: no bar for the right candidate
Bonus: Up to 10% of base
About the Company
AbleCredit builds production-grade AI systems for BFSI enterprises, reducing OPEX by up to 70% across onboarding, credit, collections, and claims.
We run our own LLMs on GPUs, operate high-concurrency inference systems, and build AI workflows that must scale reliably under real enterprise traffic.
Role Summary (What We’re Really Hiring For)
We are looking for a strong backend / systems engineer who can:
- Deploy AI models on GPUs
- Expose them via APIs
- Scale inference under high parallel load using async systems and queues
This is not a prompt-engineering or UI-AI role.
Core Responsibilities
- Deploy and operate LLMs on GPU infrastructure (cloud or on-prem).
- Run inference servers such as vLLM / TGI / SGLang / Triton or equivalents.
- Build FastAPI / gRPC APIs on top of AI models.
- Design async, queue-based execution for AI workflows (fan-out, retries, backpressure).
- Plan and reason about capacity & scaling:
- GPU count vs RPS
- batching vs latency
- cost vs throughput
- Add observability around latency, GPU usage, queue depth, failures.
- Work closely with AI researchers to productionize models safely.
Must-Have Skills
- Strong backend engineering fundamentals (distributed systems, async workflows).
- Hands-on experience running GPU workloads in production.
- Proficiency in Python (Golang acceptable).
- Experience with Docker + Kubernetes (or equivalent).
- Practical knowledge of queues / workers (Redis, Kafka, SQS, Celery, Temporal, etc.).
- Ability to reason quantitatively about performance, reliability, and cost.
Strong Signals (Recruiter Screening Clues)
Look for candidates who have:
- Personally deployed models on GPUs
- Debugged GPU memory / latency / throughput issues
- Scaled compute-heavy backends under load
- Designed async systems instead of blocking APIs
Nice to Have
- Familiarity with LangChain / LlamaIndex (as infra layers, not just usage).
- Experience with vector DBs (Qdrant, Pinecone, Weaviate).
- Prior work on multi-tenant enterprise systems.
Not a Fit If
- Only experience is calling OpenAI / Anthropic APIs.
- Primarily a prompt engineer or frontend-focused AI dev.
- No hands-on ownership of infra, scaling, or production reliability.
About the Role
We are seeking a highly skilled Tech Lead who can drive the development of high-performance, scalable products while contributing hands-on to coding, architecture, and system design. You will work closely with a small engineering team, guiding them in technology decisions, quality delivery, and best practices while also taking ownership of key modules yourself.
Key Responsibilities
- Lead end-to-end product development from design to deployment.
- Act as both a technical contributor and mentor within a small, agile team.
- Architect scalable, robust, and secure backend and frontend systems.
- Participate in and guide:
- System design & architecture decisions
- Algorithmic analysis and performance optimization
- Database modeling and API design
- Write clean, maintainable, high-quality code.
- Own deployment pipelines and ensure reliable production releases.
- Collaborate with cross-functional stakeholders to translate requirements into actionable technical plans.
- Conduct code reviews, enforce coding standards, and ensure engineering excellence.
- Troubleshoot and solve complex technical challenges in production environments.
Required Technical Skills
- Strong expertise in Node.js and/or Python.
- Hands-on experience with TypeScript and React.js for frontend development.
- Knowledge of Go (Golang) is a strong advantage.
- Deep understanding of:
- Data Structures & Algorithms
- System Design
- Distributed systems concepts
- Microservices architecture
- Experience with deployment using:
- Docker, Kubernetes, or similar orchestration tools
- CI/CD pipelines
- Cloud platforms (AWS / GCP / Azure)
- Strong understanding of databases (SQL and NoSQL).
Soft Skills
- Strong leadership and ability to guide a small development team.
- Clear communication with technical and non-technical stakeholders.
- Ownership mindset with focus on delivering quality, on time.
- Problem-solver with the ability to make quick and correct technical decisions.
Preferred Qualifications
- 4–10 years of experience in software development.
- Prior experience in a startup or fast-moving product environment.
- Experience deploying and managing real-world products at scale.
What We Offer
- Opportunity to build and influence real products end-to-end.
- A collaborative, high-ownership environment.
- Freedom to experiment and drive product and engineering decisions.
Job Description: Software Engineer - Backend ( 3-5 Years)
Location: Bangalore
WHO WE ARE:
TIFIN is a fintech platform backed by industry leaders including JP Morgan, Morningstar, Broadridge, Hamilton Lane, Franklin Templeton, Motive Partners and a who’s who of the financial service industry. We are creating engaging wealth experiences to better financial lives
through AI and investment intelligence powered personalization. We are working to change the world of wealth in ways that personalization has changed the world of movies, music and more but with the added responsibility of delivering better wealth outcomes.
We use design and behavioral thinking to enable engaging experiences through software and application programming interfaces (APIs). We use investment science and intelligence to build algorithmic engines inside the software and APIs to enable better investor outcomes.
In a world where every individual is unique, we match them to financial advice and investments with a recognition of their distinct needs and goals across our investment marketplace and our advice and planning divisions.
OUR VALUES: Go with your GUT
●Grow at the Edge: We embrace personal growth by stepping out of our comfort zones to
discover our genius zones, driven by self-awareness and integrity. No excuses.
●Understanding through Listening and Speaking the Truth: Transparency, radical candor,
and authenticity define our communication. We challenge ideas, but once decisions are
made, we commit fully.
●I Win for Teamwin: We operate within our genius zones, taking ownership of our work
and inspiring our team with energy and attitude to win together.
Responsibilities:
• Contribute to the entire implementation process including driving the definition of improvements based on business needs and architectural improvements.
• Review code for quality and implementation of best practices.
• Promote coding, testing, and deployment best practices through hands-on research and demonstration.
• Write testable code that enables extremely high levels of code coverage.
• Ability to review frameworks and design principles toward suitability in the project context.
• Candidates who will demonstrate an ability to identify an opportunity lay out a rational plan for pursuing that opportunity, and see it through to completion.
Requirements:
• Engineering graduate with 3+ years of experience in software product development.
• Proficient in Python, Django, Pandas, GitHub, and AWS.
• Good knowledge of PostgreSQL, and MongoDB.
• Strong Experience in designing REST APIs.
• Experience with working on scalable interactive web applications.
• A clear understanding of software design constructs and their implementation.
• Understanding of the threading limitations of Python and multi-process architecture.
• Familiarity with some ORM (Object Relational Mapper) libraries.
• Good understanding of Test Driven Development.
• Unit and Integration testing.
• Preferred exposure to Finance domain.
• Strong written and oral communication skills.
Key Responsibilities
- Develop and maintain Python-based applications.
- Design and optimize SQL queries and databases.
- Collaborate with cross-functional teams to define, design, and ship new features.
- Write clean, maintainable, and efficient code.
- Troubleshoot and debug applications.
- Participate in code reviews and contribute to team knowledge sharing.
Qualifications and Required Skills
- Strong proficiency in Python programming.
- Experience with SQL and database management.
- Experience with web frameworks such as Django or Flask.
- Knowledge of front-end technologies like HTML, CSS, and JavaScript.
- Familiarity with version control systems like Git.
- Strong problem-solving skills and attention to detail.
- Excellent communication and teamwork skills.
Good to Have Skills
- Experience with cloud platforms like AWS or Azure.
- Knowledge of containerization technologies like Docker.
- Familiarity with continuous integration and continuous deployment (CI/CD) pipelines
Must have exp in Django. [ Mandatory]
- The ability to problem-solve and critically think.
- High level of knowledge of Python and the Django framework.
- Familiarity with event driven programming as well as the MVC.
- Good understanding of SQL databases.
- Good understanding of REST APIs.
The Sr. Analytics Engineer would provide technical expertise in needs identification, data modeling, data movement, and transformation mapping (source to target), automation and testing strategies, translating business needs into technical solutions with adherence to established data guidelines and approaches from a business unit or project perspective.
Understands and leverages best-fit technologies (e.g., traditional star schema structures, cloud, Hadoop, NoSQL, etc.) and approaches to address business and environmental challenges.
Provides data understanding and coordinates data-related activities with other data management groups such as master data management, data governance, and metadata management.
Actively participates with other consultants in problem-solving and approach development.
Responsibilities :
Provide a consultative approach with business users, asking questions to understand the business need and deriving the data flow, conceptual, logical, and physical data models based on those needs.
Perform data analysis to validate data models and to confirm the ability to meet business needs.
Assist with and support setting the data architecture direction, ensuring data architecture deliverables are developed, ensuring compliance to standards and guidelines, implementing the data architecture, and supporting technical developers at a project or business unit level.
Coordinate and consult with the Data Architect, project manager, client business staff, client technical staff and project developers in data architecture best practices and anything else that is data related at the project or business unit levels.
Work closely with Business Analysts and Solution Architects to design the data model satisfying the business needs and adhering to Enterprise Architecture.
Coordinate with Data Architects, Program Managers and participate in recurring meetings.
Help and mentor team members to understand the data model and subject areas.
Ensure that the team adheres to best practices and guidelines.
Requirements :
- Strong working knowledge of at least 3 years of Spark, Java/Scala/Pyspark, Kafka, Git, Unix / Linux, and ETL pipeline designing.
- Experience with Spark optimization/tuning/resource allocations
- Excellent understanding of IN memory distributed computing frameworks like Spark and its parameter tuning, writing optimized workflow sequences.
- Experience of relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., Redshift, Bigquery, Cassandra, etc).
- Familiarity with Docker, Kubernetes, Azure Data Lake/Blob storage, AWS S3, Google Cloud storage, etc.
- Have a deep understanding of the various stacks and components of the Big Data ecosystem.
- Hands-on experience with Python is a huge plus
Role Description:
As a Lead Backend Engineer , you will be responsible for designing and deploying scalable, highly available, and fault-tolerant systems for Radiusagent. Experience as a Tech Lead, managing tasks with back-end engineers while coding themselves will be a bonus point for the ideal candidate.
Responsibilities will include-
- You will contribute to all aspects of an agile software development lifecycle including design, architecture, development, documentation, testing, and operations.
- You will build cutting-edge scalable systems by writing simple and efficient code.
- You will push your design and architecture limits for new product development.
- You will ensure compliance with the build/release and configuration management process.
Skills needed:
- be able to design and build modules from ground up
- proficient in any one programming language out of php/golang/node/python/javascript/java/ruby and ability to pick up others
- experience with mysql/postgres/mongodb/couchdb
- experience with redis/memcached
- experience with rabbitmq / kafka
- experience in having written cron jobs, troubleshooting downtime
- ability to work with ambiguous requirements
- good understanding of OS concepts
Nice to haves:
- experience with managing VMs' on aws/gcp/digitalocean
- experience with docker/kubernetes
- experience elasticsearch or lucene
- Expert in Python. Comfortable with Web frameworks, such as Flask or Django
- Familiarity with Object Relational Mapping Libraries and ability to Integrate with Multiple Data Sources into One System
- Understanding of Limitations of Python and Multi Process Architecture. Understanding of Design Principles of Scalable Application
- Has familiarity with data frameworks in Python - Pandas or
- Good object-oriented design skills and knowledge of design
- Knowledge of key-value stores, caching, search, messaging queues
- Minimum 5 years of experience in the above










