
Looking for a Python Lead with 5 to 8 years of experience who can design, architect, and build secure, scalable, production-grade backend systems. The candidate should have expert-level proficiency in Python with strong hands-on experience in Django or Flask, RESTful API design, and Websockets. Must have a solid foundation in design patterns, algorithms, and data structures, along with advanced database skills across PostgreSQL, MySQL, and MongoDB. Should be well-versed in caching systems like Redis, queuing systems like RabbitMQ or Kafka, and container-based deployments using Docker. Proficient in cloud platforms — specifically AWS (EC2, S3, Lambda) and GCP (Compute Engine, Cloud Storage, Cloud Functions) — with experience setting up and managing CI/CD pipelines using Jenkins, GitLab CI, or GitHub Actions. Strong command of Git, Linux, and version control workflows is expected. Preferred candidates will have working knowledge of AI/ML concepts, hands-on experience with RAG pipelines, LLM integrations, and Vector Databases such as Pinecone or ChromaDB. The ideal candidate has a builder mindset, takes ownership from idea to production, writes clean and well-tested code, and is comfortable mentoring junior developers in a fast-paced startup environment. Must be based in India — open to working from Coimbatore or fully remote. Immediate or 30-day joiners preferred.

About Techjays
About
Company social profiles
Similar jobs
Job Title: Python Backend / GenAI Engineer (4+ Years)
Job Summary
Looking for a Python Backend Engineer with experience in Generative AI, LangGraph workflows, data engineering, and AI evaluation using Arize AI.
Responsibilities
* Develop backend APIs using Python (FastAPI / Flask / Django)
* Build Generative AI and RAG-based applications
* Design LangGraph / agent workflows
* Create data engineering pipelines (ETL, data processing)
* Implement LLM monitoring and evaluation using Arize AI
* Integrate vector databases and AI services
* Maintain scalable and production-ready backend systems
Required Skills
* 4+ years of Python backend development
* Experience in Generative AI / LLM applications
* Knowledge of LangGraph / LangChain
* Experience in data engineering pipelines
* Familiarity with Arize AI or model evaluation tools
* Understanding of REST APIs, databases, Docker
Good to Have
* Cloud platforms (Azure / AWS )
* Vector databases (FAISS, Pinecone, Azure AI Search)
Job Description: Python Developer
Experience: 2+ Years
Job Location: Nr. Iskcon Mega Mall, SG Highway, Ahmedabad
Timings: 10 AM to 7 PM
Job Description:
Technical Skills :
- Good knowledge of Python with 2+ years of minimum experience
- Strong understanding of various Python Libraries, APIs, and toolkits.
- Good experience in Django, Django REST Framework, and Flask framework.
- Understanding of AWS Serverless implementation using Lambda and API Gateway
- Hands-on Experience in Databases like Mysql, PostgreSQL.
- Good experience/understanding in Agentic AI / RAG.
- Proficient in NoSQL document databases especially MongoDB, Redis.
- Stronghold in Data Structures and Algorithm
- Thorough understanding of version control system concepts especially GIT.
- Understanding of the whole web stack and how all the pieces fit together (front-end, database, network layer, etc.) and how they impact the performance of your application.
- Excellent understanding of MVC and OOP. Bonus for the understanding of prevalent design patterns.
- Excellent debugging and optimization skills
Job Responsibilities :
- Building big, robust, scalable, and maintainable applications.
- Debugging, Fixing bugs, Identifying Performance Issues, and Improving App Performance.
- Continuously discover, evaluate, and implement new technologies to maximize development efficiency.
- Handling complex technical issues related to web app development & discussing solutions with the team.
- Developing, Deploying, and maintaining Multistage, Multi-tier applications.
- To write high-performing code and will be participating in key architectural decisions.
- Project Execution & Client Interaction
- Scrum Implementation
Python Developer - AI/MLYour Responsibilities
- Develop, train, and optimize ML models using PyTorch, TensorFlow, and Keras.
- Build end-to-end LLM and RAG pipelines using LangChain and LangGraph.
- Work with LLM APIs (OpenAI, Anthropic Claude, Azure OpenAI) and implement prompt engineering strategies.
- Utilize Hugging Face Transformers for model fine-tuning and deployment.
- Integrate embedding models for semantic search and retrieval systems.
- Work with transformer-based architectures (BERT, GPT, LLaMA, Mistral) for production use cases.
- Implement LLM evaluation frameworks (RAGAS, LangSmith) and performance optimization.
- Design and maintain Python microservices using FastAPI with REST/GraphQL APIs.
- Implement real-time communication with FastAPI WebSockets.
- Implement pgvector for embedding storage and similarity search with efficient indexing strategies.
- Integrate vector databases (pgvector, Pinecone, Weaviate, FAISS, Milvus) for retrieval pipelines.
- Containerize AI services with Docker and deploy on Kubernetes (EKS/GKE/AKS).
- Configure AWS infrastructure (EC2, S3, RDS, SageMaker, Lambda, CloudWatch) for AI/ML workloads.
- Version ML experiments using MLflow, Weights & Biases, or Neptune.
- Deploy models using serving frameworks (TorchServe, BentoML, TensorFlow Serving).
- Implement model monitoring, drift detection, and automated retraining pipelines.
- Build CI/CD pipelines for automated testing and deployment with ≥80% test coverage (pytest).
- Follow security best practices for AI systems (prompt injection prevention, data privacy, API key management).
- Participate in code reviews, tech talks, and AI learning sessions.
- Follow Agile/Scrum methodologies and Git best practices.
Required Qualifications
- Bachelor's or Master's degree in Computer Science, AI/ML, or related field.
- 2–5 years of Python development experience (Python 3.9+) with strong AI/ML background.
- Hands-on experience with LangChain and LangGraph for building LLM-powered workflows and RAG systems.
- Deep learning experience with PyTorch or TensorFlow.
- Experience with Hugging Face Transformers and model fine-tuning.
- Proficiency with LLM APIs (OpenAI, Anthropic, Azure OpenAI) and prompt engineering.
- Strong experience with FastAPI frameworks.
- Proficiency in PostgreSQL with pgvector extension for embedding storage and similarity search.
- Experience with vector databases (pgvector, Pinecone, Weaviate, FAISS, or Milvus).
- Experience with model versioning tools (MLflow, Weights & Biases, or Neptune).
- Hands-on with Docker, Kubernetes basics, and AWS cloud services.
- Skilled in Git workflows, automated testing (pytest), and CI/CD practices.
- Understanding of security principles for AI systems.
- Excellent communication and analytical thinking.
Nice to Have
- Experience with multiple vector databases (Pinecone, Weaviate, FAISS, Milvus).
- Knowledge of advanced LLM fine-tuning (LoRA, QLoRA, PEFT) and RLHF.
- Experience with model serving frameworks and distributed training.
- Familiarity with workflow orchestration tools (Airflow, Prefect, Dagster).
- Knowledge of quantization and model compression techniques.
- Experience with infrastructure as code (Terraform, CloudFormation).
- Familiarity with data versioning tools (DVC) and AutoML.
- Experience with Streamlit or Gradio for ML demos.
- Background in statistics, optimization, or applied mathematics.
- Contributions to AI/ML or LangChain/LangGraph open-source projects.
Designation: Python Developer
Experienced in AI/ML
Location: Turbhe, Navi Mumbai
CTC: 6-12 LPA
Years of Experience: 2-5 years
At Arcitech.ai, we’re redefining the future with AI-powered software solutions across education, recruitment, marketplaces, and beyond. We’re looking for a Python Developer passionate about AI/ML, who’s ready to work on scalable, cloud-native platforms and help build the next generation of intelligent, LLM-driven products.
💼 Your Responsibilities
AI/ML Engineering
- Develop, train, and optimize ML models using PyTorch/TensorFlow/Keras.
- Build end-to-end LLM and RAG (Retrieval-Augmented Generation) pipelines using LangChain.
- Collaborate with data scientists to convert prototypes into production-grade AI applications.
- Integrate NLP, Computer Vision, and Recommendation Systems into scalable products.
- Work with transformer-based architectures (BERT, GPT, LLaMA, etc.) for real-world AI use cases.
Backend & Systems Development
- Design, develop, and maintain robust Python microservices with REST/GraphQL APIs.
- Implement real-time communication with Django Channels/WebSockets.
- Containerize AI services with Docker and deploy on Kubernetes (EKS/GKE/AKS).
- Configure and manage AWS (EC2, S3, RDS, SageMaker, CloudWatch) for AI/ML workloads.
Reliability & Automation
- Develop background task queues with Celery, ensuring smart retries and monitoring.
- Implement CI/CD pipelines for automated model training, testing, and deployment.
- Write automated unit & integration tests (pytest/unittest) with ≥80% coverage.
Collaboration
- Contribute to MLOps best practices and mentor peers in LangChain/AI integration.
- Participate in tech talks, code reviews, and AI learning sessions within the team.
🎓 Required Qualifications
- Bachelor’s or Master’s degree in Computer Science, AI/ML, or related field.
- 2–5 years of experience in Python development with strong AI/ML exposure.
- Hands-on experience with LangChain for building LLM-powered workflows and RAG systems.
- Deep learning experience with PyTorch or TensorFlow.
- Experience deploying ML models and LLM apps into production systems.
- Familiarity with REST/GraphQL APIs and cloud platforms (AWS/Azure/GCP).
- Skilled in Git workflows, automated testing, and CI/CD practices.
🌟 Nice to Have
- Experience with vector databases (Pinecone, Weaviate, FAISS, Milvus) for retrieval pipelines.
- Knowledge of LLM fine-tuning, prompt engineering, and evaluation frameworks.
- Familiarity with Airflow/Prefect/Dagster for data and model pipelines.
- Background in statistics, optimization, or applied mathematics.
- Contributions to AI/ML or LangChain open-source projects.
- Experience with model monitoring and drift detection in production.
🎁 Why Join Us
- Competitive compensation and benefits 💰
- Work on cutting-edge LLM and AI/ML applications 🤖
- A collaborative, innovation-driven work culture 📚
- Opportunities to grow into AI/ML leadership roles 🚀
Role Overview
We are looking for a Senior Backend Developer with strong Node.js expertise who is comfortable working in an AI-assisted development environment. The ideal candidate should be proficient in leveraging modern AI developer tools to accelerate development, improve code quality, and help modernize legacy systems.
This role involves working on backend systems, microservices architecture, and AI-powered development workflows, including refactoring legacy applications into modern Node.js services.
Key Responsibilities
- Design, build, and maintain scalable backend services using Node.js.
- Leverage AI-powered developer tools such as GitHub Copilot, Cursor, or Amazon Q to improve development efficiency.
- Refactor and modernize legacy .NET applications into Node.js-based microservices using AI-assisted workflows.
- Design and optimize database schemas and complex SQL queries for performance and scalability.
- Work with PostgreSQL and Redis for high-performance data storage and caching.
- Build and integrate AI-based features and services into backend applications.
- Automate development workflows through custom CLI tools or IDE extensions.
- Collaborate with cross-functional teams to deliver high-quality backend solutions.
Required Skills & Experience
- 5+ years of backend development experience
- Strong expertise in Node.js backend development
- Hands-on experience with AI-assisted development tools (GitHub Copilot, Cursor, Amazon Q, or similar)
- Experience working with PostgreSQL and Redis
- Strong understanding of microservices architecture and API design
- Experience with AI engineering (building or integrating AI-powered features)
- Strong problem-solving and debugging skills
Good to Have
- Prior experience with .NET / C#
- Experience modernizing legacy systems
- Experience building developer productivity tools (CLI tools, IDE extensions, automation)
- Experience optimizing complex SQL queries and database migrations
A LITTLE BIT ABOUT THE COMPANY:
Established in 2017, Fountane Inc is one part a Digital Product Studio that specializes in building superior product experiences, and one part Ventures Lab incubating and investing in new competitive technology businesses from scratch. Thus far, we’ve created half a dozen multi million valuation companies in the US, and a handful of sister ventures for large corporations including Target, US Ventures, Imprint Engine.
We’re a team of 100 strong from around the world that are radically open minded and believes in excellence, respecting one another and pushing our boundaries to furthest its ever been.
Job Title: AI/ML Engineer – Voice (2–3 Years)
Location: Bengaluru (On-site)
Employment Type: Full-time
About Impacto Digifin Technologies
Impacto Digifin Technologies enables enterprises to adopt digital transformation through intelligent, AI-powered solutions. Our platforms reduce manual work, improve accuracy, automate complex workflows, and ensure compliance—empowering organizations to operate with speed, clarity, and confidence.
We combine automation where it’s fastest with human oversight where it matters most. This hybrid approach ensures trust, reliability, and measurable efficiency across fintech and enterprise operations.
Role Overview
We are looking for an AI Engineer Voice with strong applied experience in machine learning, deep learning, NLP, GenAI, and full-stack voice AI systems.
This role requires someone who can design, build, deploy, and optimize end-to-end voice AI pipelines, including speech-to-text, text-to-speech, real-time streaming voice interactions, voice-enabled AI applications, and voice-to-LLM integrations.
You will work across core ML/DL systems, voice models, predictive analytics, banking-domain AI applications, and emerging AGI-aligned frameworks. The ideal candidate is an applied engineer with strong fundamentals, the ability to prototype quickly, and the maturity to contribute to R&D when needed.
This role is collaborative, cross-functional, and hands-on.
Key Responsibilities
Voice AI Engineering
- Build end-to-end voice AI systems, including STT, TTS, VAD, audio processing, and conversational voice pipelines.
- Implement real-time voice pipelines involving streaming interactions with LLMs and AI agents.
- Design and integrate voice calling workflows, bi-directional audio streaming, and voice-based user interactions.
- Develop voice-enabled applications, voice chat systems, and voice-to-AI integrations for enterprise workflows.
- Build and optimize audio preprocessing layers (noise reduction, segmentation, normalization)
- Implement voice understanding modules, speech intent extraction, and context tracking.
Machine Learning & Deep Learning
- Build, deploy, and optimize ML and DL models for prediction, classification, and automation use cases.
- Train and fine-tune neural networks for text, speech, and multimodal tasks.
- Build traditional ML systems where needed (statistical, rule-based, hybrid systems).
- Perform feature engineering, model evaluation, retraining, and continuous learning cycles.
NLP, LLMs & GenAI
- Implement NLP pipelines including tokenization, NER, intent, embeddings, and semantic classification.
- Work with LLM architectures for text + voice workflows
- Build GenAI-based workflows and integrate models into production systems.
- Implement RAG pipelines and agent-based systems for complex automation.
Fintech & Banking AI
- Work on AI-driven features related to banking, financial risk, compliance automation, fraud patterns, and customer intelligence.
- Understand fintech data structures and constraints while designing AI models.
Engineering, Deployment & Collaboration
- Deploy models on cloud or on-prem (AWS / Azure / GCP / internal infra).
- Build robust APIs and services for voice and ML-based functionalities.
- Collaborate with data engineers, backend developers, and business teams to deliver end-to-end AI solutions.
- Document systems and contribute to internal knowledge bases and R&D.
Security & Compliance
- Follow fundamental best practices for AI security, access control, and safe data handling.
- Awareness of financial compliance standards (plus, not mandatory).
- Follow internal guidelines on PII, audio data, and model privacy.
Primary Skills (Must-Have)
Core AI
- Machine Learning fundamentals
- Deep Learning architectures
- NLP pipelines and transformers
- LLM usage and integration
- GenAI development
- Voice AI (STT, TTS, VAD, real-time pipelines)
- Audio processing fundamentals
- Model building, tuning, and retraining
- RAG systems
- AI Agents (orchestration, multi-step reasoning)
Voice Engineering
- End-to-end voice application development
- Voice calling & telephony integration (framework-agnostic)
- Realtime STT ↔ LLM ↔ TTS interactive flows
- Voice chat system development
- Voice-to-AI model integration for automation
Fintech/Banking Awareness
- High-level understanding of fintech and banking AI use cases
- Data patterns in core banking analytics (advantageous)
Programming & Engineering
- Python (strong competency)
- Cloud deployment understanding (AWS/Azure/GCP)
- API development
- Data processing & pipeline creation
Secondary Skills (Good to Have)
- MLOps & CI/CD for ML systems
- Vector databases
- Prompt engineering
- Model monitoring & evaluation frameworks
- Microservices experience
- Basic UI integration understanding for voice/chat
- Research reading & benchmarking ability
Qualifications
- 2–3 years of practical experience in AI/ML/DL engineering.
- Bachelor’s/Master’s degree in CS, AI, Data Science, or related fields.
- Proven hands-on experience building ML/DL/voice pipelines.
- Experience in fintech or data-intensive domains preferred.
Soft Skills
- Clear communication and requirement understanding
- Curiosity and research mindset
- Self-driven problem solving
- Ability to collaborate cross-functionally
- Strong ownership and delivery discipline
- Ability to explain complex AI concepts simply
We're looking for AI/ML enthusiasts who build, not just study. If you've implemented transformers from scratch, fine-tuned LLMs, or created innovative ML solutions, we want to see your work!
What You’ll Do
-Build autonomous AI agents using LangChain, LangGraph, and similar frameworks.
- Develop RAG pipelines with vector DBs like FAISS, Pinecone, or ChromaDB.
- Create FastAPI endpoints to expose agent functionality.
- Implement Model Context Protocol (MCP) for tool-agent integrations.
- Optimize prompts, workflows, and retrieval strategies for real performance.
- Contribute to new agentic AI design patterns and innovations.
Who Should Apply
We’re looking for freshers who are:
-Strong in Python and love experimenting with AI/ML projects.
- Familiar with one or more of these: LangChain/LangGraph, HuggingFace, PyTorch/TensorFlow, RAG pipelines.
- Active on GitHub with 2–3 well-documented projects (clean code + clear README).
- Curious, hands-on builders who want to learn by doing.
Bonus Points if you’ve dabbled with:
- LLM fine-tuning (LoRA, QLoRA), memory systems. AutoGen, CrewAI, MCP, or other agent frameworks.
- Docker, async programming, API integrations.
Education:
- Completed/Pursuing Bachelor's in Computer Science or related field
- Strong foundation in ML theory and practice
Apply if:
- You have done projects using GenAI, Machine Learning, Deep Learning.
- You must have strong Python coding experience.
- Someone who is available immediately to start with us in the office(Hyderabad).
- Someone who has the hunger to learn something new always and aims to step up at a high pace.
We value quality implementations and thorough documentation over quantity. Show us how you think through problems and implement solutions!
We're Hiring: Senior Developer (AI & Machine Learning)** 🚀
🔧 **Tech Stack**: Python, Neo4j, FAISS, LangChain, React.js, AWS/GCP/Azure
🧠 **Role**: AI/ML development, backend architecture, cloud deployment
🌍 **Location**: Remote (India)
💼 **Experience**: 5-10 years
If you're passionate about making an impact in EdTech and want to help shape the future of learning with AI, we want to hear from you!
About Synorus
Synorus is building a next-generation ecosystem of AI-first products. Our flagship legal-AI platform LexVault is redefining legal research, drafting, knowledge retrieval, and case intelligence using domain-tuned LLMs, private RAG pipelines, and secure reasoning systems.
If you are passionate about AI, legaltech, and training high-performance models — this internship will put you on the front line of innovation.
Role Overview
We are seeking passionate AI/LLM Engineering Interns who can:
- Fine-tune LLMs for legal domain use-cases
- Train and experiment with open-source foundation models
- Work with large datasets efficiently
- Build RAG pipelines and text-processing frameworks
- Run model training workflows on Google Colab / Kaggle / Cloud GPUs
This is a hands-on engineering and research internship — you will work directly with senior founders & technical leadership.
Key Responsibilities
- Fine-tune transformer-based models (Llama, Mistral, Gemma, etc.)
- Build and preprocess legal datasets at scale
- Develop efficient inference & training pipelines
- Evaluate models for accuracy, hallucinations, and trustworthiness
- Implement RAG architectures (vector DBs + embeddings)
- Work with GPU environments (Colab/Kaggle/Cloud)
- Contribute to model improvements, prompt engineering & safety tuning
Must-Have Skills
- Strong knowledge of Python & PyTorch
- Understanding of LLMs, Transformers, Tokenization
- Hands-on experience with HuggingFace Transformers
- Familiarity with LoRA/QLoRA, PEFT training
- Data wrangling: Pandas, NumPy, tokenizers
- Ability to handle multi-GB datasets efficiently
Bonus Skills
(Not mandatory — but a strong plus)
- Experience with RAG / vector DBs (Chroma, Qdrant, LanceDB)
- Familiarity with vLLM, llama.cpp, GGUF
- Worked on summarization, Q&A or document-AI projects
- Knowledge of legal texts (Indian laws/case-law/statutes)
- Open-source contributions or research work
What You Will Gain
- Real-world training on LLM fine-tuning & legal AI
- Exposure to production-grade AI pipelines
- Direct mentorship from engineering leadership
- Research + industry project portfolio
- Letter of experience + potential full-time offer
Ideal Candidate
- You experiment with models on weekends
- You love pushing GPUs to their limits
- You prefer research + implementation over theory alone
- You want to build AI that matters — not just demos
Location - Remote
Stipend - 5K - 10K











