Cutshort logo
Remote python jobs

50+ Remote Python Jobs in India

Apply to 50+ Remote Python Jobs on CutShort.io. Find your next job, effortlessly. Browse Python Jobs and apply today!

icon
codeviorcom
prakash rawat
Posted by prakash rawat
Remote only
0 - 2 yrs
₹1L - ₹3L / yr
skill iconPython
skill iconJavascript
skill iconNodeJS (Node.js)
skill iconPostgreSQL
SQL
+3 more

Primary Title: Software Engineer

About the Opportunity

A product engineering organization operating in the SaaS and cloud-native web applications sector, delivering scalable B2B platforms and developer-facing tooling. We build API-first, microservices‑oriented solutions and focus on high-quality user experiences, performance, and continuous delivery. This is a fully remote role for engineers based in India who want to own end-to-end features and ship production code.

Join Codevior’s engineering team to work closely with product, design, and QA to design, build, and operate customer-facing features and internal platform services. You’ll help raise engineering standards, improve observability, and accelerate delivery for fast-growing product lines.

Role & Responsibilities

  • Design, implement, and maintain production-grade web services and front-end features using modern JavaScript/TypeScript stacks.
  • Build and consume RESTful APIs; collaborate on API contracts and ensure backward compatibility for clients.
  • Write clean, testable code and maintain automated unit and integration tests to support CI/CD pipelines.
  • Diagnose and resolve performance, scalability, and reliability issues in cloud-hosted environments.
  • Participate in code reviews, drive technical design discussions, and contribute to engineering best practices and documentation.
  • Collaborate with cross-functional teams to deliver features from concept through deployment and monitoring in production.

Skills & Qualifications

Must-Have

  • Proven experience building web applications using JavaScript or TypeScript.
  • Backend development with Node.js and experience designing RESTful APIs.
  • Front-end experience with React or a comparable component framework.
  • Database proficiency with SQL and data modeling for transactional systems.
  • Version control with Git and familiarity with collaborative branching workflows.
  • Containerization experience with Docker and basic familiarity with cloud deployment patterns.

Preferred

  • Experience with CI/CD tooling and automated test frameworks (unit and integration testing).
  • Familiarity with cloud platforms such as AWS, Azure, or GCP and serverless / microservices patterns.
  • Knowledge of performance tuning, observability (logs/metrics/tracing), and production incident handling.

Benefits & Culture Highlights

  • 100% remote and flexible work hours — work from anywhere in India.
  • Learning & development stipend, regular tech talks, and mentorship from senior engineers.
  • Collaborative, outcome-focused culture with emphasis on engineering craftsmanship and ownership.

How to apply: If you are a pragmatic, delivery-focused engineer who enjoys ownership and building reliable systems, we’d love to hear from you. Include links to your GitHub or portfolio and a short note on your most impactful project.

Read more
Tonomo
Remote only
7 - 12 yrs
$18K - $21.6K / yr
Artificial Intelligence (AI)
skill iconFlutter
skill iconAndroid Development
skill iconiOS App Development
skill iconPython
+16 more

The Mission Tonomo is revolutionizing e-commerce with an intelligent, autonomous platform powered by IoT and AI. We are in the Beta phase, rapidly iterating based on user feedback. We need an "Unblocker"—a senior engineer who owns the mobile experience but can dive into the Python backend to build the endpoints they need to move fast.

The Engineering Culture We believe in AI-Augmented Engineering. We expect you to use tools like Cursor, Copilot, Gemini, GPT-4 and alike, to handle boilerplate code, allowing you to focus on complex native bridges, system architecture, and "on-the-spot" bug resolution.

Core Responsibilities

  • Flutter Mastery: Lead the development of our cross-platform Beta app (Android, iOS, and Web) using Flutter.
  • Backend Independence: Build and modify REST APIs and microservices in Python (FastAPI) to unblock frontend features.
  • AI coding: tools like Cursor, Copilot, Gemini, GPT-4 and alike
  • Agile Troubleshooting: Fix critical UI and logical bugs "on the spot" as reported by users. Experience with UI/UX best practices.
  • Performance & Debugging: Proactively monitor app health, experienced with Sentry, Firebase Crashlytics, and Flutter DevTools
  • IoT & Integration: Work with IoT telemetry protocols (MQTT) and integrate third-party services for payments (Stripe) and Firebase.
  • Native Depth: Develop custom plugins and MethodChannels to bridge Flutter with native iOS/Android functionalities.
  • Dashboard Ownership: Own dashboards end-to-end. Design and build internal dashboards for: Business Intelligence. System health and operational metrics. IoT and backend activity insights.
  • Frontend Development Build modern, responsive web dashboards using React (or similar). Implement advanced data visualizations. Focus on clarity, performance, and usability for non-technical stakeholders.
  • BI & Data Integration: Integrate dashboards with: Backend APIs (Python / FastAPI). Databases (PostgreSQL). Analytics / metrics sources (Grafana, Prometheus, or BI tools). Work with product & ops to define what should be measured.
  • Monitoring & Insights: Build visual views on top of monitoring data (Grafana or embedded views). Help translate raw metrics into actionable insights. Support ad-hoc analysis and investigation workflows.
  • Execution & Iteration: Move fast in a startup environment: iterate dashboards based on real feedback. Improve data quality, consistency, and trust over time.

Technical Requirements

  • Mobile Experience: 7+ years in mobile development with at least 5 highly distributed apps published.
  • The Stack: * Frontend: Expert Flutter/Dart skills
  • Backend: Proficient Python developer with experience in FastAPI, SQLAlchemy, and PostgreSQL.
  • Data & Backend Awareness: Comfortable consuming REST APIs and working with structured data.
  • Ability to collaborate on schema design and API contracts.
  • BI / Analytics (Nice to Have): Experience with BI tools or platforms (Grafana, Metabase, Superset, Looker, etc.).
  • Understanding of KPIs, funnels, and business metrics.
  • Experience embedding dashboards or analytics into web apps.
  • Architecture: Mastery of design patterns for both mobile (MVVM/MVC) and backend microservices.
  • Infrastructure: Experience with Google Cloud Platform and IoT telemetry (mandatory).
  • Execution: Proactive attitude toward learning and the ability to "own" a feature from DB schema to UI implementation.
  • Experience with Atlassian Jira

 

Soft skills:

·      Self-Directed Ownership: flags blockers early and suggests improvements without being asked. You are well experienced professional... You don't wait for a Jira ticket to be perfect; you ask the right questions and move the needle forward

·      Transparency: Extreme honesty about timelines—if a task is more complex than estimated, you communicate it immediately, not at the deadline.

·      Clear communicator with engineers and non-technical stakeholders.

 

The Deal

  • Part-time Retainer: 100 hours per month.
  • Rate: $15 – $18 USD per hour (Performance-based).
  • Impact: Direct partnership with the founding team in a fast-paced, AI-driven startup.
  • Location: We value the stability and focus of Tier-2 rockstars Kochi, Indore, Jaipur, or Ahmedabad and alike.

How to Apply If you are a self-starter who codes with AI and can bridge the gap between frontend and backend, send your resume and links to your 3 best live apps

Read more
Product company

Product company

Agency job
via Trinity consulting by Priyanka G
Remote, Bengaluru (Bangalore)
3 - 7 yrs
₹4L - ₹8L / yr
Artificial Intelligence (AI)
skill iconMachine Learning (ML)
Retrieval Augmented Generation (RAG)
skill iconDocker
skill iconKubernetes
+1 more

Experience: 3+ years


Responsibilities:


  • Build, train and fine tune ML models
  • Develop features to improve model accuracy and outcomes.
  • Deploy models into production using Docker, kubernetes and cloud services.
  • Proficiency in Python, MLops, expertise in data processing and large scale data set.
  • Hands on experience in Cloud AI/ML services.
  • Exposure in RAG Architecture
Read more
Proximity Works

at Proximity Works

1 video
5 recruiters
Eman Khan
Posted by Eman Khan
Remote only
4 - 8 yrs
Upto ₹40L / yr (Varies
)
skill iconScala
skill iconJava
skill iconPython
Apache Spark
Apache Hive
+6 more

We are looking for a highly skilled Data Engineer. The ideal candidate will have hands-on expertise in Big Data technologies, with a strong foundation in distributed data processing, real-time pipelines, and large-scale data systems. You will design, build, and optimize data solutions that power insights, reporting, and decision-making for our advertising ecosystem.


Key Responsibilities

  • Design, develop, and maintain large-scale data pipelines to support Ads reporting, attribution, and analytics use cases.
  • Work extensively with Hive, Spark, SQL, Scala, and Kafka to process and manage petabyte-scale datasets.
  • Optimize data workflows for performance, scalability, and cost efficiency.
  • Partner with data scientists, product managers, and platform engineers to deliver high-quality, reliable datasets and APIs.
  • Ensure data quality, integrity, and consistency across multiple data sources.
  • Troubleshoot and resolve issues in real-time streaming pipelines and batch data jobs.
  • Continuously evaluate new technologies to enhance the Ads Data platform.

Requirements

  • Strong programming experience in Scala (preferred), Java, or Python.
  • Hands-on experience with Apache Spark (batch & streaming) for large-scale data processing.
  • Proficiency in Hive, SQL, and data modeling for analytical workloads.
  • Experience working with Kafka for real-time event streaming.
  • Solid understanding of Big Data ecosystems (S4, Hive, Presto, Delta etc.).
  • Strong debugging, performance tuning, and problem-solving skills.
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or related field.

Nice to Have

  • Experience in AdTech, Attribution, or Campaign Analytics.
  • Familiarity with cloud-based big data solutions (AWS EMR, GCP BigQuery, Databricks, etc.).
  • Familiarity with scheduling services like AirFlow
  • Knowledge of data governance, security, and compliance best practices.

Benefits

  • Best in class salary: We hire only the best, and we pay accordingly.
  • Proximity Talks: Meet other designers, engineers, and product geeks — and learn from experts in the field.
  • Keep on learning with a world-class team: Work with the best in the field, challenge yourself constantly, and learn something new every day.
Read more
ByteFoundry AI

at ByteFoundry AI

4 candid answers
Bisman Gill
Posted by Bisman Gill
Remote only
3 - 8 yrs
Upto ₹40L / yr (Varies
)
skill iconReact.js
skill iconNodeJS (Node.js)
skill iconPython
SQL
skill iconAmazon Web Services (AWS)
+3 more

About the Role

We are looking for a motivated Full Stack Developer with 2–5 years of hands-on experience in building scalable web applications. You will work closely with senior engineers and product teams to develop new features, improve system performance, and ensure high-

quality code delivery.

Responsibilities

- Develop and maintain full-stack applications.

- Implement clean, maintainable, and efficient code.

- Collaborate with designers, product managers, and backend engineers.

- Participate in code reviews and debugging.

- Work with REST APIs/GraphQL.

- Contribute to CI/CD pipelines.

- Ability to work independently as well as within a collaborative team environment.


Required Technical Skills

- Strong knowledge of JavaScript/TypeScript.

- Experience with React.js, Next.js.

- Backend experience with Node.js, Express, NestJS.

- Understanding of SQL/NoSQL databases.

- Experience with Git, APIs, debugging tools.ß

- Cloud familiarity (AWS/GCP/Azure).

AI and System Mindset

Experience working with AI-powered systems is a strong plus. Candidates should be comfortable integrating AI agents, third-party APIs, and automation workflows into applications, and should demonstrate curiosity and adaptability toward emerging AI technologies.

Soft Skills

- Strong problem-solving ability.

- Good communication and teamwork.

- Fast learner and adaptable.

Education

Bachelor's degree in Computer Science / Engineering or equivalent.

Read more
OIP Insurtech

at OIP Insurtech

2 candid answers
Katarina Vasic
Posted by Katarina Vasic
Remote, Hyderabad
3 - 7 yrs
₹11L - ₹15L / yr
skill iconPython
skill iconDjango
API
skill iconFlask
RESTFUL

OIP Insurtech is an innovative business technology company founded on the belief that one company can transform the insurance industry. Our mission is to deliver outstanding and forward-thinking tech solutions that redefine how insurance processes are managed. We provide a wide range of services, including full-cycle automation of insurance processes, software development, staff augmentation, and IT support.

We foster a safe, agile environment where every idea is valued and encouraged, making it an ideal place for developers passionate about building innovative solutions.


Job Overview:

We are looking for a skilled Python Developer to join our Insurtech team. You will be responsible for developing high-quality, scalable software solutions, collaborating with cross-functional teams, and contributing to the full software development lifecycle.


Key Responsibilities:

  • Write clean, efficient, and maintainable Python code.
  • Develop, test, and deploy desktop/web applications, APIs, or backend systems.
  • Collaborate with product managers, designers, and other developers to understand requirements and deliver solutions.
  • Troubleshoot, debug, and optimize existing applications.
  • Maintain code quality through code reviews, testing, and documentation.
  • Stay up-to-date with Python frameworks, libraries, and best practices.


Required Skills and Qualifications:

  • Strong proficiency in Python and its standard libraries.
  • Experience with web frameworks such as Django, Flask, or FastAPI.
  • Experience with data analysis/engineering principles.
  • Familiarity with RESTful APIs, database systems (SQL/NoSQL), and ORM tools.
  • Understanding of version control systems (Git).
  • Knowledge of front-end technologies (HTML, CSS, JavaScript) is a plus.
  • Excellent problem-solving skills and attention to detail.
  • Strong communication and collaboration abilities.

Preferred Qualifications:

  • Experience with cloud platforms (AWS, Azure, GCP).
  • Knowledge of containerization tools like Docker or Kubernetes.
  • Experience with testing frameworks (PyTest, Unittest).
  • Understanding of Agile development methodologies.
Read more
Fluxon

at Fluxon

3 candid answers
Ariba Khan
Posted by Ariba Khan
Remote only
6 - 12 yrs
₹30L - ₹60L / yr
skill iconNodeJS (Node.js)
skill iconPython
skill iconGo Programming (Golang)
skill iconJava
skill iconReact.js
+1 more

About the company:

We are Fluxon, a product development team founded by ex-Googlers and startup founders. We offer full-cycle software development: from ideation and design to build and go-to-market. We partner with visionary companies, ranging from fast-growing startups to tech leaders like Google and Stripe, to turn bold ideas into products with the power to transform the world. 


Qualifications:

  • 7+ years of industry experience in software development
  • Experience leading development through the full product lifecycle, including CI/CD, testing, release management, deployment, monitoring and incident response
  • Fluent in the design and implementation of scalable system architectures, data structures and algorithms, and effective development practices


About the role:

As a Staff Software Engineer at Fluxon, you'll drive the end-to-end delivery of products to market while learning, contributing, and growing in partnership with our leadership team

You'll be responsible for:

  • Guiding project delivery all the way to the user, leading projects, building and iterating in a dynamic environment
  • Partnering directly with clients to understand their needs, and achieve business goals
  • Defining product requirements, identifying appropriate system designs and planning development in partnership with our Product and Design teams
  • Supporting development of a healthy and effective engineering culture

You'll work with a diversity of technologies, including:

  • Languages
  • TypeScript/JavaScript, Java, .Net, Python, Golang, Rust, Ruby on Rails, Kotlin, Swift
  • Frameworks
  • Next.js, React, Angular, Spring, Expo, FastAPI, Django, SwiftUI
  • Cloud Service Providers
  • Google Cloud Platform, Amazon Web Services, Microsoft Azure
  • Cloud Services
  • Compute Engine, AWS Amplify, Fargate, Cloud Run
  • Apache Kafka, SQS, GCP CMS
  • S3, GCS
  • Technologies
  • AI/ML, LLMs, Crypto, SPA, Mobile apps, Architecture redesign 
  • Google Gemini, OpenAI ChatGPT, Vertex AI, Anthropic Claude, Huggingface
  • Databases
  • Firestore(Firebase), PostgreSQL, MariaDB, BigQuery, Supabase
  • Redis, Memcache


What we offer:

  • Exposure to high-profile SV startups and enterprise companies
  • Competitive salary
  • Fully remote work with flexible hours
  • Flexible paid time off
  • Profit-sharing program
  • Healthcare
  • Parental leave, including adoption and fostering
  • Gym membership and tuition reimbursement
  • Hands-on career development
Read more
Procedure

at Procedure

4 candid answers
3 recruiters
Adithya K
Posted by Adithya K
Remote only
5 - 10 yrs
₹40L - ₹60L / yr
Software Development
skill iconAmazon Web Services (AWS)
skill iconPython
TypeScript
skill iconPostgreSQL
+3 more

Procedure is hiring for Drover.


This is not a DevOps/SRE/cloud-migration role — this is a hands-on backend engineering and architecture role where you build the platform powering our hardware at scale.


About Drover

Ranching is getting harder. Increased labor costs and a volatile climate are placing mounting pressure to provide for a growing population. Drover is empowering ranchers to efficiently and sustainably feed the world by making it cheaper and easier to manage livestock, unlock productivity gains, and reduce carbon footprint with rotational grazing. Not only is this a $46B opportunity, you'll be working on a climate solution with the potential for real, meaningful impact.


We use patent-pending low-voltage electrical muscle stimulation (EMS) to steer and contain cows, replacing the need for physical fences or electric shock. We are building something that has never been done before, and we have hundreds of ranches on our waitlist.


Drover is founded by Callum Taylor (ex-Harvard), who comes from 5 generations of ranching, and Samuel Aubin, both of whom grew up in Australian ranching towns and have an intricate understanding of the problem space. We are well-funded and supported by Workshop Ventures, a VC firm with experience in building unicorn IoT companies.


We're looking to assemble a team of exceptional talent with a high eagerness to dive headfirst into understanding the challenges and opportunities within ranching.


About The Role

As our founding cloud engineer, you will be responsible for building and scaling the infrastructure that powers our IoT platform, connecting thousands of devices across ranches nationwide.


Because we are an early-stage startup, you will have high levels of ownership in what you build. You will play a pivotal part in architecting our cloud infrastructure, building robust APIs, and ensuring our systems can scale reliably. We are looking for someone who is excited about solving complex technical challenges at the intersection of IoT, agriculture, and cloud computing.


What You'll Do

  • Develop Drover IoT cloud architecture from the ground up (it’s a green field project)
  • Design and implement services to support wearable devices, mobile app, and backend API
  • Implement data processing and storage pipelines
  • Create and maintain Infrastructure-as-Code
  • Support the engineering team across all aspects of early-stage development -- after all, this is a startup


Requirements

  • 5+ years of experience developing cloud architecture on AWS
  • In-depth understanding of various AWS services, especially those related to IoT
  • Expertise in cloud-hosted, event-driven, serverless architectures
  • Expertise in programming languages suitable for AWS micro-services (eg: TypeScript, Python)
  • Experience with networking and socket programming
  • Experience with Kubernetes or similar orchestration platforms
  • Experience with Infrastructure-as-Code tools (e.g., Terraform, AWS CDK)
  • Familiarity with relational databases (PostgreSQL)
  • Familiarity with Continuous Integration and Continuous Deployment (CI/CD)


Nice To Have

  • Bachelor’s or Master’s degree in Computer Science, Software Engineering, Electrical Engineering, or a related field


Read more
Sun King

at Sun King

2 candid answers
1 video
Reshika Mendiratta
Posted by Reshika Mendiratta
Remote only
2yrs+
Best in industry
Test Automation (QA)
Software Testing (QA)
Manual testing
skill iconPython
skill iconJava
+8 more

About Sun King

Sun King is the world’s leading off-grid solar energy company, providing affordable solar solutions to the 1.8 billion people without reliable access to electricity. By combining product design, fintech, and field operations, Sun King has connected over 20 million homes to solar power across Africa and Asia, adding more than 200,000 new homes each month. Through ‘pay-as-you-go’ financing, customers make small payments to eventually own their solar systems, saving money and reducing reliance on harmful energy sources like kerosene.


Sun King employs 2,800 staff across 12 countries, with expertise in product design, data science, logistics, customer service, and more. The company is expanding its product range to include clean cooking, electric mobility, and entertainment solutions, all while supporting a diverse workforce — with women making up 44% of the team.


About the role:

The role involves designing, executing, and maintaining robust functional, regression, and integration testing to ensure product quality and reliability, along with thorough defect tracking, analysis, and resolution. The individual will develop and maintain UI and API automation frameworks to improve test coverage, minimize manual effort, and enhance release efficiency. Close collaboration with development teams is expected to reproduce issues, validate fixes, and ensure high-quality releases. The role also includes integrating automated tests into CI/CD pipelines, supporting production issue analysis, and verifying hotfixes in live environments. Additionally, the candidate will actively participate in requirement and design reviews to ensure testability and clarity, maintain comprehensive QA documentation, and continuously improve testing frameworks, tools, and overall QA processes.


What you will be expected to do:

  • Design, execute, and maintain test cases, test plans, and test scripts for functional, regression, and integration testing.
  • Identify software defects, document them clearly, and track them through to closure.
  • Analyze bugs and provide detailed insights to help developers understand root causes.
  • Partner closely with the development team to reproduce issues, validate fixes, and ensure overall product quality.
  • Develop, maintain, and improve automated test suites (API/UI) to enhance test coverage, reduce manual effort, and improve release confidence.
  • Work with CI/CD pipelines to integrate automated tests into the deployment workflow.
  • Validate production issues, support troubleshooting, and verify hotfixes in real-time environments.
  • Recommend improvements in product performance, usability, and reliability based on test findings.
  • Participate in requirement and design reviews to ensure clarity, completeness, and testability.
  • Benchmark against competitor products and suggest enhancements based on industry trends.
  • Maintain detailed test documentation, including test results, defect logs, and release readiness assessments.
  • Continuously improve QA processes, automation frameworks, and testing methodologies.

You might be a strong candidate if you have/are:

  • Bachelor’s Degree in Computer Science, Information Technology, or a related field.
  • 2+ years of hands-on experience in software testing (manual + exposure to automation).
  • Strong understanding of QA methodologies, testing types, and best practices.
  • Experience in designing and executing test cases, test plans, and regression suites.
  • Exposure to automation tools/frameworks such as Selenium, Playwright, Cypress, TestNG, JUnit, or similar.
  • Basic programming or scripting knowledge (Java/Python preferred).
  • Good understanding of SQL for backend and data validation testing.
  • Familiarity with API testing tools such as Postman or RestAssured.
  • Experience with defect tracking and test management tools (Jira, TestRail, etc.).
  • Strong analytical and debugging skills with the ability to identify root causes.
  • Ability to work effectively in Agile/Scrum environments and partner with developers, product, and DevOps teams.
  • Strong ownership mindset — having contributed to high-quality, near bug-free releases.

Good to have:

  • Exceptional attention to detail and a strong focus on product quality.
  • Experience with performance, load, or security testing (JMeter, Gatling, OWASP tools, etc.).
  • Exposure to advanced automation frameworks or building automation scripts from scratch.
  • Familiarity with CI/CD pipelines and integrating automated tests.
  • Experience working with observability tools like Grafana, Kibana, and Prometheus for production verification.
  • Good understanding of microservices, distributed systems, or cloud platforms.

What Sun King offers:

  • Professional growth in a dynamic, rapidly expanding, high-social-impact industry
  • An open-minded, collaborative culture made up of enthusiastic colleagues who are driven by the challenge of innovation towards profound impact on people and the planet.
  • A truly multicultural experience: you will have the chance to work with and learn from people from different geographies, nationalities, and backgrounds.
  • Structured, tailored learning and development programs that help you become a better leader, manager, and professional through the Sun King Center for Leadership.
Read more
Koolioai
Aishwaria SterlingJames
Posted by Aishwaria SterlingJames
Remote only
0 - 1 yrs
₹15000 - ₹20000 / mo
skill iconPython
skill iconReact.js
skill iconHTML/CSS
skill iconJavascript
skill iconRedux/Flux
+2 more

About koolio.ai


Website: www.koolio.ai


Koolio Inc. is a cutting-edge Silicon Valley startup dedicated to transforming how stories are told through audio. Our mission is to democratize audio content creation by empowering individuals and businesses to effortlessly produce high-quality, professional-grade content. Leveraging AI and intuitive web-based tools, koolio.ai enables creators to craft, edit, and distribute audio content—from storytelling to educational materials, brand marketing, and beyond. We are passionate about helping people and organizations share their voices, fostering creativity, collaboration, and engaging storytelling for a wide range of use cases.


About the Internship Position

We are looking for a motivated full-time Software Development Intern to join our innovative team. As an intern at koolio.ai, you’ll have the opportunity to work on a next-gen AI-powered platform and gain hands-on experience developing and optimizing backend systems that power our platform. This internship is ideal for students or recent graduates who are passionate about backend technologies and eager to learn in a dynamic, fast-paced startup environment.


Key Responsibilities:

  • Assist in the development and maintenance of backend systems and APIs.
  • Write reusable, testable, and efficient code to support scalable web applications.
  • Work with cloud services and server-side technologies to manage data and optimize performance.
  • Collaborate with cross-functional teams to integrate frontend features with backend logic.
  • Collaborate with the product and design teams to develop and implement user-friendly web interfaces
  • Ensure responsive design and optimize web pages for performance and scalability across devices
  • Debug and resolve issues, improving the overall user experience on the platform and ensuring reliability
  • Assist in integrating APIs and frontend services with the backend
  • Stay up-to-date with the latest trends and suggest improvements to enhance the platform’s functionality


Requirements and Skills:

  • Education: Currently pursuing or recently completed a degree in Computer Science, Engineering, or a related field.
  • Technical Skills:
  • Good understanding of server-side technologies like Python
  • Familiarity with REST APIs and database systems (e.g., MySQL, PostgreSQL, or NoSQL databases).
  • Exposure to cloud platforms like AWS, Google Cloud, or Azure is a plus.
  • Knowledge of version control systems such as Git.
  • Strong proficiency in JavaScript frameworks like ReactJS or Redux
  • Proficiency in frontend languages such as HTML, CSS, and JavaScript (ES6+)
  • Soft Skills:
  • Eagerness to learn and adapt in a fast-paced environment.
  • Strong problem-solving and critical-thinking skills.
  • Effective communication and teamwork capabilities.
  • Other Skills: Familiarity with CI/CD pipelines and basic knowledge of containerization (e.g., Docker) is a bonus.


Why Join Us?

  • Gain real-world experience working on a cutting-edge platform.
  • Work alongside a talented and passionate team committed to innovation.
  • Receive mentorship and guidance from industry experts.
  • Opportunity to transition to a full-time role based on performance and company needs.


This internship is an excellent opportunity to kickstart your career in software development, build critical skills, and contribute to a product that has a real-world impact.

Read more
KGiSL MICROCOLLEGE
Hiring Recruitment
Posted by Hiring Recruitment
Remote, Erode
1 - 3 yrs
₹1L - ₹3.5L / yr
skill iconHTML/CSS
JAWA
skill iconPython
Artificial Intelligence (AI)

Salary: ₹3.5 LPA( Based on the performance)

Experience: 1–3 Years (ONLY FOR FEMALES)


We are looking for a Technical Trainer skilled in HTML, Java, Python, and AI to conduct technical trainer. The trainer will create learning materials, deliver sessions, assess student performance, and support learners throughout the training. Strong communication skills and the ability to explain technical concepts clearly are essential.

Read more
Capital Squared
Remote only
5 - 10 yrs
₹25L - ₹55L / yr
MLOps
DevOps
Google Cloud Platform (GCP)
CI/CD
skill iconPostgreSQL
+4 more

Role: Full-Time, Long-Term Required: Docker, GCP, CI/CD Preferred: Experience with ML pipelines


OVERVIEW

We are seeking a DevOps engineer to join as a core member of our technical team. This is a long-term position for someone who wants to own infrastructure and deployment for a production machine learning system. You will ensure our prediction pipeline runs reliably, deploys smoothly, and scales as needed.


The ideal candidate thinks about failure modes obsessively, automates everything possible, and builds systems that run without constant attention.


CORE TECHNICAL REQUIREMENTS

Docker (Required): Deep experience with containerization. Efficient Dockerfiles, layer caching, multi-stage builds, debugging container issues. Experience with Docker Compose for local development.


Google Cloud Platform (Required): Strong GCP experience: Cloud Run for serverless containers, Compute Engine for VMs, Artifact Registry for images, Cloud Storage, IAM. You can navigate the console but prefer scripting everything.


CI/CD (Required): Build and maintain deployment pipelines. GitHub Actions required. You automate testing, building, pushing, and deploying. You understand the difference between continuous integration and continuous deployment.


Linux Administration (Required): Comfortable on the command line. SSH, diagnose problems, manage services, read logs, fix things. Bash scripting is second nature.


PostgreSQL (Required): Database administration basics—backups, monitoring, connection management, basic performance tuning. Not a DBA, but comfortable keeping a production database healthy.


Infrastructure as Code (Preferred): Terraform, Pulumi, or similar. Infrastructure should be versioned, reviewed, and reproducible—not clicked together in a console.


WHAT YOU WILL OWN

Deployment Pipeline: Maintaining and improving deployment scripts and CI/CD workflows. Code moves from commit to production reliably with appropriate testing gates.


Cloud Run Services: Managing deployments for model fitting, data cleansing, and signal discovery services. Monitor health, optimize cold starts, handle scaling.


VM Infrastructure: PostgreSQL and Streamlit on GCP VMs. Instance management, updates, backups, security.


Container Registry: Managing images in GitHub Container Registry and Google Artifact Registry. Cleanup policies, versioning, access control.


Monitoring and Alerting: Building observability. Logging, metrics, health checks, alerting. Know when things break before users tell us.


Environment Management: Configuration across local and production. Secrets management. Environment parity where it matters.


WHAT SUCCESS LOOKS LIKE

Deployments are boring—no drama, no surprises. Systems recover automatically from transient failures. Engineers deploy with confidence. Infrastructure changes are versioned and reproducible. Costs are reasonable and resources scale appropriately.


ENGINEERING STANDARDS

Automation First: If you do something twice, automate it. Manual processes are bugs waiting to happen.


Documentation: Runbooks, architecture diagrams, deployment guides. The next person can understand and operate the system.


Security Mindset: Secrets never in code. Least-privilege access. You think about attack surfaces.


Reliability Focus: Design for failure. Backups are tested. Recovery procedures exist and work.


CURRENT ENVIRONMENT

GCP (Cloud Run, Compute Engine, Artifact Registry, Cloud Storage), Docker, Docker Compose, GitHub Actions, PostgreSQL 16, Bash deployment scripts with Python wrapper.


WHAT WE ARE LOOKING FOR

Ownership Mentality: You see a problem, you fix it. You do not wait for assignment.


Calm Under Pressure: When production breaks, you diagnose methodically.


Communication: You explain infrastructure decisions to non-infrastructure people. You document what you build.


Long-Term Thinking: You build systems maintained for years, not quick fixes creating tech debt.


EDUCATION

University degree in Computer Science, Engineering, or related field preferred. Equivalent demonstrated expertise also considered.


TO APPLY

Include: (1) CV/resume, (2) Brief description of infrastructure you built or maintained, (3) Links to relevant work if available, (4) Availability and timezone.

Read more
Remote only
5 - 10 yrs
₹25L - ₹55L / yr
Data engineering
Databases
skill iconPython
SQL
skill iconPostgreSQL
+4 more

Role: Full-Time, Long-Term Required: Python, SQL Preferred: Experience with financial or crypto data


OVERVIEW

We are seeking a data engineer to join as a core member of our technical team. This is a long-term position for someone who wants to build robust, production-grade data infrastructure and grow with a small, focused team. You will own the data layer that feeds our machine learning pipeline—from ingestion and validation through transformation, storage, and delivery.


The ideal candidate is meticulous about data quality, thinks deeply about failure modes, and builds systems that run reliably without constant attention. You understand that downstream ML models are only as good as the data they consume.


CORE TECHNICAL REQUIREMENTS

Python (Required): Professional-level proficiency. You write clean, maintainable code for data pipelines—not throwaway scripts. Comfortable with Pandas, NumPy, and their performance characteristics. You know when to use Python versus push computation to the database.


SQL (Required): Advanced SQL skills. Complex queries, query optimization, schema design, execution plans. PostgreSQL experience strongly preferred. You think about indexing, partitioning, and query performance as second nature.


Data Pipeline Design (Required): You build pipelines that handle real-world messiness gracefully. You understand idempotency, exactly-once semantics, backfill strategies, and incremental versus full recomputation tradeoffs. You design for failure—what happens when an upstream source is late, returns malformed data, or goes down entirely. Experience with workflow orchestration required: Airflow, Prefect, Dagster, or similar.


Data Quality (Required): You treat data quality as a first-class concern. You implement validation checks, anomaly detection, and monitoring. You know the difference between data that is missing versus data that should not exist. You build systems that catch problems before they propagate downstream.


WHAT YOU WILL BUILD

Data Ingestion: Pipelines pulling from diverse sources—crypto exchanges, traditional market feeds, on-chain data, alternative data. Handling rate limits, API quirks, authentication, and source-specific idiosyncrasies.


Data Validation: Checks ensuring completeness, consistency, and correctness. Schema validation, range checks, freshness monitoring, cross-source reconciliation.


Transformation Layer: Converting raw data into clean, analysis-ready formats. Time series alignment, handling different frequencies and timezones, managing gaps.


Storage and Access: Schema design optimized for both write patterns (ingestion) and read patterns (ML training, feature computation). Data lifecycle and retention management.

Monitoring and Alerting: Observability into pipeline health. Knowing when something breaks before it affects downstream systems.


DOMAIN EXPERIENCE

Preference for candidates with experience in financial or crypto data—understanding market data conventions, exchange-specific quirks, and point-in-time correctness. You know why look-ahead bias is dangerous and how to prevent it.


Time series data at scale—hundreds of symbols with years of history, multiple frequencies, derived features. You understand temporal joins, windowed computations, and time-aligned data challenges.


High-dimensional feature stores—we work with hundreds of thousands of derived features. Experience managing, versioning, and serving large feature sets is valuable.


ENGINEERING STANDARDS

Reliability: Pipelines run unattended. Failures are graceful with clear errors, not silent corruption. Recovery is straightforward.


Reproducibility: Same inputs and code version produce identical outputs. You version schemas, track lineage, and can reconstruct historical states.


Documentation: Schemas, data dictionaries, pipeline dependencies, operational runbooks. Others can understand and maintain your systems.


Testing: You write tests for pipelines—validation logic, transformation correctness, edge cases. Untested pipelines are broken pipelines waiting to happen.


TECHNICAL ENVIRONMENT

PostgreSQL, Python, workflow orchestration (flexible on tool), cloud infrastructure (GCP preferred but flexible), Git.


WHAT WE ARE LOOKING FOR

Attention to Detail: You notice when something is slightly off and investigate rather than ignore.


Defensive Thinking: You assume sources will send bad data, APIs will fail, schemas will change. You build accordingly.


Self-Direction: You identify problems, propose solutions, and execute without waiting to be told.


Long-Term Orientation: You build systems you will maintain for years.


Communication: You document clearly, explain data issues to non-engineers, and surface problems early.


EDUCATION

University degree in a quantitative/technical field preferred: Computer Science, Mathematics, Statistics, Engineering. Equivalent demonstrated expertise also considered.


TO APPLY

Include: (1) CV/resume, (2) Brief description of a data pipeline you built and maintained, (3) Links to relevant work if available, (4) Availability and timezone.

Read more
Palcode.ai

at Palcode.ai

2 candid answers
Team Palcode
Posted by Team Palcode
Remote only
1 - 2 yrs
₹4L - ₹5L / yr
skill iconPython
FastAPI
skill iconFlask

At Palcode.ai, We're on a mission to fix the massive inefficiencies in pre-construction. Think about it - in a $10 trillion industry, estimators still spend weeks analyzing bids, project managers struggle with scattered data, and costly mistakes slip through complex contracts. We're fixing this with purpose-built AI agents that work. Our platform can do “magic” to Preconstruction workflows from Weeks to Hours. It's not just about AI – it's about bringing real, measurable impact to an industry ready for change. We are backed by names like AWS for Startups, Upekkha Accelerator, and Microsoft for Startups.



Why Palcode.ai


Tackle Complex Problems: Build AI that reads between the lines of construction bids, spots hidden risks in contracts, and makes sense of fragmented project data

High-Impact Code: Your code won't sit in a backlog – it goes straight to estimators and project managers who need it yesterday

Tech Challenges That Matter: Design systems that process thousands of construction documents, handle real-time pricing data, and make intelligent decisions

Build & Own: Shape our entire tech stack, from data processing pipelines to AI model deployment

Quick Impact: Small team, huge responsibility. Your solutions directly impact project decisions worth millions

Learn & Grow: Master the intersection of AI, cloud architecture, and construction tech while working with founders who've built and scaled construction software


Your Role:

  • Design and build our core AI services and APIs using Python
  • Create reliable, scalable backend systems that handle complex data
  • Help set up cloud infrastructure and deployment pipelines
  • Collaborate with our AI team to integrate machine learning models
  • Write clean, tested, production-ready code


You'll fit right in if:

  • You have 1 year of hands-on Python development experience
  • You're comfortable with full-stack development and cloud services
  • You write clean, maintainable code and follow good engineering practices
  • You're curious about AI/ML and eager to learn new technologies
  • You enjoy fast-paced startup environments and take ownership of your work


How we will set you up for success

  • You will work closely with the Founding team to understand what we are building.
  • You will be given comprehensive training about the tech stack, with an opportunity to avail virtual training as well.
  • You will be involved in a monthly one-on-one with the founders to discuss feedback
  • A unique opportunity to learn from the best - we are Gold partners of AWS, Razorpay, and Microsoft Startup programs, having access to rich talent to discuss and brainstorm ideas.
  • You’ll have a lot of creative freedom to execute new ideas. As long as you can convince us, and you’re confident in your skills, we’re here to back you in your execution.


Location: Bangalore, Remote


Compensation: Competitive salary + Meaningful equity


If you get excited about solving hard problems that have real-world impact, we should talk.


All the best!!

Read less


Read more
Hashone Careers
Madhavan I
Posted by Madhavan I
Remote only
5 - 10 yrs
₹20L - ₹40L / yr
DevOps
skill iconAmazon Web Services (AWS)
skill iconKubernetes
cicd
skill iconPython
+1 more

Job Description: DevOps Engineer

Location: Bangalore / Hybrid / Remote

Company: LodgIQ

Industry: Hospitality / SaaS / Machine Learning


About LodgIQ

Headquartered in New York, LodgIQ delivers a revolutionary B2B SaaS platform to the

travel industry. By leveraging machine learning and artificial intelligence, we enable precise

forecasting and optimized pricing for hotel revenue management. Backed by Highgate

Ventures and Trilantic Capital Partners, LodgIQ is a well-funded, high-growth startup with a

global presence.


Role Summary:

We are seeking a Senior DevOps Engineer with 5+ years of strong hands-on experience in

AWS, Kubernetes, CI/CD, infrastructure as code, and cloud-native technologies. This

role involves designing and implementing scalable infrastructure, improving system

reliability, and driving automation across our cloud ecosystem.


Key Responsibilities:

• Architect, implement, and manage scalable, secure, and resilient cloud

infrastructure on AWS

• Lead DevOps initiatives including CI/CD pipelines, infrastructure automation,

and monitoring

• Deploy and manage Kubernetes clusters and containerized microservices

• Define and implement infrastructure as code using

Terraform/CloudFormation

• Monitor production and staging environments using tools like CloudWatch,

Prometheus, and Grafana

• Support MongoDB and MySQL database administration and optimization

• Ensure high availability, performance tuning, and cost optimization

• Guide and mentor junior engineers, and enforce DevOps best practices

• Drive system security, compliance, and audit readiness in cloud environments

• Collaborate with engineering, product, and QA teams to streamline release

processes


Required Qualifications:

• 5+ years of DevOps/Infrastructure experience in production-grade environments

• Strong expertise in AWS services: EC2, EKS, IAM, S3, RDS, Lambda, VPC, etc.

• Proven experience with Kubernetes and Docker in production

• Proficient with Terraform, CloudFormation, or similar IaC tools

• Hands-on experience with CI/CD pipelines using Jenkins, GitHub Actions, or

similar

• Advanced scripting in Python, Bash, or Go

• Solid understanding of networking, firewalls, DNS, and security protocols

• Exposure to monitoring and logging stacks (e.g., ELK, Prometheus, Grafana)

• Experience with MongoDB and MySQL in cloud environments

Preferred Qualifications:

• AWS Certified DevOps Engineer or Solutions Architect

• Experience with service mesh (Istio, Linkerd), Helm, or ArgoCD

• Familiarity with Zero Downtime Deployments, Canary Releases, and Blue/Green

Deployments

• Background in high-availability systems and incident response

• Prior experience in a SaaS, ML, or hospitality-tech environment


Tools and Technologies You’ll Use:

• Cloud: AWS

• Containers: Docker, Kubernetes, Helm

• CI/CD: Jenkins, GitHub Actions

• IaC: Terraform, CloudFormation

• Monitoring: Prometheus, Grafana, CloudWatch

• Databases: MongoDB, MySQL

• Scripting: Bash, Python

• Collaboration: Git, Jira, Confluence, Slack


Why Join Us?

• Competitive salary and performance bonuses.

• Remote-friendly work culture.

• Opportunity to work on cutting-edge tech in AI and ML.

• Collaborative, high-growth startup environment.

• For more information, visit http://www.lodgiq.com

Read more
Analytical Brains Education
Remote only
1 - 5 yrs
₹8L - ₹12L / yr
skill iconPython
Shell Scripting
Powershell
SQL
skill iconJava

Job Description

We are looking for motivated IT professionals with at least one year of industry experience. The ideal candidate should have hands-on experience in AWS, Azure, AI, or Cloud technologies, or should be enthusiastic and ready to upskill and shift to new and emerging technologies. This role is primarily remote; however, candidates may be required to visit the office occasionally for meetings or project needs.

Key Requirements

  • Minimum 1 year of experience in the IT industry
  • Exposure to AWS / Azure / AI / Cloud platforms (any one or more)
  • Willingness to learn and adapt to new technologies
  • Strong problem-solving and communication skills
  • Ability to work independently in a remote setup
  • Must have a proper work-from-home environment (laptop, stable internet, quiet workspace)

Education Qualification

  • B.Tech / BE / MCA / M.Sc (IT) / equivalent


Read more
EDUPOWERPO SOLUTIONS

at EDUPOWERPO SOLUTIONS

1 candid answer
CHANDAN RAI
Posted by CHANDAN RAI
Remote only
1 - 2 yrs
₹1.2L - ₹2L / yr
SEO management
skill iconHTML/CSS
skill iconPython
skill iconDjango

What We’re Looking For

• Hands-on experience in keyword research, competitor analysis and content gap identification

• Strong understanding of on-page SEO: meta tags, schema, internal linking, URL structure and content optimization

• Experience managing technical SEO: site audits, crawling issues, indexing, page speed and mobile optimization

• Ability to plan and execute backlink strategies using safe, high-quality methods

• Familiarity with tools like Google Search Console, Google Analytics, Ahrefs, SEMrush, Screaming Frog

• Experience working with blogs, landing pages and long-form content

• Ability to coordinate with writers, developers and designers for SEO requirements

• Proven experience in improving rankings for competitive keywords

• Understanding of local SEO and structured data markup

• Comfortable working in a fast-moving, bootstrapped startup environment

• Bonus: Experience with Django or basic HTML/CSS is useful but not mandatory

What You Will Work On

• Improving rankings for keywords like “Top boarding schools in India”, “Best boarding schools”, etc.

• Conducting monthly audits and pushing technical SEO fixes

• Growing EduPowerPro’s organic traffic through structured content planning

• Managing backlink acquisition and partnerships

• Tracking performance and presenting monthly insights

Read more
Mira Network

at Mira Network

1 candid answer
Nikita Sinha
Posted by Nikita Sinha
Remote only
5 - 10 yrs
Upto ₹70L / yr (Varies
)
Systems design
skill iconGo Programming (Golang)
skill iconRust
skill iconReact.js
skill iconNodeJS (Node.js)
+2 more

Mira is building the foundational trust and verification layer for agentic commerce - the emerging landscape where autonomous AI agents interact, transact, and deliver value across modern digital systems. Our work extends into next-generation consumer finance, blending intelligent automation, verifiable execution, and new forms of digital value movement.


We operate as a senior, high-caliber engineering team that ships foundational infrastructure for intelligent consumer experiences, where correctness, reliability, and clarity of execution matter deeply.

We are building a consumer-facing financial product at Mira.


Think crypto-native neobank + agentic automation, where large parts of money movement, compliance, and execution are stitched together via existing providers - but orchestrated with strong guarantees, clear invariants, and excellent user experience.

This is a 0 to 1 product. Architecture decisions made early will compound for years.



We’re looking for a Lead Architect / Lead Full-Stack Engineer to act as a technical owner for this product.

You will define and build the core product architecture across backend, integrations, and frontend surfaces. This role is not about inventing new primitives - it’s about correctly and safely stitching together high-risk systems (money movement, KYC, wallets, verification, agent flows) into a coherent, scalable product.


You’ll work closely with product leadership to shape the technical direction, while staying deeply hands-on in production code and also support in building and leading the technical team.


WHAT YOU WILL DO:

Your primary responsibility is to design, build, and own the core full-stack system end to end.

Specifically, you will:

  • Architect and implement backend systems for correctness-critical workflows, including ledgers, balances, transaction state, and orchestration logic.
  • Design and build integration-heavy systems, stitching together KYC, wallet providers, verification services, vaults, and external financial APIs.
  • Own system invariants and failure modes: ensuring money movement, retries, reconciliation, and edge cases behave safely under stress.
  • Build and evolve agentic automation flows that coordinate execution across multiple systems while remaining observable and debuggable.
  • Develop core frontend surfaces (web, admin, internal tools) and collaborate closely with React Native engineers for the consumer app.
  • Set up and maintain DevOps foundations: environments, CI/CD, monitoring, alerts, and operational playbooks.
  • Act as a technical decision-maker and mentor, raising the bar for system design, code quality, and reliability across the team.
  • Collaborate with the broader engineering and product team on shared primitives around verification, correctness, and trust.


WHAT YOU BRING:

You are a senior engineer who has built and owned real systems where failure is expensive.

You likely have:

  • 5–8+ years of full-stack engineering experience, with strong depth on the backend.
  • Prior experience in consumer fintech, payments, wallets, or financial infrastructure, where ledgers and state correctness mattered.
  • Deep experience with schema design, data modeling, consistency models, and fault-tolerant systems.
  • Strong systems thinking: you naturally reason about invariants, race conditions, retries, idempotency, and data integrity.
  • Hands-on experience integrating and operating third-party APIs in production (KYC, payments, identity, compliance, etc.).
  • Solid DevOps instincts - comfortable owning deployments, infra decisions, and operational reliability.
  • Frontend experience with modern JS/TypeScript frameworks (React); enough to build and reason about product UX, even if you’re backend-leaning.
  • Comfort operating in ambiguity and 0→1 environments, where the problem is still being shaped alongside the solution.
  • A strong sense of ownership - you don’t wait for specs; you help define them.

NICE TO HAVE:

  • Experience working with crypto or blockchain-adjacent systems, even if via integrations rather than protocol design.
  • Familiarity with React Native or close collaboration with mobile teams.
  • Prior work building consumer-scale systems or financial-grade infrastructure.
  • Prior exposure to agent-like workflows, automation engines, or distributed orchestration systems.

WHY THIS ROLE MATTERS:

This is a keystone hire.

The person in this role will:

  • Shape how money moves through the system
  • Define early architectural patterns
  • Prevent painful rewrites later by getting fundamentals right now

If you enjoy building serious systems with real-world consequences - and want to operate at founder-level ownership inside a small, elite team - this role will stretch and reward you.


WHAT WE OFFER:

  • Competitive compensation.
  • High ownership and the opportunity to shape product direction.
  • Direct impact on foundational cryptographic and blockchain infrastructure.
  • A collaborative team that values clarity, autonomy, and velocity.


Note: This role can be remote; however Bengaluru or Mumbai candidates will be prioritized.

Read more
ConvertLens

at ConvertLens

2 candid answers
3 recruiters
Reshika Mendiratta
Posted by Reshika Mendiratta
Remote, Noida
2yrs+
Best in industry
skill iconPython
FastAPI
AI Agents
Artificial Intelligence (AI)
Large Language Models (LLM)
+9 more

🚀 About Us

At Remedo, we're building the future of digital healthcare marketing. We help doctors grow their online presence, connect with patients, and drive real-world outcomes like higher appointment bookings and better Google reviews — all while improving their SEO.


We’re also the creators of Convertlens, our generative AI-powered engagement engine that transforms how clinics interact with patients across the web. Think hyper-personalized messaging, automated conversion funnels, and insights that actually move the needle.

We’re a lean, fast-moving team with startup DNA. If you like ownership, impact, and tech that solves real problems — you’ll fit right in.


🛠️ What You’ll Do

  • Build and maintain scalable Python back-end systems that power Convertlens and internal applications.
  • Develop Agentic AI applications and workflows to drive automation and insights.
  • Design and implement connectors to third-party systems (APIs, CRMs, marketing tools) to source and unify data.
  • Ensure system reliability with strong practices in observability, monitoring, and troubleshooting.


⚙️ What You Bring

  • 2+ years of hands-on experience in Python back-end development.
  • Strong understanding of REST API design and integration.
  • Proficiency with relational databases (MySQL/PostgreSQL).
  • Familiarity with observability tools (logging, monitoring, tracing — e.g., OpenTelemetry, Prometheus, Grafana, ELK).
  • Experience maintaining production systems with a focus on reliability and scalability.
  • Bonus: Exposure to Node.js and modern front-end frameworks like ReactJs.
  • Strong problem-solving skills and comfort working in a startup/product environment.
  • A builder mindset — scrappy, curious, and ready to ship.


💼 Perks & Culture

  • Flexible work setup — remote-first for most, hybrid if you’re in Delhi NCR.
  • A high-growth, high-impact environment where your code goes live fast.
  • Opportunities to work with Agentic AI and cutting-edge tech.
  • Small team, big vision — your work truly matters here.
Read more
Automate Accounts

at Automate Accounts

2 candid answers
Namrata Das
Posted by Namrata Das
Remote only
2 - 6 yrs
₹6L - ₹15L / yr
skill iconPython
skill iconNodeJS (Node.js)
skill iconSpring Boot
Debugging
RESTful APIs
+1 more

Responsibilities


Develop and maintain web and backend components using Python, Node.js, and Zoho tools


Design and implement custom workflows and automations in Zoho


Perform code reviews to maintain quality standards and best practices


Debug and resolve technical issues promptly


Collaborate with teams to gather and analyze requirements for effective solutions


Write clean, maintainable, and well-documented code


Manage and optimize databases to support changing business needs


Contribute individually while mentoring and supporting team members


Adapt quickly to a fast-paced environment and meet expectations within the first month



Leadership Opportunities


Lead and mentor junior developers in the team


Drive projects independently while collaborating with the broader team


Act as a technical liaison between the team and stakeholders to deliver effective solutions



Selection Process


1. HR Screening: Review of qualifications and experience


2. Online Technical Assessment: Test coding and problem-solving skills


3. Technical Interview: Assess expertise in web development, Python, Node.js, APIs, and Zoho


4. Leadership Evaluation: Evaluate team collaboration and leadership abilities


5. Management Interview: Discuss cultural fit and career opportunities


6. Offer Discussion: Finalize compensation and role specifics



Experience Required


2-6 years of relevant experience as a Software Developer


Proven ability to work as a self-starter and contribute individually


Strong technical and interpersonal skills to support team members effectively


Read more
Talent Pro
Mayank choudhary
Posted by Mayank choudhary
Remote only
3 - 5 yrs
₹20L - ₹25L / yr
skill iconPython
skill iconReact.js
skill iconNodeJS (Node.js)

Role: Technical Co-Founder


Experience: 3+ years (Mandatory)


Compensation: Equity Only (No Salary)


Requirements:


Strong full-stack development skills


Experience building web applications from scratch


Able to manage complete tech independently


Startup mindset & owne

rship attitude

Read more
CFRA

at CFRA

4 candid answers
2 recruiters
Bisman Gill
Posted by Bisman Gill
Remote only
4yrs+
Upto ₹23L / yr (Varies
)
skill iconAmazon Web Services (AWS)
SQL
skill iconPython
skill iconNodeJS (Node.js)
skill iconJava
+1 more

The Senior Software Developer is responsible for development of CFRA’s report generation framework using a modern technology stack: Python on AWS cloud infrastructure, SQL, and Web technologies. This is an opportunity to make an impact on both the team and the organization by being part of the design and development of a new customer-facing report generation framework that will serve as the foundation for all future report development at CFRA.

The ideal candidate has a passion for solving business problems with technology and can effectively communicate business and technical needs to stakeholders. We are looking for candidates that value collaboration with colleagues and having an immediate, tangible impact for a leading global independent financial insights and data company.


Key Responsibilities

  • Analyst Workflows: Design and development of CFRA’s integrated content publishing platform using a proprietary 3rd party editorial and publishing platform for integrated digital publishing.
  • Designing and Developing APIs: Design and development of robust, scalable, and secure APIs on AWS, considering factors like performance, reliability, and cost-efficiency.
  • AWS Service Integration: Integrate APIs with various AWS services such as AWS Lambda, Amazon API Gateway, Amazon SQS, Amazon SNS, AWS Glue, and others, to build comprehensive and efficient solutions.
  • Performance Optimization: Identify and implement optimizations to improve performance, scalability, and efficiency, leveraging AWS services and tools.
  • Security and Compliance: Ensure APIs are developed following best security practices, including authentication, authorization, encryption, and compliance with relevant standards and regulations.
  • Monitoring and Logging: Implement monitoring and logging solutions for APIs using AWS CloudWatch, AWS X-Ray, or similar tools, to ensure availability, performance, and reliability.
  • Continuous Integration and Deployment (CI/CD): Establish and maintain CI/CD pipelines for API development, automating testing, deployment, and monitoring processes on AWS.
  • Documentation and Training: Create and maintain comprehensive documentation for internal and external users, and provide training and support to developers and stakeholders.
  • Team Collaboration: Collaborate effectively with cross-functional teams, including product managers, designers, and other developers, to deliver high-quality solutions that meet business requirements.
  • Problem Solving: troubleshooting efforts, identifying root causes and implementing solutions to ensure system stability and performance.
  • Stay Updated: Stay updated with the latest trends, tools, and technologies related to development on AWS, and continuously improve your skills and knowledge.

Desired Skills and Experience

  • Development: 5+ years of extensive experience in designing, developing, and deploying using modern technologies, with a focus on scalability, performance, and security.
  • AWS Services: proficiency in using AWS services such as AWS Lambda, Amazon API Gateway, Amazon SQS, Amazon SNS, Amazon SES, Amazon RDS, Amazon DynamoDB, and others, to build and deploy API solutions.
  • Programming Languages: Proficiency in programming languages commonly used for development, such as Python, Node.js, or others, as well as experience with serverless frameworks like AWS.
  • Architecture Design: Ability to design scalable and resilient API architectures using microservices, serverless, or other modern architectural patterns, considering factors like performance, reliability, and cost-efficiency.
  • Security: Strong understanding of security principles and best practices, including authentication, authorization, encryption, and compliance with standards like OAuth, OpenID Connect, and AWS IAM.
  • DevOps Practices: Familiarity with DevOps practices and tools, including CI/CD pipelines, infrastructure as code (IaC), and automated testing, to ensure efficient and reliable deployment on AWS.
  • Problem-solving Skills: Excellent problem-solving skills, with the ability to troubleshoot complex issues, identify root causes, and implement effective solutions to ensure the stability and performance.
  • Communication Skills: Strong communication skills, with the ability to effectively communicate technical concepts to both technical and non-technical stakeholders, and collaborate with cross- functional teams.
  • Agile Methodologies: Experience working in Agile development environments, following practices like Scrum or Kanban, and ability to adapt to changing requirements and priorities.
  • Continuous Learning: A commitment to continuous learning and staying updated with the latest trends, tools, and technologies related to development and AWS services.
  • Bachelor's Degree: A bachelor's degree in Computer Science, Software Engineering, or a related field is often preferred, although equivalent experience and certifications can also be valuable.




Read more
CFRA

at CFRA

4 candid answers
2 recruiters
Bisman Gill
Posted by Bisman Gill
Remote only
7yrs+
Upto ₹36L / yr (Varies
)
skill iconAmazon Web Services (AWS)
SQL
skill iconPython
skill iconNodeJS (Node.js)
skill iconJava

The Lead Software Developer is responsible for development of CFRA’s report generation framework using a modern technology stack: Python on AWS cloud infrastructure, SQL, and Web technologies. This is an opportunity to make an impact on both the team and the organization by being part of the design and development of a new customer-facing report generation framework that will serve as the foundation for all future report development at CFRA.

The ideal candidate has a passion for solving business problems with technology and can effectively communicate business and technical needs to stakeholders. We are looking for candidates that value collaboration with colleagues and having an immediate, tangible impact for a leading global independent financial insights and data company.


Key Responsibilities

  • Analyst Workflows: Lead the design and development of CFRA’s integrated content publishing platform using a proprietary 3rd party editorial and publishing platform for integrated digital publishing.
  • Designing and Developing APIs: Lead the design and development of robust, scalable, and secure APIs on AWS, considering factors like performance, reliability, and cost-efficiency.
  • Architecture Planning: Collaborate with architects and stakeholders to define architecture, including API gateway, microservices, and serverless components, ensuring alignment with business goals and AWS best practices.
  • Technical Leadership: Provide technical guidance and leadership to the development team, ensuring adherence to coding standards, best practices, and AWS guidelines.
  • AWS Service Integration: Integrate APIs with various AWS services such as AWS Lambda, Amazon API Gateway, Amazon SQS, Amazon SNS, AWS Glue, and others, to build comprehensive and efficient solutions.
  • Performance Optimization: Identify and implement optimizations to improve performance, scalability, and efficiency, leveraging AWS services and tools.
  • Security and Compliance: Ensure APIs are developed following best security practices, including authentication, authorization, encryption, and compliance with relevant standards and regulations.
  • Monitoring and Logging: Implement monitoring and logging solutions for APIs using AWS CloudWatch, AWS X-Ray, or similar tools, to ensure availability, performance, and reliability.
  • Continuous Integration and Deployment (CI/CD): Establish and maintain CI/CD pipelines for API development, automating testing, deployment, and monitoring processes on AWS.
  • Documentation and Training: Create and maintain comprehensive documentation for internal and external users, and provide training and support to developers and stakeholders.
  • Team Collaboration: Collaborate effectively with cross-functional teams, including product managers, designers, and other developers, to deliver high-quality solutions that meet business requirements.
  • Problem Solving: Lead troubleshooting efforts, identifying root causes and implementing solutions to ensure system stability and performance.
  • Stay Updated: Stay updated with the latest trends, tools, and technologies related to development on AWS, and continuously improve your skills and knowledge.


Desired Skills and Experience

  • Development: 10+ years of extensive experience in designing, developing, and deploying using modern technologies, with a focus on scalability, performance, and security.
  • AWS Services: Strong proficiency in using AWS services such as AWS Lambda, Amazon API Gateway, Amazon SQS, Amazon SNS, Amazon SES, Amazon RDS, Amazon DynamoDB, and others, to build and deploy API solutions.
  • Programming Languages: Proficiency in programming languages commonly used for development, such as Python, Node.js, or others, as well as experience with serverless frameworks like AWS.
  • Architecture Design: Ability to design scalable and resilient API architectures using microservices, serverless, or other modern architectural patterns, considering factors like performance, reliability, and cost-efficiency.
  • Security: Strong understanding of security principles and best practices, including authentication, authorization, encryption, and compliance with standards like OAuth, OpenID Connect, and AWS IAM.
  • DevOps Practices: Familiarity with DevOps practices and tools, including CI/CD pipelines, infrastructure as code (IaC), and automated testing, to ensure efficient and reliable deployment on AWS.
  • Problem-solving Skills: Excellent problem-solving skills, with the ability to troubleshoot complex issues, identify root causes, and implement effective solutions to ensure the stability and performance.
  • Team Leadership: Experience leading and mentoring a team of developers, providing technical guidance, code reviews, and fostering a collaborative and innovative environment.
  • Communication Skills: Strong communication skills, with the ability to effectively communicate technical concepts to both technical and non-technical stakeholders, and collaborate with cross- functional teams.
  • Agile Methodologies: Experience working in Agile development environments, following practices like Scrum or Kanban, and ability to adapt to changing requirements and priorities.
  • Continuous Learning: A commitment to continuous learning and staying updated with the latest trends, tools, and technologies related to development and AWS services.
  • Bachelor's Degree: A bachelor's degree in Computer Science, Software Engineering, or a related field is often preferred, although equivalent experience and certifications can also be valuable.
Read more
Remote only
0 - 2 yrs
₹3L - ₹4L / yr
LLM Integration
Retrieval Augmented Generation (RAG)
Vector database
langchain
AI/ML pipelines
+2 more

We’re looking for a full-stack generalist who can handle the entire product lifecycle from frontend, backend, APIs, AI integrations, deployment, and everything in between. Someone who enjoys owning features from concept to production and working across the entire stack

Read more
Vy Systems

at Vy Systems

1 recruiter
Kalki K
Posted by Kalki K
Remote only
4 - 12 yrs
₹18L - ₹28L / yr
databricks
skill iconAmazon Web Services (AWS)
SQL
skill iconPython
PySpark

Job Summary


We are seeking an experienced Databricks Developer with strong skills in PySpark, SQL, Python, and hands-on experience deploying data solutions on AWS (preferred), Azure. The role involves designing, developing, and optimizing scalable data pipelines and analytics workflows on the Databricks platform.


Key Responsibilities

- Develop and optimize ETL/ELT pipelines using Databricks and PySpark.

- Build scalable data workflows on AWS (EC2, S3, Glue, Lambda, IAM) or Azure (ADF, ADLS, Synapse).

- Implement and manage Delta Lake (ACID, schema evolution, time travel).

- Write efficient, complex SQL for transformation and analytics.

- Build and support batch and streaming ingestion (Kafka, Kinesis, EventHub).

- Optimize Databricks clusters, jobs, notebooks, and PySpark performance.

- Collaborate with cross-functional teams to deliver reliable data solutions.

- Ensure data governance, security, and compliance.

- Troubleshoot pipelines and support CI/CD deployments.


Required Skills & Experience

- 4–8 years in Data Engineering / Big Data development.

- Strong hands-on experience with Databricks (clusters, jobs, workflows).

- Advanced PySpark and strong Python skills.

- Expert-level SQL (complex queries, window functions).

- Practical experience with AWS (preferred) or Azure cloud services.

- Experience with Delta Lake, Parquet, and data lake architectures.

- Familiarity with CI/CD tools (GitHub Actions, Azure DevOps, Jenkins).

- Good understanding of data modeling, optimization, and distributed systems.

Read more
Remote only
5 - 15 yrs
₹10L - ₹15L / yr
FastAPI
skill iconPython
RESTful APIs
SQL
NOSQL Databases
+5 more


Summary:

We are seeking a highly skilled Python Backend Developer with proven expertise in FastAPI to join our team as a full-time contractor for 12 months. The ideal candidate will have 5+ years of experience in backend development, a strong understanding of API design, and the ability to deliver scalable, secure solutions. Knowledge of front-end technologies is an added advantage. Immediate joiners are preferred. This role requires full-time commitment—please apply only if you are not engaged in other projects.

Job Type:

Full-Time Contractor (12 months)

Location:

Remote / On-site (Jaipur preferred, as per project needs)

Experience:

5+ years in backend development

Key Responsibilities:

  • Design, develop, and maintain robust backend services using Python and FastAPI.
  •  Implement and manage Prisma ORM for database operations.
  • Build scalable APIs and integrate with SQL databases and third-party services.
  • Deploy and manage backend services using Azure Function Apps and Microsoft Azure Cloud.
  • Collaborate with front-end developers and other team members to deliver high-quality web applications.
  • Ensure application performance, security, and reliability.
  • Participate in code reviews, testing, and deployment processes.

Required Skills:

  • Expertise in Python backend development with strong experience in FastAPI.
  • Solid understanding of RESTful API design and implementation.
  • Proficiency in SQL databases and ORM tools (preferably Prisma)
  • Hands-on experience with Microsoft Azure Cloud and Azure Function Apps.
  • Familiarity with CI/CD pipelines and containerization (Docker).
  • Knowledge of cloud architecture best practices.

Added Advantage:

  • Front-end development knowledge (React, Angular, or similar frameworks).
  • Exposure to AWS/GCP cloud platforms.
  • Experience with NoSQL databases.

Eligibility:

  • Minimum 5 years of professional experience in backend development.
  • Available for full-time engagement.
  • Please excuse if you are currently engaged in other projects—we require dedicated availability.

 

Read more
Grey Chain Technology

at Grey Chain Technology

5 candid answers
Pratikshya Pusty
Posted by Pratikshya Pusty
Remote only
3 - 6 yrs
₹10L - ₹14L / yr
skill iconPython
pytest
Selenium

About Grey chain (https://greychaindesign.com/)


A Generative AI-as-a-service, Mobile & Digital Transformation firm helping organizations reimagine user experiences with Disruptive & Transformational thinking and partnership.

Trusted by: UNICEF, BOSE, KFINTECH, WHO and many Fortune 500 Companies

Home of over 110 Product Engineering nerds building the next generation of Digital Products.Primary Industries: Banking & Financial Services, Non-Profits, Retail & eCommerce, Consumer Goods and Consulting.


Location: Remote


Job Summary:


We're seeking an experienced QA Automation Engineer to join our team. The ideal candidate will have strong technical skills, excellent communication, and a proactive approach to issue resolution. The QA Automation Engineer will design, develop, and maintain automated testing solutions to ensure the quality and reliability of our software applications.


Key Responsibilities:


- Design, develop, and maintain automated testing frameworks using Python (Pytest)

- Create and execute automated test scripts to ensure thorough coverage of application functionality

- Collaborate with cross-functional teams to identify and prioritize testing needs

- Proactively identify and report issues, risks, and opportunities for process improvement

- Work independently with minimal supervision, taking ownership of tasks and deliverables

- Develop and maintain pipelines, integrating automated testing into CI/CD workflows

- Apply OOPS concepts, code reusability principles, and fuzzy logic to write efficient, modular code

- Analyze application flows and functionalities to identify potential issues and improve test coverage


Requirements:


- 3-4 years of experience in automation testing

- Proficient in Python (2-3 years) with expertise in Pytest

- Strong understanding of pipelines and ability to make required changes or learn quickly

- Excellent communication and collaboration skills

- Ability to work independently with minimal supervision

- Proactive approach to issue identification and risk management

- Strong grasp of OOPS concepts, code reusability, and fuzzy logic

Read more
Inflectionio
Renu Philip
Posted by Renu Philip
Remote only
5 - 8 yrs
₹48L - ₹60L / yr
skill iconJava
skill iconSpring Boot
Microservices
skill iconPython
skill iconGo Programming (Golang)
+4 more

Inflection.io is a venture-backed B2B marketing automation company, enabling to communicate with their customers and prospects from one platform. We're used by leading SaaS companies like Sauce Labs, Sigma Computing, BILL, Mural, and Elastic, many of which pay more than $100K/yr (1 crore rupee).

And,... it’s working! We have world-class stats: our largest deal is over 3 crore, we have a 5 star rating on G2, over 100% NRR, and constantly break sales and customer records. We’ve raised $14M in total since 2021 with $7.6M of fresh funding in 2024, giving us many years of runway.

However, we’re still in startup mode with approximately 30 employees and looking for the next SDE3 to help propel Inflection forward. Do you want to join a fast growing startup that is aiming to build a very large company?

Key Responsibilities:

  • Lead the design, development, and deployment of complex software systems and applications.
  • Collaborate with engineers and product managers to define and implement innovative solutions
  • Provide technical leadership and mentorship to junior engineers, promoting best practices and fostering a culture of continuous improvement.
  • Write clean, maintainable and efficient code, ensuring high performance and scalability of the software.
  • Conduct code reviews and provide constructive feedback to ensure code quality and adherence to coding standards.
  • Troubleshoot and resolve complex technical issues, optimizing system performance and reliability.
  • Stay updated with the latest industry trends and technologies, evaluating their potential for adoption in our projects.
  • Participate in the full software development lifecycle, from requirements gathering to deployment and monitoring.

Qualifications:

  • 5+ years of professional software development experience, with a strong focus on backend development.
  • Proficiency in one or more programming languages such as Java, Python, Golang or C#
  • Strong understanding of database systems, both relational (e.g., MySQL, PostgreSQL) and NoSQL (e.g., MongoDB, Cassandra).
  • Hands-on experience with message brokers such as Kafka, RabbitMQ, or Amazon SQS.
  • Experience with cloud platforms (AWS or Azure or Google Cloud) and containerization technologies (Docker, Kubernetes).
  • Proven track record of designing and implementing scalable, high-performance systems.
  • Excellent problem-solving skills and the ability to think critically and creatively.
  • Strong communication and collaboration skills, with the ability to work effectively in a fast-paced, team-oriented environment.


Read more
PGAGI
Javeriya Shaik
Posted by Javeriya Shaik
Remote only
0 - 0.6 yrs
₹2L - ₹2L / yr
skill iconPython
Large Language Models (LLM)
Natural Language Processing (NLP)
skill iconDeep Learning
FastAPI
+1 more

We're at the forefront of creating advanced AI systems, from fully autonomous agents that provide intelligent customer interaction to data analysis tools that offer insightful business solutions. We are seeking enthusiastic interns who are passionate about AI and ready to tackle real-world problems using the latest technologies.


Duration: 6 months


Perks:

- Hands-on experience with real AI projects.

- Mentoring from industry experts.

- A collaborative, innovative and flexible work environment

After completion of the internship period, there is a chance to get a full-time opportunity as AI/ML engineer (Up to 12 LPA).


Compensation:

- Joining Bonus: A one-time bonus of INR 2,500 will be awarded upon joining.

- Stipend: Base is INR 8000/- & can increase up to 20000/- depending upon performance matrix.

Key Responsibilities

  • Experience working with python, LLM, Deep Learning, NLP, etc..
  • Utilize GitHub for version control, including pushing and pulling code updates.
  • Work with Hugging Face and OpenAI platforms for deploying models and exploring open-source AI models.
  • Engage in prompt engineering and the fine-tuning process of AI models.

Requirements

  • Proficiency in Python programming.
  • Experience with GitHub and version control workflows.
  • Familiarity with AI platforms such as Hugging Face and OpenAI.
  • Understanding of prompt engineering and model fine-tuning.
  • Excellent problem-solving abilities and a keen interest in AI technology.


To Apply Click below link and submit the Assignment

https://pgagi.in/jobs/28df1e98-f0c3-4d58-9509-d5b1a4ea9754

Read more
MyOperator - VoiceTree Technologies

at MyOperator - VoiceTree Technologies

1 video
3 recruiters
Vijay Muthu
Posted by Vijay Muthu
Remote only
3 - 5 yrs
₹12L - ₹20L / yr
skill iconPython
skill iconDjango
MySQL
skill iconPostgreSQL
Microservices architecture
+26 more

About Us:

MyOperator and Heyo are India’s leading conversational platforms, empowering 40,000+ businesses with Call and WhatsApp-based engagement. We’re a product-led SaaS company scaling rapidly, and we’re looking for a skilled Software Developer to help build the next generation of scalable backend systems.


Role Overview:

We’re seeking a passionate Python Developer with strong experience in backend development and cloud infrastructure. This role involves building scalable microservices, integrating AI tools like LangChain/LLMs, and optimizing backend performance for high-growth B2B products.


Key Responsibilities:

  • Develop robust backend services using Python, Django, and FastAPI
  • Design and maintain a scalable microservices architecture
  • Integrate LangChain/LLMs into AI-powered features
  • Write clean, tested, and maintainable code with pytest
  • Manage and optimize databases (MySQL/Postgres)
  • Deploy and monitor services on AWS
  • Collaborate across teams to define APIs, data flows, and system architecture

Must-Have Skills:

  • Python and Django
  • MySQL or Postgres
  • Microservices architecture
  • AWS (EC2, RDS, Lambda, etc.)
  • Unit testing using pytest
  • LangChain or Large Language Models (LLM)
  • Strong grasp of Data Structures & Algorithms
  • AI coding assistant tools (e.g., Chat GPT & Gemini)

Good to Have:

  • MongoDB or ElasticSearch
  • Go or PHP
  • FastAPI
  • React, Bootstrap (basic frontend support)
  • ETL pipelines, Jenkins, Terraform

Why Join Us?

  • 100% Remote role with a collaborative team
  • Work on AI-first, high-scale SaaS products
  • Drive real impact in a fast-growing tech company
  • Ownership and growth from day one


Read more
Hypersonix Inc

at Hypersonix Inc

2 candid answers
1 product
Reshika Mendiratta
Posted by Reshika Mendiratta
Remote only
4yrs+
Upto ₹30L / yr (Varies
)
SQL
skill iconPython
Data engineering
Big Data
skill iconAmazon Web Services (AWS)
+1 more

About the Company

Hypersonix.ai is disrupting the e-commerce space with AI, ML and advanced decision capabilities to drive real-time business insights. Hypersonix.ai has been built ground up with new age technology to simplify the consumption of data for our customers in various industry verticals. Hypersonix.ai is seeking a well-rounded, hands-on product leader to help lead product management of key capabilities and features.


About the Role

We are looking for talented and driven Data Engineers at various levels to work with customers to build the data warehouse, analytical dashboards and ML capabilities as per customer needs.


Roles and Responsibilities

  • Create and maintain optimal data pipeline architecture
  • Assemble large, complex data sets that meet functional / non-functional business requirements; should write complex queries in an optimized way
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and AWS ‘big data’ technologies
  • Run ad-hoc analysis utilizing the data pipeline to provide actionable insights
  • Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs
  • Keep our data separated and secure across national boundaries through multiple data centers and AWS regions
  • Work with analytics and data scientist team members and assist them in building and optimizing our product into an innovative industry leader


Requirements

  • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases
  • Experience building and optimizing ‘big data’ data pipelines, architectures and data sets
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement
  • Strong analytic skills related to working with unstructured datasets
  • Build processes supporting data transformation, data structures, metadata, dependency and workload management
  • A successful history of manipulating, processing and extracting value from large disconnected datasets
  • Working knowledge of message queuing, stream processing, and highly scalable ‘big data’ data stores
  • Experience supporting and working with cross-functional teams in a dynamic environment
  • We are looking for a candidate with 4+ years of experience in a Data Engineer role, who has attained a Graduate degree in Computer Science, Information Technology or completed MCA.
Read more
Appiness Interactive
Remote only
6 - 10 yrs
₹10L - ₹14L / yr
skill iconPython
skill iconDjango
FastAPI
skill iconFlask
pandas
+9 more

Position Overview: The Lead Software Architect - Python & Data Engineering is a senior technical leadership role responsible for designing and owning end-to-end architecture for data-intensive, AI/ML, and analytics platforms, while mentoring developers and ensuring technical excellence across the organization. 


Key Responsibilities: 

  • Design end-to-end software architecture for data-intensive applications, AI/ML pipelines, and analytics platforms
  • Evaluate trade-offs between competing technical approaches 
  • Define data models, API approach, and integration patterns across systems 
  • Create technical specifications and architecture documentation 
  • Lead by example through production-grade Python code and mentor developers on engineering fundamentals 
  • Conduct design and code reviews focused on architectural soundness 
  • Establish engineering standards, coding practices, and design patterns for the team 
  • Translate business requirements into technical architecture 
  • Collaborate with data scientists, analysts, and other teams to design integrated solutions 
  • Whiteboard and defend system design and architectural choices 
  • Take responsibility for system performance, reliability, and maintainability 
  • Identify and resolve architectural bottlenecks proactively 


Required Skills:  

  • 8+ years of experience in software architecture and development  
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field 
  • Strong foundations in data structures, algorithms, and computational complexity 
  • Experience in system design for scale, including caching strategies, load balancing, and asynchronous processing  
  • 6+ years of Python development experience 
  • Deep knowledge of Django, Flask, or FastAPI 
  • Expert understanding of Python internals including GIL and memory management 
  • Experience with RESTful API design and event-driven architectures (Kafka, RabbitMQ) 
  • Proficiency in data processing frameworks such as Pandas, Apache Spark, and Airflow 
  • Strong SQL optimization and database design experience (PostgreSQL, MySQL, MongoDB)  Experience with AWS, GCP, or Azure cloud platforms 
  • Knowledge of containerization (Docker) and orchestration (Kubernetes) 
  • Hands-on experience designing CI/CD pipelines Preferred (Bonus) 


Skills

  • Experience deploying ML models to production (MLOps, model serving, monitoring) Understanding of ML system design including feature stores and model versioning 
  • Familiarity with ML frameworks such as scikit-learn, TensorFlow, and PyTorch  
  • Open-source contributions or technical blogging demonstrating architectural depth 
  • Experience with modern front-end frameworks for full-stack perspective


Read more
Middleware

at Middleware

2 candid answers
1 video
Jayant Bhawal
Posted by Jayant Bhawal
Remote only
4 - 8 yrs
₹25L - ₹40L / yr
TypeScript
skill iconPython
skill iconPostgreSQL
skill iconJavascript
skill iconReact.js
+2 more

Why Middleware? 💡


Sick of the endless waiting?

Waiting on code reviews, QA feedback, or that "quick call"?

At Middleware, we’re all about freeing up engineers like you to do what you love—build.


We’ve created a cockpit that gives engineering leaders the insights they need to unblock teams, cut bottlenecks, and let engineers focus on impact.


What You’ll Do 🎨

  • Own the Product: Shape a product that engineers rely on daily.
  • Build Stunning UIs: Craft data-rich, intuitive designs that solve real problems.
  • Shape Middleware’s Architecture: Make our systems robust, seamless, and introduce mechanisms that allow high visibility into our automated pipelines.


What We’re Looking For 🔍

  • React + Typescript: You know your way around these tools and have launched awesome projects.
  • Python + Postgres: You've build complete backend systems, not just basic CRUD apps.
  • Passionate Builder: Hungry to grow, build, and make an impact.


Bonus Points ⭐️

  • Eye for Design: You have a sense for clean, user-friendly visuals.
  • Understanding of distributed systems: Not everything runs on a single machine, and you know how to make things work across a lot of those.
  • DSA Know-how: Familiarity with data structures (graphs, linked lists, etc.) because our product (even frontend) actually uses DSA concepts.


Why You'll Love Working with Us ❤️

We’re engineers at heart.

Middleware was founded by ex-Uber and Maersk engineers who know what it’s like to be stuck in meeting loops and endless waiting. If you're here to build, to make things happen, and to change the game for engineering teams everywhere, let’s chat!


Ready to jump in? Explore Middleware (https://www.middlewarehq.com/) or check out our demo (https://demo.middlewarehq.com/).

Read more
Remote only
2 - 10 yrs
₹10L - ₹20L / yr
skill iconPython
JSON
skill iconPostgreSQL
PGSQL
Large Language Models (LLM)

Role & Responsibilities:


We are seeking a Software Developer with 5-10 year’s experience with strong foundations in Python, databases, and AI technologies.


The ideal candidate will support the development of AI-powered solutions, focusing on LLM integration, prompt engineering, and database-driven workflows.


This is a hands-on role with opportunities to learn and grow into advanced AI engineering responsibilities.


Key Responsibilities :


• Develop, test, and maintain Python-based applications and APIs.


• Design and optimize prompts for Large Language Models (LLMs) to improve accuracy and performance.


• Work with JSON-based data structures for request/response handling.


• Integrate and manage PostgreSQL (pgSQL) databases, including writing queries and handling data pipelines.


• Collaborate with the product and AI teams to implement new features.


• Debug, troubleshoot, and optimize performance of applications and workflows.


• Stay updated on advancements in LLMs, AI frameworks, and generative AI tools.

Read more
Remote only
2 - 9 yrs
₹5L - ₹12L / yr
skill iconPython
python scripting
Databases
LLM integration
Prompt engineering
+11 more

Role & Responsibilities:

We are seeking a Software Developer with 2-10 year’s experience with strong foundations in Python, databases, and AI technologies. The ideal candidate will support the development of AI-powered solutions, focusing on LLM integration, prompt engineering, and database-driven workflows. This is a hands-on role with opportunities to learn and grow into advanced AI engineering responsibilities.


Key Responsibilities

• Develop, test, and maintain Python-based applications and APIs.

• Design and optimize prompts for Large Language Models (LLMs) to improve accuracy and performance.

• Work with JSON-based data structures for request/response handling. • Integrate and manage PostgreSQL (pgSQL) databases, including writing queries and handling data pipelines.

• Collaborate with the product and AI teams to implement new features. • Debug, troubleshoot, and optimize performance of applications and workflows.

• Stay updated on advancements in LLMs, AI frameworks, and generative AI tools.


Required Skills & Qualifications


• Strong knowledge of Python (scripting, APIs, data handling).

• Basic understanding of Large Language Models (LLMs) and prompt engineering techniques.

• Experience with JSON data parsing and transformations.

• Familiarity with PostgreSQL or other relational databases.

• Ability to write clean, maintainable, and well-documented code.

• Strong problem-solving skills and eagerness to learn.

• Bachelor’s degree in Computer Science, Engineering, or related field (or equivalent practical experience).


Nice-to-Have (Preferred)

• Exposure to AI/ML frameworks (e.g., LangChain, Hugging Face, OpenAI APIs).

• Experience working in startups or fast-paced environments.

• Familiarity with version control (Git/GitHub) and cloud platforms (AWS, GCP, or Azure).


What We Offer

• The opportunity to define the future of GovTech through AI-powered solutions.

• A strategic leadership role in a fast-scaling startup with direct impact on product direction and market success.

• Collaborative and innovative environment with cross-functional exposure.

• Growth opportunities backed by a strong leadership team.

• Remote flexibility and work-life balance.

Read more
NA

NA

Agency job
via eTalent Services by JaiPrakash Bharti
Remote only
3 - 8 yrs
₹5L - ₹14L / yr
skill iconPython
skill iconMachine Learning (ML)
Windows Azure
TensorFlow
MLFlow
+6 more

Role: Azure AI Tech Lead

Exp-3.5-7 Years

Location: Remote / Noida (NCR)

Notice Period: Immediate to 15 days

 

Mandatory Skills: Python, Azure AI/ML, PyTorch, TensorFlow, JAX, HuggingFace, LangChain, Kubeflow, MLflow, LLMs, RAG, MLOps, Docker, Kubernetes, Generative AI, Model Deployment, Prometheus, Grafana

 

JOB DESCRIPTION

As the Azure AI Tech Lead, you will serve as the principal technical expert leading the design, development, and deployment of advanced AI and ML solutions on the Microsoft Azure platform. You will guide a team of engineers, establish robust architectures, and drive end-to-end implementation of AI projects—transforming proof-of-concepts into scalable, production-ready systems.

 

Key Responsibilities:

  • Lead architectural design and development of AI/ML solutions using Azure AI, Azure OpenAI, and Cognitive Services.
  • Develop and deploy scalable AI systems with best practices in MLOps across the full model lifecycle.
  • Mentor and upskill AI/ML engineers through technical reviews, training, and guidance.
  • Implement advanced generative AI techniques including LLM fine-tuning, RAG systems, and diffusion models.
  • Collaborate cross-functionally to translate business goals into innovative AI solutions.
  • Enforce governance, responsible AI practices, and performance optimization standards.
  • Stay ahead of trends in LLMs, agentic AI, and applied research to shape next-gen solutions.

 

Qualifications:

  • Bachelor’s or Master’s in Computer Science or related field.
  • 3.5–7 years of experience delivering end-to-end AI/ML solutions.
  • Strong expertise in Azure AI ecosystem and production-grade model deployment.
  • Deep technical understanding of ML, DL, Generative AI, and MLOps pipelines.
  • Excellent analytical and problem-solving abilities; applied research or open-source contributions preferred.


Read more
venanalytics

at venanalytics

2 candid answers
Rincy jain
Posted by Rincy jain
Remote, Mumbai
3 - 4 yrs
₹7L - ₹10L / yr
skill iconPython
SQL
PowerBI
Client Servicing
Team Management
+6 more

About Ven Analytics


At Ven Analytics, we don’t just crunch numbers — we decode them to uncover insights that drive real business impact. We’re a data-driven analytics company that partners with high-growth startups and enterprises to build powerful data products, business intelligence systems, and scalable reporting solutions. With a focus on innovation, collaboration, and continuous learning, we empower our teams to solve real-world business problems using the power of data.


Role Overview


We’re looking for a Power BI Data Analyst who is not just proficient in tools but passionate about building insightful, scalable, and high-performing dashboards. The ideal candidate should have strong fundamentals in data modeling, a flair for storytelling through data, and the technical skills to implement robust data solutions using Power BI, Python, and SQL..


Key Responsibilities


  • Technical Expertise: Develop scalable, accurate, and maintainable data models using Power BI, with a clear understanding of Data Modeling, DAX, Power Query, and visualization principles.


  • Programming Proficiency: Use SQL and Python for complex data manipulation, automation, and analysis.


  • Business Problem Translation: Collaborate with stakeholders to convert business problems into structured data-centric solutions considering performance, scalability, and commercial goals.


  • Hypothesis Development: Break down complex use-cases into testable hypotheses and define relevant datasets required for evaluation.


  • Solution Design: Create wireframes, proof-of-concepts (POC), and final dashboards in line with business requirements.


  • Dashboard Quality: Ensure dashboards meet high standards of data accuracy, visual clarity, performance, and support SLAs.


  • Performance Optimization: Continuously enhance user experience by improving performance, maintainability, and scalability of Power BI solutions.


  • Troubleshooting & Support: Quick resolution of access, latency, and data issues as per defined SLAs.


  • Power BI Development: Use power BI desktop for report building and service for distribution 


  • Backend development: Develop optimized SQL queries that are easy to consume, maintain and debug.


  • Version Control: Strict control on versions by tracking CRs and Bugfixes. Ensuring the maintenance of Prod and Dev dashboards. 


  • Client Servicing : Engage with clients to understand their data needs, gather requirements, present insights, and ensure timely, clear communication throughout project cycles.


  • Team Management : Lead and mentor a small team by assigning tasks, reviewing work quality, guiding technical problem-solving, and ensuring timely delivery of dashboards and reports..


Must-Have Skills


  • Strong experience building robust data models in Power BI
  • Hands-on expertise with DAX (complex measures and calculated columns)
  • Proficiency in M Language (Power Query) beyond drag-and-drop UI
  • Clear understanding of data visualization best practices (less fluff, more insight)
  • Solid grasp of SQL and Python for data processing
  • Strong analytical thinking and ability to craft compelling data stories
  • Client Servicing Background.


Good-to-Have (Bonus Points)


  • Experience using DAX Studio and Tabular Editor
  • Prior work in a high-volume data processing production environment
  • Exposure to modern CI/CD practices or version control with BI tools

 

Why Join Ven Analytics?


  • Be part of a fast-growing startup that puts data at the heart of every decision.
  • Opportunity to work on high-impact, real-world business challenges.
  • Collaborative, transparent, and learning-oriented work environment.
  • Flexible work culture and focus on career development.


Read more
Ekloud INC
Kratika Agarwal
Posted by Kratika Agarwal
Remote only
8 - 14 yrs
₹7L - ₹18L / yr
m365
m365 developer
ms teams
MS SharePoint
Microsoft Exchange
+12 more

Requires that any candidate know the M365 Collaboration environment. SharePoint Online, MS Teams. Exchange Online, Entra and Purview. Need developer that possess a strong understanding of Data Structure, Problem Solving abilities, SQL, PowerShell, MS Teams App Development, Python, Visual Basic, C##, JavaScript, Java, HTML, PHP, C.

Need a strong understanding of the development lifecycle, and possess debugging skills time management, business acumen, and have a positive attitude is a must and open to continual growth.

Capability to code appropriate solutions will be tested in any interview.

Knowledge of a wide variety of Generative AI models

Conceptual understanding of how large language models work

Proficiency in coding languages for data manipulation (e.g., SQL) and machine learning & AI development (e.g., Python)

Experience with dashboarding tools such as Power BI and Tableau (beneficial but not essential)

Read more
Forbes Advisor

at Forbes Advisor

3 candid answers
Bisman Gill
Posted by Bisman Gill
Remote only
4yrs+
Upto ₹40L / yr (Varies
)
skill iconPython
SQL
Database performance tuning
Data-flow analysis
Data modeling

About Forbes Advisor

Forbes Digital Marketing Inc. is a high-growth digital media and technology company dedicated to helping consumers make confident, informed decisions about their money, health, careers, and everyday life.

We do this by combining data-driven content, rigorous product comparisons, and user-first design — all built on top of a modern, scalable platform. Our global teams bring deep expertise across journalism, product, performance marketing, data, and analytics.

 

The Role

We’re hiring a Data Scientist to help us unlock growth through advanced analytics and machine learning. This role sits at the intersection of marketing performance, product optimization, and decision science.


You’ll partner closely with Paid Media, Product, and Engineering to build models, generate insight, and influence how we acquire, retain, and monetize users. From campaign ROI to user segmentation and funnel optimization, your work will directly shape how we grow.This role is ideal for someone who thrives on business impact, communicates clearly, and wants to build re-usable, production-ready insights — not just run one-off analyses.

 

What You’ll Do

Marketing & Revenue Modelling

• Own end-to-end modelling of LTV, user segmentation, retention, and marketing

efficiency to inform media optimization and value attribution.

• Collaborate with Paid Media and RevOps to optimize SEM performance, predict high-

value cohorts, and power strategic bidding and targeting.

Product & Growth Analytics

• Work closely with Product Insights and General Managers (GMs) to define core metrics, KPIs, and success frameworks for new launches and features.

• Conduct deep-dive analysis of user behaviour, funnel performance, and product engagement to uncover actionable insights.

• Monitor and explain changes in key product metrics, identifying root causes and business impact.

• Work closely with Data Engineering to design and maintain scalable data pipelines that

support machine learning workflows, model retraining, and real-time inference.

Predictive Modelling & Machine Learning

• Build predictive models for conversion, churn, revenue, and engagement using regression, classification, or time-series approaches.

• Identify opportunities for prescriptive analytics and automation in key product and marketing workflows.

• Support development of reusable ML pipelines for production-scale use cases in product recommendation, lead scoring, and SEM planning.

Collaboration & Communication

• Present insights and recommendations to a variety of stakeholders — from ICs to executives — in a clear and compelling manner.

• Translate business needs into data problems, and complex findings into strategic action plans.

• Work cross-functionally with Engineering, Product, BI, and Marketing to deliver and deploy your work.

 

What You’ll Bring

Minimum Qualifications

• Bachelor’s degree in a quantitative field (Mathematics, Statistics, CS, Engineering, etc.).

• 4+ years in data science, growth analytics, or decision science roles.

• Strong SQL and Python skills (Pandas, Scikit-learn, NumPy).

• Hands-on experience with Tableau, Looker, or similar BI tools.

• Familiarity with LTV modelling, retention curves, cohort analysis, and media attribution.

• Experience with GA4, Google Ads, Meta, or other performance marketing platforms.

• Clear communication skills and a track record of turning data into decisions.


Nice to Have

• Experience with BigQuery and Google Cloud Platform (or equivalent).

• Familiarity with affiliate or lead-gen business models.

• Exposure to NLP, LLMs, embeddings, or agent-based analytics.

• Ability to contribute to model deployment workflows (e.g., using Vertex AI, Airflow, or Composer).

 

Why Join Us?

• Remote-first and flexible — work from anywhere in India with global exposure.

• Monthly long weekends (every third Friday off).

• Generous wellness stipends and parental leave.

• A collaborative team where your voice is heard and your work drives real impact.

• Opportunity to help shape the future of data science at one of the world’s most trusted

brands.

Read more
Quanteon Solutions
DurgaPrasad Sannamuri
Posted by DurgaPrasad Sannamuri
Remote only
4 - 5 yrs
₹10L - ₹15L / yr
Artificial Intelligence (AI)
Generative AI
Large Language Models (LLM) tuning
skill iconMachine Learning (ML)
skill iconPython
+4 more

We are looking for an 4-5 years of experienced AI/ML Engineer who can design, develop, and deploy scalable machine learning models and AI-driven solutions. The ideal candidate should have strong expertise in data processing, model building, and production deployment, along with solid programming and problem-solving skills.

Key Responsibilities

  • Develop and deploy machine learning, deep learning, and NLP models for various business use cases.
  • Build end-to-end ML pipelines including data preprocessing, feature engineering, training, evaluation, and production deployment.
  • Optimize model performance and ensure scalability in production environments.
  • Work closely with data scientists, product teams, and engineers to translate business requirements into AI solutions.
  • Conduct data analysis to identify trends and insights.
  • Implement MLOps practices for versioning, monitoring, and automating ML workflows.
  • Research and evaluate new AI/ML techniques, tools, and frameworks.
  • Document system architecture, model design, and development processes.

Required Skills

  • Strong programming skills in Python (NumPy, Pandas, Scikit-learn, TensorFlow, PyTorch, Keras).
  • Hands-on experience in building and deploying ML/DL models in production.
  • Good understanding of machine learning algorithms, neural networks, NLP, and computer vision.
  • Experience with REST APIs, Docker, Kubernetes, and cloud platforms (AWS/GCP/Azure).
  • Working knowledge of MLOps tools such as MLflow, Airflow, DVC, or Kubeflow.
  • Familiarity with data pipelines and big data technologies (Spark, Hadoop) is a plus.
  • Strong analytical skills and ability to work with large datasets.
  • Excellent communication and problem-solving abilities.

Preferred Qualifications

  • Bachelor’s or Master’s degree in Computer Science, Data Science, AI, Machine Learning, or related fields.
  • Experience in deploying models using cloud services (AWS Sagemaker, GCP Vertex AI, etc.).
  • Experience in LLM fine-tuning or generative AI is an added advantage.


Read more
Big Rattle Technologies
Sreelakshmi Nair (Big Rattle Technologies)
Posted by Sreelakshmi Nair (Big Rattle Technologies)
Remote, Mumbai
5 - 7 yrs
₹8L - ₹12L / yr
skill iconPython
SQL
skill iconMachine Learning (ML)
Data profiling
E2E
+8 more

Position: QA Engineer – Machine Learning Systems (5 - 7 years)

Location: Remote (Company in Mumbai)

Company: Big Rattle Technologies Private Limited


Immediate Joiners only.


Summary:

The QA Engineer will own quality assurance across the ML lifecycle—from raw data validation through feature engineering checks, model training/evaluation verification, batch prediction/optimization validation, and end-to-end (E2E) workflow testing. The role is hands-on with Python automation, data profiling, and pipeline test harnesses in Azure ML and Azure DevOps. Success means probably correct data, models, and outputs at production scale and cadence.


Key Responsibilities:

Test Strategy & Governance

  • ○ Define an ML-specific Test Strategy covering data quality KPIs, feature consistency
  • checks, model acceptance gates (metrics + guardrails), and E2E run acceptance
  • (timeliness, completeness, integrity).
  • ○ Establish versioned test datasets & golden baselines for repeatable regression of
  • features, models, and optimizers.


Data Quality & Transformation

  • Validate raw data extracts and landed data lake data: schema/contract checks, null/outlier thresholds, time-window completeness, duplicate detection, site/material coverage.
  • Validate transformed/feature datasets: deterministic feature generation, leakage detection, drift vs. historical distributions, feature parity across runs (hash or statistical similarity tests).
  • Implement automated data quality checks (e.g., Great Expectations/pytest + Pandas/SQL) executed in CI and AML pipelines.

Model Training & Evaluation

  • Verify training inputs (splits, windowing, target leakage prevention) and hyperparameter configs per site/cluster.
  • Automate metric verification (e.g., MAPE/MAE/RMSE, uplift vs. last model, stability tests) with acceptance thresholds and champion/challenger logic.
  • Validate feature importance stability and sensitivity/elasticity sanity checks (price/volume monotonicity where applicable).
  • Gate model registration/promotion in AML based on signed test artifacts and reproducible metrics.


Predictions, Optimization & Guardrails

  • Validate batch predictions: result shapes, coverage, latency, and failure handling.
  • Test model optimization outputs and enforced guardrails: detect violations and prove idempotent writes to DB.
  • Verify API push to third party system (idempotency keys, retry/backoff, delivery receipts).


Pipelines & E2E

  • Build pipeline test harnesses for AML pipelines (data-gen nightly, training weekly,
  • prediction/optimization) including orchestrated synthetic runs and fault injection
  • (missing slice, late competitor data, SB backlog).
  • Run E2E tests from raw data store -> ADLS -> AML -> RDBMS -> APIM/Frontend, assert
  • freshness SLOs and audit event completeness (Event Hubs -> ADLS immutable).


Automation & Tooling

  • Develop Python-based automated tests (pytest) for data checks, model metrics, and API contracts; integrate with Azure DevOps (pipelines, badges, gates).
  • Implement data-driven test runners (parameterized by site/material/model-version) and store signed test artifacts alongside models in AML Registry.
  • Create synthetic test data generators and golden fixtures to cover edge cases (price gaps, competitor shocks, cold starts).


Reporting & Quality Ops

  • Publish weekly test reports and go/no-go recommendations for promotions; maintain a defect taxonomy (data vs. model vs. serving vs. optimization).
  • Contribute to SLI/SLO dashboards (prediction timeliness, queue/DLQ, push success, data drift) used for release gates.


Required Skills (hands-on experience in the following):

  • Python automation (pytest, pandas, NumPy), SQL (PostgreSQL/Snowflake), and CI/CD (Azure
  • DevOps) for fully automated ML QA.
  • Strong grasp of ML validation: leakage checks, proper splits, metric selection
  • (MAE/MAPE/RMSE), drift detection, sensitivity/elasticity sanity checks.
  • Experience testing AML pipelines (pipelines/jobs/components), and message-driven integrations
  • (Service Bus/Event Hubs).
  • API test skills (FastAPI/OpenAPI, contract tests, Postman/pytest-httpx) + idempotency and retry
  • patterns.
  • Familiar with feature stores/feature engineering concepts and reproducibility.
  • Solid understanding of observability (App Insights/Log Analytics) and auditability requirements.


Required Qualifications:

  • Bachelor’s or Master’s degree in Computer Science, Information Technology, or related field.
  • 5–7+ years in QA with 3+ years focused on ML/Data systems (data pipelines + model validation).
  • Certification in Azure Data or ML Engineer Associate is a plus.



Why should you join Big Rattle?

Big Rattle Technologies specializes in AI/ ML Products and Solutions as well as Mobile and Web Application Development. Our clients include Fortune 500 companies. Over the past 13 years, we have delivered multiple projects for international and Indian clients from various industries like FMCG, Banking and Finance, Automobiles, Ecommerce, etc. We also specialise in Product Development for our clients.

Big Rattle Technologies Private Limited is ISO 27001:2022 certified and CyberGRX certified.

What We Offer:

  • Opportunity to work on diverse projects for Fortune 500 clients.
  • Competitive salary and performance-based growth.
  • Dynamic, collaborative, and growth-oriented work environment.
  • Direct impact on product quality and client satisfaction.
  • 5-day hybrid work week.
  • Certification reimbursement.
  • Healthcare coverage.

How to Apply:

Interested candidates are invited to submit their resume detailing their experience. Please detail out your work experience and the kind of projects you have worked on. Ensure you highlight your contributions and accomplishments to the projects.


Read more
Intineri infosol Pvt Ltd

at Intineri infosol Pvt Ltd

2 candid answers
Shivani Pandey
Posted by Shivani Pandey
Remote only
4 - 10 yrs
₹5L - ₹15L / yr
skill iconPython
2D Geometry Concept
3D Geometry Concept
NumPy
SciPy
+9 more

Job Title: Python Developer

Experience Level: 4+ years

 

Job Summary:

We are seeking a skilled Python Developer with strong experience in developing and maintaining APIs. Familiarity with 2D and 3D geometry concepts is a strong plus. The ideal candidate will be passionate about clean code, scalable systems, and solving complex geometric and computational problems.


Key Responsibilities:

·       Design, develop, and maintain robust and scalable APIs using Python.

·       Work with geometric data structures and algorithms (2D/3D).

·       Collaborate with cross-functional teams including front-end developers, designers, and product managers.

·       Optimize code for performance and scalability.

·       Write unit and integration tests to ensure code quality.

·       Participate in code reviews and contribute to best practices.

 

Required Skills:

·       Strong proficiency in Python.

·       Experience with RESTful API development (e.g., Flask, FastAPI, Django REST Framework).

·       Good understanding of 2D/3D geometry, computational geometry, or CAD-related concepts.

·       Familiarity with libraries such as NumPySciPyShapelyOpen3D, or PyMesh.

·       Experience with version control systems (e.g., Git).

·       Strong problem-solving and analytical skills.

 

Good to Have:

·       Experience with 3D visualization tools or libraries (e.g., VTK, Blender API, Three.js via Python bindings).

·       Knowledge of mathematical modeling or simulation.

·       Exposure to cloud platforms (AWS, Azure, GCP).

·       Familiarity with CI/CD pipelines.

 

Education:

·       Bachelor’s or Master’s degree in Computer Science, Engineering, Mathematics, or a related field.

Read more
Slooze

at Slooze

2 candid answers
Hari Krishna
Posted by Hari Krishna
Remote, Coimbatore
0 - 4 yrs
₹2L - ₹12L / yr
skill iconPython
GraphQL
crawlers
data pipeline
Integration

About the Role

We’re looking for a Data Engineer who can turn messy, unstructured information into clean, usable insights. You’ll be building crawlers, integrating APIs, and setting up data flows that power our analytics and AI layers. If you love data plumbing as much as data puzzles — this role is for you.

 

Responsibilities

- Build and maintain Python-based data pipelines, crawlers, and integrations with 3rd-party APIs.

Perform brute-force analytics and exploratory data work on crawled datasets to surface trends and anomalies.

- Develop and maintain ETL workflows — from raw ingestion to clean, structured outputs.

- Collaborate with product and ML teams to make data discoverable, queryable, and actionable.

- Optimize data collection for performance, reliability, and scalability.

 

Requirements

- Strong proficiency in Python and Jupyter notebooks.

- Experience building web crawlers / scrapers and integrating with REST / GraphQL APIs.

- Solid understanding of data structures and algorithms (DSA).

- Comfort with quick, hands-on analytics — slicing and validating data directly.

 

Good to Have

- Experience with schema design and database modeling.

- Exposure to both SQL and NoSQL databases; familiarity with vector databases is a plus.

- Knowledge of data orchestration tools (Dagster preferred).

- Understanding of data lifecycle management — from raw to enriched layers.

 

Why Join Us

We’re not offering employment — we’re offering ownership.

If you’re here for a job, this isn’t your place.

We’re building the data spine of a new-age supply chain intelligence platform — and we need people who can crush constraints, move fast, and make impossible things work.

You’ll have room to think, build, break, and reinvent — not follow.

If you thrive in chaos and create clarity, you’ll fit right in.

 

Screening Challenge

 

Before we schedule a call, we have an exciting challenge for you. Please go through the link below and submit your solution to us with the Subject line: SUB: [Role] [Full Name]

 

Link - https://dub.sh/slooze-takehome

Read more
Synorus
Synorus Admin
Posted by Synorus Admin
Remote only
0 - 1 yrs
₹0.2L - ₹1L / yr
Google colab
Retrieval Augmented Generation (RAG)
Large Language Models (LLM) tuning
skill iconPython
PyTorch
+3 more

About Synorus

Synorus is building a next-generation ecosystem of AI-first products. Our flagship legal-AI platform LexVault is redefining legal research, drafting, knowledge retrieval, and case intelligence using domain-tuned LLMs, private RAG pipelines, and secure reasoning systems.

If you are passionate about AI, legaltech, and training high-performance models — this internship will put you on the front line of innovation.


Role Overview

We are seeking passionate AI/LLM Engineering Interns who can:

  • Fine-tune LLMs for legal domain use-cases
  • Train and experiment with open-source foundation models
  • Work with large datasets efficiently
  • Build RAG pipelines and text-processing frameworks
  • Run model training workflows on Google Colab / Kaggle / Cloud GPUs

This is a hands-on engineering and research internship — you will work directly with senior founders & technical leadership.

Key Responsibilities

  • Fine-tune transformer-based models (Llama, Mistral, Gemma, etc.)
  • Build and preprocess legal datasets at scale
  • Develop efficient inference & training pipelines
  • Evaluate models for accuracy, hallucinations, and trustworthiness
  • Implement RAG architectures (vector DBs + embeddings)
  • Work with GPU environments (Colab/Kaggle/Cloud)
  • Contribute to model improvements, prompt engineering & safety tuning

Must-Have Skills

  • Strong knowledge of Python & PyTorch
  • Understanding of LLMs, Transformers, Tokenization
  • Hands-on experience with HuggingFace Transformers
  • Familiarity with LoRA/QLoRA, PEFT training
  • Data wrangling: Pandas, NumPy, tokenizers
  • Ability to handle multi-GB datasets efficiently

Bonus Skills

(Not mandatory — but a strong plus)

  • Experience with RAG / vector DBs (Chroma, Qdrant, LanceDB)
  • Familiarity with vLLM, llama.cpp, GGUF
  • Worked on summarization, Q&A or document-AI projects
  • Knowledge of legal texts (Indian laws/case-law/statutes)
  • Open-source contributions or research work

What You Will Gain

  • Real-world training on LLM fine-tuning & legal AI
  • Exposure to production-grade AI pipelines
  • Direct mentorship from engineering leadership
  • Research + industry project portfolio
  • Letter of experience + potential full-time offer

Ideal Candidate

  • You experiment with models on weekends
  • You love pushing GPUs to their limits
  • You prefer research + implementation over theory alone
  • You want to build AI that matters — not just demos


Location - Remote

Stipend - 5K - 10K

Read more
Palcode.ai

at Palcode.ai

2 candid answers
Team Palcode
Posted by Team Palcode
Remote only
1 - 2 yrs
₹3L - ₹4L / yr
skill iconPython
FastAPI
skill iconFlask

At Palcode.ai, We're on a mission to fix the massive inefficiencies in pre-construction. Think about it - in a $10 trillion industry, estimators still spend weeks analyzing bids, project managers struggle with scattered data, and costly mistakes slip through complex contracts. We're fixing this with purpose-built AI agents that work. Our platform can do “magic” to Preconstruction workflows from Weeks to Hours. It's not just about AI – it's about bringing real, measurable impact to an industry ready for change. We are backed by names like AWS for Startups, Upekkha Accelerator, and Microsoft for Startups.



Why Palcode.ai


Tackle Complex Problems: Build AI that reads between the lines of construction bids, spots hidden risks in contracts, and makes sense of fragmented project data

High-Impact Code: Your code won't sit in a backlog – it goes straight to estimators and project managers who need it yesterday

Tech Challenges That Matter: Design systems that process thousands of construction documents, handle real-time pricing data, and make intelligent decisions

Build & Own: Shape our entire tech stack, from data processing pipelines to AI model deployment

Quick Impact: Small team, huge responsibility. Your solutions directly impact project decisions worth millions

Learn & Grow: Master the intersection of AI, cloud architecture, and construction tech while working with founders who've built and scaled construction software


Your Role:

  • Design and build our core AI services and APIs using Python
  • Create reliable, scalable backend systems that handle complex data
  • Help set up cloud infrastructure and deployment pipelines
  • Collaborate with our AI team to integrate machine learning models
  • Write clean, tested, production-ready code


You'll fit right in if:

  • You have 1 year of hands-on Python development experience
  • You're comfortable with full-stack development and cloud services
  • You write clean, maintainable code and follow good engineering practices
  • You're curious about AI/ML and eager to learn new technologies
  • You enjoy fast-paced startup environments and take ownership of your work


How we will set you up for success

  • You will work closely with the Founding team to understand what we are building.
  • You will be given comprehensive training about the tech stack, with an opportunity to avail virtual training as well.
  • You will be involved in a monthly one-on-one with the founders to discuss feedback
  • A unique opportunity to learn from the best - we are Gold partners of AWS, Razorpay, and Microsoft Startup programs, having access to rich talent to discuss and brainstorm ideas.
  • You’ll have a lot of creative freedom to execute new ideas. As long as you can convince us, and you’re confident in your skills, we’re here to back you in your execution.


Location: Bangalore, Remote


Compensation: Competitive salary + Meaningful equity


If you get excited about solving hard problems that have real-world impact, we should talk.


All the best!!

Read more
Remote only
0 - 3 yrs
₹0.20000000000000004 - ₹1.2 / mo
PyTorch
TensorFlow
OpenCV
FFmpeg
skill iconDeep Learning
+11 more

About the Role

We are looking for a passionate AI Engineer Intern (B.Tech, M.Tech / M.S. or equivalent) with strong foundations in Artificial Intelligence, Computer Vision, and Deep Learning to join our R&D team.

You will help us build and train realistic face-swap and deepfake video models, powering the next generation of AI-driven video synthesis technology.

This is a remote, individual-contributor role offering exposure to cutting-edge AI model development in a startup-like environment.


Key Responsibilities

  • Research, implement, and fine-tune face-swap / deepfake architectures (e.g., FaceSwap, SimSwap, DeepFaceLab, LatentSync, Wav2Lip).
  • Train and optimize models for realistic facial reenactment and temporal consistency.
  • Work with GANs, VAEs, and diffusion models for video synthesis.
  • Handle dataset creation, cleaning, and augmentation for face-video tasks.
  • Collaborate with the AI core team to deploy trained models in production environments.
  • Maintain clean, modular, and reproducible pipelines using Git and experiment-tracking tools.

Required Qualifications

  • B.Tech, M.Tech / M.S. (or equivalent) in AI / ML / Computer Vision / Deep Learning.
  • Certifications in AI or Deep Learning (DeepLearning.AI, NVIDIA DLI, Coursera, etc.).
  • Proficiency in PyTorch or TensorFlow, OpenCV, FFmpeg.
  • Understanding of CNNs, Autoencoders, GANs, Diffusion Models.
  • Familiarity with datasets like CelebA, VoxCeleb, FFHQ, DFDC, etc.
  • Good grasp of data preprocessing, model evaluation, and performance tuning.

Preferred Skills

  • Prior hands-on experience with face-swap or lip-sync frameworks.
  • Exposure to 3D morphable models, NeRF, motion transfer, or facial landmark tracking.
  • Knowledge of multi-GPU training and model optimization.
  • Familiarity with Rust / Python backend integration for inference pipelines.

What We Offer

  • Work directly on production-grade AI video synthesis systems.
  • Remote-first, flexible working hours.
  • Mentorship from senior AI researchers and engineers.
  • Opportunity to transition into a full-time role upon outstanding performance.


Location: Remote | Stipend: ₹10,000/month | Duration: 3–6 months

Read more
Remote only
5 - 20 yrs
₹12L - ₹25L / yr
skill iconMachine Learning (ML)
Artificial Intelligence (AI)
skill iconPython
Generative AI
Large Language Models (LLM)

We are building an AI-powered chatbot platform and looking for an AI/ML Engineer with strong backend skills as our first technical hire. You will be responsible for developing the core chatbot engine using LLMs, creating backend APIs, and building scalable RAG pipelines.

You should be comfortable working independently, shipping fast, and turning ideas into real product features. This role is ideal for someone who loves building with modern AI tools and wants to be part of a fast-growing product from day one.

Responsibilities

• Build the core AI chatbot engine using LLMs (OpenAI, Claude, Gemini, Llama etc.)

• Develop backend services and APIs using Python (FastAPI/Flask)

• Create RAG pipelines using vector databases (Pinecone, FAISS, Chroma)

• Implement embeddings, prompt flows, and conversation logic

• Integrate chatbot with web apps, WhatsApp, CRMs and 3rd-party APIs

• Ensure system reliability, performance, and scalability

• Work directly with the founder in shaping the product and roadmap

Requirements

• Strong experience with LLMs & Generative AI

• Excellent Python skills with FastAPI/Flask

• Hands-on experience with LangChain or RAG architectures

• Vector database experience (Pinecone/FAISS/Chroma)

• Strong understanding of REST APIs and backend development

• Ability to work independently, experiment fast, and deliver clean code

Nice to Have

• Experience with cloud (AWS/GCP)

• Node.js knowledge

• LangGraph, LlamaIndex

• MLOps or deployment experience


Read more
Remote only
6 - 15 yrs
₹10L - ₹30L / yr
skill iconNextJs (Next.js)
skill iconFlutter
FastAPI
skill iconAmazon Web Services (AWS)
TypeScript
+8 more

Mission

Own architecture across web + backend, ship reliably, and establish patterns the team can scale on.

Responsibilities

  • Lead system architecture for Next.js (web) and FastAPI (backend); own code quality, reviews, and release cadence.
  • Build and maintain the web app (marketing, auth, dashboard) and a shared TS SDK (@revilo/contracts, @revilo/sdk).
  • Integrate StripeMaps, analytics; enforce accessibility and performance baselines.
  • Define CI/CD (GitHub Actions), containerization (Docker), env/promotions (staging → prod).
  • Partner with Mobile and AI engineers on API/tool schemas and developer experience.

Requirements

  • 6–10+ years; expert TypeScript, strong Python.
  • Next.js (App Router), TanStack Query, shadcn/ui; FastAPI, Postgres, pydantic/SQLModel.
  • Auth (OTP/JWT/OAuth), payments, caching, pagination, API versioning.
  • Practical CI/CD and observability (logs/metrics/traces).

Nice-to-haves

  • OpenAPI typegen (Zod), feature flags, background jobs/queues, Vercel/EAS.

Key Outcomes (ongoing)

  • Stable architecture with typed contracts; <2% crash/error on web, p95 API latency in budget, reliable weekly releases.


Read more
Beyond Seek Technologies Pvt Ltd
Remote only
4 - 7 yrs
₹12L - ₹24L / yr
Process automation
RPA
skill iconPython
AWS Lambda
AWS Simple Queuing Service (SQS)
+7 more

Responsibilities:


  • Design, build, and maintain backend services and APIs using Python frameworks such as FastAPI or Django.
  • Implement RAG-based features and services, including document ingestion pipelines, vector indexing, and retrieval logic using modern LLM tooling.
  • Build robust data ingestion, scraping, and automation workflows (web scraping, headless browsers, APIs) to integrate with third-party systems and internal tools.
  • Develop and operate ETL/ELT pipelines to move, clean, and transform data across databases, file stores, and external platforms.
  • Own reliability, performance, and observability of services: logging, metrics, alerting, and debugging in production.
  • Collaborate closely with product and business stakeholders to translate ambiguous workflows into clear technical designs and automation logic.
  • Write clean, testable code with solid unit/integration coverage, and contribute to internal libraries, tooling, and best practices documentation.
  • Participate in code reviews, architectural discussions, and mentor junior engineers on Python, RAG patterns, and automation best practices.


Requirements:


  • 4–6 years of hands-on experience as a Python engineer building production systems (FastAPI, Django, or similar).
  • Strong understanding of backend fundamentals: REST APIs, authentication/authorisation, async patterns, background jobs, and task queues.
  • Experience with event-driven architectures (Kafka, SQS, RabbitMQ) and workflow engines (e.g., Temporal, Airflow, Prefect).
  • Practical experience with at least one LLM/RAG stack (e.g., LangChain, LlamaIndex, custom vector store integrations) and working with embeddings, chunking, and retrieval.
  • Solid experience with web scraping and automation: requests/HTTP clients, Selenium/Playwright or similar, rate-limiting, anti-bot handling, and resilient scraping patterns.
  • Experience building data pipelines or ETLs: extracting from APIs/files/DBs, transforming/cleaning, and loading into relational or NoSQL stores.
  • Hands-on experience with AWS or similar cloud platforms (e.g., Lambda, S3, API Gateway, ECS/Fargate, or equivalent).
  • Strong debugging skills and comfort with distributed, asynchronous systems and eventual consistency.
  • Ability to take loosely defined business workflows and design clean, maintainable technical solutions.
  • Strong communication skills and a habit of documenting decisions, APIs, and workflows.


Nice to Have:


  • Experience with vector databases (e.g., Pinecone, Weaviate, Qdrant, OpenSearch vector, etc.) and search tuning.
  • Exposure to building internal tools or low-code-like automation platforms for operations or support teams.
  • Prior experience integrating with ERP/CRM/marketplace or other enterprise/legacy systems.
  • Understanding of cloud security, IAM, and secret management best practices.
Read more
Remote only
2 - 4 yrs
₹4L - ₹8L / yr
skill iconPython
JSON
LLMS
oops
skill iconJava
+4 more

Role Overview

We are seeking a Junior Developer with 1-3 year’s experience with strong foundations in Python, databases, and AI technologies. The ideal candidate will support the development of AI-powered solutions, focusing on LLM integration, prompt engineering, and database-driven workflows. This is a hands-on role with opportunities to learn and grow into advanced AI engineering responsibilities.

Key Responsibilities

  • Develop, test, and maintain Python-based applications and APIs.
  • Design and optimize prompts for Large Language Models (LLMs) to improve accuracy and performance.
  • Work with JSON-based data structures for request/response handling.
  • Integrate and manage PostgreSQL (pgSQL) databases, including writing queries and handling data pipelines.
  • Collaborate with the product and AI teams to implement new features.
  • Debug, troubleshoot, and optimize performance of applications and workflows.
  • Stay updated on advancements in LLMs, AI frameworks, and generative AI tools.

Required Skills & Qualifications

  • Strong knowledge of Python (scripting, APIs, data handling).
  • Basic understanding of Large Language Models (LLMs) and prompt engineering techniques.
  • Experience with JSON data parsing and transformations.
  • Familiarity with PostgreSQL or other relational databases.
  • Ability to write clean, maintainable, and well-documented code.
  • Strong problem-solving skills and eagerness to learn.
  • Bachelor’s degree in Computer Science, Engineering, or related field (or equivalent practical experience).

Nice-to-Have (Preferred)

  • Exposure to AI/ML frameworks (e.g., LangChain, Hugging Face, OpenAI APIs).
  • Experience working in startups or fast-paced environments.
  • Familiarity with version control (Git/GitHub) and cloud platforms (AWS, GCP, or Azure).

What We Offer

  • Opportunity to work on cutting-edge AI applications in permitting & compliance.
  • Collaborative, growth-focused, and innovation-driven work culture.
  • Mentorship and learning opportunities in AI/LLM development.
  • Competitive compensation with performance-based growth.


Read more
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Why apply via Cutshort?
Connect with actual hiring teams and get their fast response. No spam.
Find more jobs
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort