10+ PyTorch Jobs in Mumbai | PyTorch Job openings in Mumbai
Apply to 10+ PyTorch Jobs in Mumbai on CutShort.io. Explore the latest PyTorch Job opportunities across top companies like Google, Amazon & Adobe.
Artificial Intelligence Resercher (Computer vision)
Responsibility
• Work on Various SOTA Computer Vision Models, Dataset Augmentation & Dataset Generation
Techniques that help improve model accuracy & precision.
• Work on development & improvement of End-to-End Pipeline use cases running at scale.
• Programming skills with multi-threaded GPU CUDA computing and API Solutions.
• Proficient with Training of Detection, Classification & Segmentation Models with TensorFlow,
Pytorch, MX Net etc
Required Skills
• Strong development skills required in Python and C++.
• Ability to architect a solution based on given requirements and convert the business requirements into a technical computer vision problem statement.
• Ability to work in a fast-paced environment and coordinate across different parts of different projects.
• Bringing in the technical expertise around the implementation of best coding standards and
practices across the team.
• Extensive experience of working on edge devices like Jetson Nano, Raspberry Pi and other GPU powered low computational devices.
• Experience with using Docker, Nvidia Docker, Nvidia NGC containers for Computer Vision Deep
Learning
• Experience with Scalable Cloud Deployment Architecture for Video Analytics(Involving Kubernetes
and or Kafka)
• Good experience with any of one cloud technologies like AWS, Azure and Google Cloud.
• Experience in working with Model Optimisation for Nvidia Hardware (Tensors Conversion of both TensorFlow & Pytorch models.
• Proficient understanding of code versioning tools, such as Git.
• Proficient in Data Structures & Algorithms.
• Well versed in software design paradigms and good development practices.
• Experience with Scalable Cloud Deployment Architecture for Video Analytics(Involving Kubernetes
and or Kafka).
Responsibilities
- Work on execution and scheduling of all tasks related to assigned projects' deliverable dates
- Optimize and debug existing codes to make them scalable and improve performance
- Design, development, and delivery of tested code and machine learning models into production environments
- Work effectively in teams, managing and leading teams
- Provide effective, constructive feedback to the delivery leader
- Manage client expectations and work with an agile mindset with machine learning and AI technology
- Design and prototype data-driven solutions
Eligibility
- Highly experienced in designing, building, and shipping scalable and production-quality machine learning algorithms in the field of Python applications
- Working knowledge and experience in NLP core components (NER, Entity Disambiguation, etc.)
- In-depth expertise in Data Munging and Storage (Experienced in SQL, NoSQL, MongoDB, Graph Databases)
- Expertise in writing scalable APIs for machine learning models
- Experience with maintaining code logs, task schedulers, and security
- Working knowledge of machine learning techniques, feed-forward, recurrent and convolutional neural networks, entropy models, supervised and unsupervised learning
- Experience with at least one of the following: Keras, Tensorflow, Caffe, or PyTorch
The Role:
As an ML Engineer at TIFIN, you will be responsible for driving research and innovation in a result-oriented direction. Your role will involve staying updated on the latest research trends and exploring advancements in Natural Language Understanding (NLU) and their applications in conversational AI. You will also play a mentoring role and contribute to improving NLU capabilities using transfer learning from historical conversational data.
Requirements:
- Experience with training and fine-tuning machine learning models on large text datasets.
- Strong computer science fundamentals and at least 3 years of software development experience.
- Track record of thinking big and finding simple solutions while dealing with ambiguity.
- Proven experience as a Natural Language Processing Engineer or in similar roles.
- Good understanding of NLP tricks and techniques for semantic extraction, data structure, and modeling.
- Familiarity with text representation techniques, algorithms, and statistics.
- Knowledge of programming languages such as R, Python, and Java.
- Proficiency in Machine Learning frameworks like TensorFlow, PyTorch, etc.
- Ability to design software architectures and solve complex problems.
- Strong analytical and problem-solving skills.
- Experience in projects related to information retrieval, machine comprehension, entity recognition, text classification, semantic frame parsing, or machine translation is a plus.
- Publications, patents, or conference talks in relevant fields are a bonus.
About UpSolve
We built and deliver complex AI solutions which help drive business decisions faster and more accurately. We are a typical AI company and have a range of solutions developed on Video, Image and Text.
What you will do
- Stay informed on new technologies and implement cautiously
- Maintain necessary documentation for the project
- Fix the issues reported by application users
- Plan, build, and design solutions with a mental note of future requirements
- Coordinate with the development team to manage fixes, code changes, and merging
Location: Mumbai
Working Mode: Remote
What are we looking for
- Bachelor's or Master's degree in Computer Science, Software Engineering, or a related field.
- Minimum 2 years of professional experience in software development, with a focus on machine learning and full stack development.
- Strong proficiency in Python programming language and its machine learning libraries such as TensorFlow, PyTorch, or scikit-learn.
- Experience in developing and deploying machine learning models in production environments.
- Proficiency in web development technologies including HTML, CSS, JavaScript, and front-end frameworks such as React, Angular, or Vue.js.
- Experience in designing and developing RESTful APIs and backend services using frameworks like Flask or Django.
- Knowledge of databases and SQL for data storage and retrieval.
- Familiarity with version control systems such as Git.
- Strong problem-solving and analytical skills.
- Excellent communication and collaboration abilities.
- Ability to work effectively in a fast-paced and dynamic team environment.
- Good to have Cloud Exposure
Accrete.ai
Responsibilities:
- Data science model review, run the code refactoring and optimization, containerization, deployment, versioning, and monitoring of its quality.
- Design and implement cloud solutions, build MLOps on the cloud (preferably AWS)
- Work with workflow orchestration tools like Kubeflow, Airflow, Argo, or similar tools
- Data science models testing, validation, and test automation.
- Communicate with a team of data scientists, data engineers, and architects, and document the processes.
Eligibility:
- Rich hands-on experience in writing object-oriented code using python
- Min 3 years of MLOps experience (Including model versioning, model and data lineage, monitoring, model hosting and deployment, scalability, orchestration, continuous learning, and Automated pipelines)
- Understanding of Data Structures, Data Systems, and software architecture
- Experience in using MLOps frameworks like Kubeflow, MLFlow, and Airflow Pipelines for building, deploying, and managing multi-step ML workflows based on Docker containers and Kubernetes.
- Exposure to deep learning approaches and modeling frameworks (PyTorch, Tensorflow, Keras, etc. )
Who Are We
A research-oriented company with expertise in computer vision and artificial intelligence, at its core, Orbo is a comprehensive platform of AI-based visual enhancement stack. This way, companies can find a suitable product as per their need where deep learning powered technology can automatically improve their Imagery.
ORBO's solutions are helping BFSI, beauty and personal care digital transformation and Ecommerce image retouching industries in multiple ways.
WHY US
- Join top AI company
- Grow with your best companions
- Continuous pursuit of excellence, equality, respect
- Competitive compensation and benefits
You'll be a part of the core team and will be working directly with the founders in building and iterating upon the core products that make cameras intelligent and images more informative.
To learn more about how we work, please check out
Description:
We are looking for a computer vision engineer to lead our team in developing a factory floor analytics SaaS product. This would be a fast-paced role and the person will get an opportunity to develop an industrial grade solution from concept to deployment.
Responsibilities:
- Research and develop computer vision solutions for industries (BFSI, Beauty and personal care, E-commerce, Defence etc.)
- Lead a team of ML engineers in developing an industrial AI product from scratch
- Setup end-end Deep Learning pipeline for data ingestion, preparation, model training, validation and deployment
- Tune the models to achieve high accuracy rates and minimum latency
- Deploying developed computer vision models on edge devices after optimization to meet customer requirements
Requirements:
- Bachelor’s degree
- Understanding about depth and breadth of computer vision and deep learning algorithms.
- Experience in taking an AI product from scratch to commercial deployment.
- Experience in Image enhancement, object detection, image segmentation, image classification algorithms
- Experience in deployment with OpenVINO, ONNXruntime and TensorRT
- Experience in deploying computer vision solutions on edge devices such as Intel Movidius and Nvidia Jetson
- Experience with any machine/deep learning frameworks like Tensorflow, and PyTorch.
- Proficient understanding of code versioning tools, such as Git
Our perfect candidate is someone that:
- is proactive and an independent problem solver
- is a constant learner. We are a fast growing start-up. We want you to grow with us!
- is a team player and good communicator
What We Offer:
- You will have fun working with a fast-paced team on a product that can impact the business model of E-commerce and BFSI industries. As the team is small, you will easily be able to see a direct impact of what you build on our customers (Trust us - it is extremely fulfilling!)
- You will be in charge of what you build and be an integral part of the product development process
- Technical and financial growth!
Who Are We
A research-oriented company with expertise in computer vision and artificial intelligence, at its core, Orbo is a comprehensive platform of AI-based visual enhancement stack. This way, companies can find a suitable product as per their need where deep learning powered technology can automatically improve their Imagery.
ORBO's solutions are helping BFSI, beauty and personal care digital transformation and Ecommerce image retouching industries in multiple ways.
WHY US
- Join top AI company
- Grow with your best companions
- Continuous pursuit of excellence, equality, respect
- Competitive compensation and benefits
You'll be a part of the core team and will be working directly with the founders in building and iterating upon the core products that make cameras intelligent and images more informative.
To learn more about how we work, please check out
Description:
We are looking for a computer vision engineer to lead our team in developing a factory floor analytics SaaS product. This would be a fast-paced role and the person will get an opportunity to develop an industrial grade solution from concept to deployment.
Responsibilities:
- Research and develop computer vision solutions for industries (BFSI, Beauty and personal care, E-commerce, Defence etc.)
- Lead a team of ML engineers in developing an industrial AI product from scratch
- Setup end-end Deep Learning pipeline for data ingestion, preparation, model training, validation and deployment
- Tune the models to achieve high accuracy rates and minimum latency
- Deploying developed computer vision models on edge devices after optimization to meet customer requirements
Requirements:
- Bachelor’s degree
- Understanding about depth and breadth of computer vision and deep learning algorithms.
- 4+ years of industrial experience in computer vision and/or deep learning
- Experience in taking an AI product from scratch to commercial deployment.
- Experience in Image enhancement, object detection, image segmentation, image classification algorithms
- Experience in deployment with OpenVINO, ONNXruntime and TensorRT
- Experience in deploying computer vision solutions on edge devices such as Intel Movidius and Nvidia Jetson
- Experience with any machine/deep learning frameworks like Tensorflow, and PyTorch.
- Proficient understanding of code versioning tools, such as Git
Our perfect candidate is someone that:
- is proactive and an independent problem solver
- is a constant learner. We are a fast growing start-up. We want you to grow with us!
- is a team player and good communicator
What We Offer:
- You will have fun working with a fast-paced team on a product that can impact the business model of E-commerce and BFSI industries. As the team is small, you will easily be able to see a direct impact of what you build on our customers (Trust us - it is extremely fulfilling!)
- You will be in charge of what you build and be an integral part of the product development process
- Technical and financial growth!
Responsibilities
At Dolat, code is our business, so naturally, the Core Engineering and Systems team is at the center of what we do. Our community of developers has designed and continues to enhance one of the fastest
trading platforms using the latest tools and technologies. As a Software Developer, you’ll draw upon your computer science, mathematical, and analytical abilities to develop complex and nimble code used to grow our business and increase the efficiency of the global financial markets.
Your responsibilities may include any of the following, which will require you to exercise discretion and independent judgment:
• Augmenting, improving, redesigning, and/or re-implementing Dolat's low-latency/highthroughput production trading environment, which collects data from and disseminates
orders to exchanges around the world
• Optimizing this platform by using network and systems programming, as well as other advanced techniques
• Developing systems that provide easy access to historical market data and trading simulations
• Building risk-management and performance-tracking tools
• Shaping the future of Dolat through regular interviewing and infrequent campus recruiting trips
• Implementing domain-optimized data structures
• Learn and internalize the theories behind current trading system
• Participate in the design, architecture and implementation of automated trading systems
• Take ownership of system from design through implementation
Job Type: Full-time
CTC Offering : 3.6L PA to 6L PA
Job Location: Remote for 6-9 months due to the pandemic, then Mumbai, Maharashtra
Required experience:
-
Minimum 1.5 to 2 year of experience in Web & Backend Development using Python with experience in some form of Machine Learning ML Algorithms
Overview
We are looking for Python developers with a strong understanding of object orientation and experience in web and backend development. Experience with Analytical algorithms and mathematical calculations using libraries such as Numpy and Pandas are a must. Experience in some form of Machine Learning. We require candidates who have working experience using Django Framework
Key Skills required (Items in Bold are mandatory keywords) :
1. Proficiency in Python 3.x based web and backend development
2. Solid understanding of Python concepts
3. Experience with some form of Machine Learning (ML)
4. Experience in using libraries such as Numpy and Pandas
5. Some form of experience with NLP and Deep Learning using any of Pytorch, Tensorflow, Keras, Scikit-learn or similar
6. Hands on experience with RDBMS such as Postgres or MySQL
7. Experience building REST APIs using DRF or Flask
8. Comfort with Git repositories, branching and deployment using Git
9. Working experience with Docker
10. Basic working knowledge of ReactJs
11. Experience in deploying Django applications to AWS,Digital Ocean or Heroku
KRAs includes:
1. Understanding the scope of work
2. Understanding and adopting the current internal development workflow and processes
3. Understanding client requirements as communicated by the project manager
4. Arriving on timelines for projects, either independently or as a part of a team
5. Executing projects either independently or as a part of a team
6. Developing products and projects using Python
7. Writing code to collect and mathematically analyse large volumes of data.
8. Creating backend modules in Python by building or reutilizing existing modules in a manner so as to
provide optimal deliveries on time
9. Writing Scalable, maintainable code
10. Building secured REST APIs
11. Setting up batch task processing environments using Celery
12. Unit testing prepared modules
13. Bug fixing issues as reported by the QA team
14. Optimization and performance tuning of code
Bonus but not mandatory
1. Nodejs
2. Redis
3. PHP
4. CI/CD
5. AWS
Job Description
We are looking for an experienced engineer to join our data science team, who will help us design, develop, and deploy machine learning models in production. You will develop robust models, prepare their deployment into production in a controlled manner, while providing appropriate means to monitor their performance and stability after deployment.
What You’ll Do will include (But not limited to):
- Preparing datasets needed to train and validate our machine learning models
- Anticipate and build solutions for problems that interrupt availability, performance, and stability in our systems, services, and products at scale.
- Defining and implementing metrics to evaluate the performance of the models, both for computing performance (such as CPU & memory usage) and for ML performance (such as precision, recall, and F1)
- Supporting the deployment of machine learning models on our infrastructure, including containerization, instrumentation, and versioning
- Supporting the whole lifecycle of our machine learning models, including gathering data for retraining, A/B testing, and redeployments
- Developing, testing, and evaluating tools for machine learning models deployment, monitoring, retraining.
- Working closely within a distributed team to analyze and apply innovative solutions over billions of documents
- Supporting solutions ranging from rule-bases, classical ML techniques to the latest deep learning systems.
- Partnering with cross-functional team members to bring large scale data engineering solutions to production
- Communicating your approach and results to a wider audience through presentations
Your Qualifications:
- Demonstrated success with machine learning in a SaaS or Cloud environment, with hands–on knowledge of model creation and deployments in production at scale
- Good knowledge of traditional machine learning methods and neural networks
- Experience with practical machine learning modeling, especially on time-series forecasting, analysis, and causal inference.
- Experience with data mining algorithms and statistical modeling techniques for anomaly detection in time series such as clustering, classification, ARIMA, and decision trees is preferred.
- Ability to implement data import, cleansing and transformation functions at scale
- Fluency in Docker, Kubernetes
- Working knowledge of relational and dimensional data models with appropriate visualization techniques such as PCA.
- Solid English skills to effectively communicate with other team members
Due to the nature of the role, it would be nice if you have also:
- Experience with large datasets and distributed computing, especially with the Google Cloud Platform
- Fluency in at least one deep learning framework: PyTorch, TensorFlow / Keras
- Experience with No–SQL and Graph databases
- Experience working in a Colab, Jupyter, or Python notebook environment
- Some experience with monitoring, analysis, and alerting tools like New Relic, Prometheus, and the ELK stack
- Knowledge of Java, Scala or Go-Lang programming languages
- Familiarity with KubeFlow
- Experience with transformers, for example the Hugging Face libraries
- Experience with OpenCV
About Egnyte
In a content critical age, Egnyte fuels business growth by enabling content-rich business processes, while also providing organizations with visibility and control over their content assets. Egnyte’s cloud-native content services platform leverages the industry’s leading content intelligence engine to deliver a simple, secure, and vendor-neutral foundation for managing enterprise content across business applications and storage repositories. More than 16,000 customers trust Egnyte to enhance employee productivity, automate data management, and reduce file-sharing cost and complexity. Investors include Google Ventures, Kleiner Perkins, Caufield & Byers, and Goldman Sachs. For more information, visit www.egnyte.com
#LI-Remote