
Python Developer
at Altimetrik
Experience : 5-8 years
Location : Bangalore,Chennai and hyderabad
Python Developer (1 Position)
Must have skills:
· Experience in advanced Python
· Experience in GUI/Test Automation tools/libraries (Robot Framework, Selenium, & Sikuli etc.)
· Ability to create UI Automation scripts to execute in Remote/Citrix servers
· Knowledge in analytical libraries like Pandas, Numpy, Scipy, PyTorch etc.
· AWS skillset
Nice to have skills:
· Experience in SQL and Big Data analytic tools like Hive and Hue
· Experience in Machine learning
· Experience in Linux administration

Similar jobs
About Us
CLOUDSUFI, a Google Cloud Premier Partner, a Data Science and Product Engineering organization building Products and Solutions for Technology and Enterprise industries. We firmly believe in the power of data to transform businesses and make better decisions. We combine unmatched experience in business processes with cutting edge infrastructure and cloud services. We partner with our customers to monetize their data and make enterprise data dance.
Our Values
We are a passionate and empathetic team that prioritizes human values. Our purpose is to elevate the quality of lives for our family, customers, partners and the community.
Equal Opportunity Statement
CLOUDSUFI is an equal opportunity employer. We celebrate diversity and are committed to creating an inclusive environment for all employees. All qualified candidates receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, and national origin status. We provide equal opportunities in employment, advancement, and all other areas of our workplace. Please explore more at https://www.cloudsufi.com/.
About Role:
The Senior Python Developer will lead the design and implementation of ACL crawler connectors for Workato’s search platform. This role requires deep expertise in building scalable Python services, integrating with various SaaS APIs and designing robust data models. The developer will mentor junior team members and ensure that the solutions meet the technical and performance requirements outlined in the Statement of Work.
Key Responsibilities:
- Architecture and design: Translate business requirements into technical designs for ACL crawler connectors. Define data models, API interactions and modular components using the Workato SDK.
- Implementation: Build Python services to authenticate, enumerate domain entities and extract ACL information from OneDrive, ServiceNow, HubSpot and GitHub. Implement incremental sync, pagination, concurrency and caching.
- Performance optimisation: Profile code, parallelise API calls and utilise asynchronous programming to meet crawl time SLAs. Implement retry logic and error handling for network‑bound operations.
- Testing and code quality: Develop unit and integration tests, perform code reviews and enforce best practices (type hints, linting). Produce performance reports and documentation.
- Mentoring and collaboration: Guide junior developers, collaborate with QA, DevOps and product teams, and participate in design reviews and sprint planning.
- Hypercare support: Provide Level 2/3 support during the initial rollout, troubleshoot issues, implement minor enhancements and deliver knowledge transfer sessions.
Must Have Skills and Experiences:
- Bachelor’s degree in Computer Science or related field.
- 3-8 years of Python development experience, including asynchronous programming and API integration.
- Knowledge of python libraries-pandas,pytest,requests,asyncio
- Strong understanding of authentication protocols (OAuth 2.0, API keys) and access‑control models.
- Experience with integration with cloud or SaaS platforms such as Microsoft Graph, ServiceNow REST API, HubSpot API, GitHub API.
- Proven ability to lead projects and mentor other engineers.
- Excellent communication skills and ability to produce clear documentation.
Optional/Good to Have Skills and Experiences:
- Experience with integration with Microsoft Graph API, ServiceNow REST API, HubSpot API, GitHub API.
- Familiarity with the following libraries, tools and technologies will be advantageous-aiohttp,PyJWT,aiofiles / aiocache
- Experience with containerisation (Docker), CI/CD pipelines and Workato’s connector SDK is also considered a plus.
Position: Lead Python Developer
Location: Ahmedabad
The Client company includes a team of experienced information services professionals who are
passionate about growing and enhancing the value of information services businesses. They provide
support with talent, technology, tools, infrastructure and expertise required to deliver across the Data ecosystem.
Position Summary
We are seeking a skilled and experienced Backend Developer with strong expertise in TypeScript, Python,
and web scraping. You will be responsible for designing, developing, and maintaining scalable backend services and APIs that power our data-driven products. Your role will involve collaborating with cross-
functional teams, optimizing system performance, ensuring data integrity, and contributing to the design of efficient and secure architectures.
Job Responsibility
● Design, develop, and maintain backend systems and services using Python and TypeScript.
● Develop and maintain web scraping solutions to extract, process, and manage large-scale data from multiple sources.
● Work with relational and non-relational databases, ensuring high availability, scalability, and performance.
● Implement authentication, authorization, and security best practices across services.
● Write clean, maintainable, and testable code following best practices and coding standards.
● Collaborate with frontend engineers, data engineers, and DevOps teams to deliver robust solutions and troubleshoot, debug, and upgrade existing applications.
● Stay updated with backend development trends, tools, and frameworks to continuously improve processes.
● Utilize core crawling experience to design efficient strategies for scraping the data from different websites and applications.
● Collaborate with technology teams, data collection teams to build end to end technology-enabled ecosystems and partner in research projects to analyze the massive data inputs.
● Responsible for the design and development of web crawlers, able to independently solve various problems encountered in the actual development process.
● Stay updated with the latest web scraping techniques, tools, and industry trends to continuously improve the scraping processes.
Job Requirements
● 4+ years of professional experience in backend development with TypeScript and Python.
● Strong understanding of TypeScript-based server-side frameworks (e.g., Node.js, NestJS, Express) and Python frameworks (e.g., FastAPI, Django, Flask).
● Experience with tools and libraries for web scraping (e.g., Scrapy, BeautifulSoup, Selenium, Puppeteer)
● Hands-on experience with Temporal for creating and orchestrating workflows.
● Proven hands-on experience in web scraping, including crawling, data extraction, deduplication, and handling dynamic websites.
● Proficient in implementing proxy solutions and handling bot-detection challenges (e.g., Cloudflare).
● Experience working with Docker, containerized deployments, and cloud environments (GCP or Azure).
● Proficiency with database systems such as MongoDB and ElasticSearch.
● Hands-on experience with designing and maintaining scalable APIs.
● Knowledge of software testing practices (unit, integration, end-to-end).
● Familiarity with CI/CD pipelines and version control systems (Git).
● Strong problem-solving skills, attention to detail, and ability to work in agile environments.
● Great communication skills and ability to navigate in undirected situations.
Job Exposure:
● Opportunity to apply creative methods in acquiring and filtering the North American government, agencies data from various websites, sources
● In depth industry exposure on data harvesting techniques to build, scale the robust and sustainable model, using open-source applications
● Effectively collaboration with IT team to design the tailor-made solutions basis upon clients’ requirement
● Unique opportunity to research on various agencies, vendors, products as well as technology tools to compose a solution
Position: Lead Python Developer
Location: Ahmedabad, Gujarat
The Client company includes a team of experienced information services professionals who are passionate about growing and enhancing the value of information services businesses. They provide support with talent, technology, tools, infrastructure and expertise required to deliver across the Data ecosystem. Position Summary We are seeking a skilled and experienced Backend Developer with strong expertise in TypeScript, Python, and web scraping. You will be responsible for designing, developing, and maintaining scalable backend services and APIs that power our data-driven products. Your role will involve collaborating with cross functional teams, optimizing system performance, ensuring data integrity, and contributing to the design of efficient and secure architectures.
Job Responsibility
● Design, develop, and maintain backend systems and services using Python and TypeScript.
● Develop and maintain web scraping solutions to extract, process, and manage large-scale data from multiple sources.
● Work with relational and non-relational
databases, ensuring high availability, scalability, and performance.
● Implement authentication, authorization, and security best practices across services.
● Write clean, maintainable, and testable code following best practices and coding standards.
● Collaborate with frontend engineers, data engineers, and DevOps teams to deliver robust solutions and troubleshoot, debug, and upgrade existing applications.
● Stay updated with backend development trends, tools, and frameworks to continuously improve processes.
● Utilize core crawling experience to design efficient strategies for scraping the data from different websites and applications.
● Collaborate with technology teams, data collection teams to build end to end technology-enabled ecosystems and partner in research projects to analyze the massive data inputs.
● Responsible for the design and development of web crawlers, able to independently solve various problems encountered in the actual development process.
● Stay updated with the latest web scraping techniques, tools, and industry trends to continuously improve the scraping processes.
Job Requirements
● 4+ years of professional experience in backend development with TypeScript and Python.
● Strong understanding of TypeScript-based server-side frameworks (e.g., Node.js, NestJS, Express) and Python frameworks (e.g., FastAPI, Django, Flask).
● Experience with tools and libraries for web scraping (e.g., Scrapy, BeautifulSoup, Selenium, Puppeteer)
● Hands-on experience with Temporal for creating and orchestrating workflows
● Proven hands-on experience in web scraping, including crawling, data extraction, deduplication, and handling dynamic websites.
● Proficient in implementing proxy solutions and handling bot-detection challenges (e.g., Cloudflare).
● Experience working with Docker, containerized deployments, and cloud environments (GCP or Azure).
● Proficiency with database systems such as MongoDB and Elastic Search.
● Hands-on experience with designing and maintaining scalable APIs.
● Knowledge of software testing practices (unit, integration, end-to-end).
● Familiarity with CI/CD pipelines and version control systems (Git).
● Strong problem-solving skills, attention to detail, and ability to work in agile environments.
● Great communication skills and ability to navigate in undirected situations.
Job Exposure:
● Opportunity to apply creative methods in acquiring and filtering the North American government, agencies data from various websites, sources
● In depth industry exposure on data harvesting techniques to build, scale the robust and sustainable model, using open-source applications ● Effectively collaboration with IT team to design the tailor-made solutions basis upon clients’ requirement
● Unique opportunity to research on various agencies, vendors, products as well as technology tools to compose a solution
Job Title : Python Django Developer
Experience : 3+ Years
Location : Gurgaon
Working Days : 6 Days (Monday to Saturday)
Job Summary :
We are looking for a skilled Python Django Developer with strong foundational knowledge in backend development, data structures, and operating system concepts.
The ideal candidate should have experience in Django and PostgreSQL, along with excellent logical thinking and multithreading knowledge.
Technical Skills : Python, Django (or Flask), PostgreSQL/MySQL, SQL & NoSQL ORM, REST API development, JSON/XML, strong knowledge of data structures, multithreading, and OS concepts.
Key Responsibilities :
- Write efficient, reusable, testable, and scalable code using the Django framework.
- Develop backend components, server-side logic, and statistical models.
- Design and implement high-availability, low-latency applications with robust data protection and security.
- Contribute to the development of highly responsive web applications.
- Collaborate with cross-functional teams on system design and integration.
Mandatory Skills :
- Strong programming skills in Python and Django (or similar frameworks like Flask).
- Proficiency with PostgreSQL / MySQL and experience in writing complex queries.
- Strong understanding of SQL and NoSQL ORM.
- Solid grasp of data structures, multithreading, and operating system concepts.
- Experience with RESTful API development and implementation of API security.
- Knowledge of JSON/XML and their use in data exchange.
Good-to-Have Skills :
- Experience with Redis, MQTT, and message queues like RabbitMQ or Kafka
- Understanding of microservice architecture and third-party API integrations (e.g., payment gateways, SMS/email APIs)
- Familiarity with MongoDB and other NoSQL databases
- Exposure to data science libraries such as Pandas, NumPy, Scikit-learn
- Knowledge in building and integrating statistical learning models.
Role / Designation : Python Developer
Location: Bangalore, India
Skills : Certification: AI900, AZ900 Technical or Key Skills: Primary Skills Python, Flask, Web development. Knowledge on Azure Cloud, Application development, API development
Profile: IT Professional with 6 +years of experience in
• Hands on experience Python libraries such as Pandas, Numpy , OpenPyxl
• Hands on experience of Python libraries with multiple document types (excel, csv, pdf and images)
• Working with huge data sets, data analysis and provide ETL and EDA analysis reports.
• 5+ years’ experience in any of the programming languages like Python(mandatory), Java and C/C++.
• Must have experience in Azure Paas, Iaas services like Azure function app, Kubernetes services, storage account, keyvault , etc
• Experience with DB such as SQL,NoSQL
• Develop methodologies for the data analysis, data extraction, data transformations, preprocessing of data.
• Experience in deploying applications, packages in Azure environment.
• Writing scalable code using Python programming language.
• Testing and debugging applications.
• Developing back-end components.
• Integrating user-facing elements using server-side logic.
• Excellent problem solving/analytical skills and complex troubleshooting methods.
• Ability to work through ambiguous situations.
• Excellent presentation, verbal, and written communication skills. Education: BE/BTech/BSc
Certification: AI900, AZ900 Technical or Key Skills: Primary Skills Python, Flask, Web development. Knowledge on Azure Cloud, Application development, API development
· The Objective:
You will play a crucial role in designing, implementing, and maintaining our data infrastructure, run tests and update the systems
· Job function and requirements
o Expert in Python, Pandas and Numpy with knowledge of Python web Framework such as Django and Flask.
o Able to integrate multiple data sources and databases into one system.
o Basic understanding of frontend technologies like HTML, CSS, JavaScript.
o Able to build data pipelines.
o Strong unit test and debugging skills.
o Understanding of fundamental design principles behind a scalable application
o Good understanding of RDBMS databases among Mysql or Postgresql.
o Able to analyze and transform raw data.
· About us
Mitibase helps companies find warm prospects every month that are most relevant, and then helps their team to act on those with automation. We do so by automatically tracking key accounts and contacts for job changes and relationships triggers and surfaces them as warm leads in your sales pipeline.
Job Description:
- 3 - 4 years of hands-on Python programming & libraries like PyData, Pandas
- Exposure to Mongo DB
- Experience in writing Unit Test cases
- Expertise in writing medium/advanced SQL Database queries
- Strong Verbal/Written communication skills
- Ability to work with onsite counterpart teams
Quantsapp is India's first Option Trading Analytics platform on mobile. With
ever-growing users,
it makes us one of the fastest growing platform for options trading in India.
Quantsapp wants to accelerate its growth even more and capture new
countries which requires the development team to grow.
At Qauntsapp we are looking for a dynamic team mate to take up a role of
Server side development to support the brain behind the application.
Job Summary :
- You will be responsible for developing new logics / products / features as
described by the business / research team.
- An ideal candidate should be strong in mathematical processes like
optimizations, matrix algebra, differential equation, simulation process etc.
and should also possess decent hands-on experience on python, sql server and
preferably Aws. IIT graduation is a plus.
Responsibilities :
- Create algorithms from scratch
- Create products and backend API's as described by the business team
- Back-test and create hypothesis as desired by the research team
- Code the backend of logics for consumption by the UI team
- Deploy websockets, rest api's & dynamic tcp, udp based data flow
- Deployment and maintenance of codes with version control
Requirements :
- Should possess a good knowledge of advanced computing and mathematical
process
- Strong hands-on on Python and optionally Matlab
- Knowledge of databases like Sql & No Sql
- Ability to work with tight time lines
- In depth knowledge and good hands on experience on Pandas and Numpy
- Knowledge of Option Markets is a plus
- Excellent organizational and multitasking ability
- Experience on AWS Cloud is a plus
Position: Senior Java Developer / Tech Lead
Work Location: Bangalore; India
Type of Employment: Full-time, Permanent
Experience: 6 to 10 years of Experience
- Proficient in Java/J2EE , core java, full stack development experience ( Database to UX)
- Must be proficient with integration with third-party applications within the enterprise and outside the organizations (API)
- Must be good in writing highly optimized code.
- Must be good at abstracting solutions - reduce time and cost to build
- Experience with Workflow Engines like Activiti will be an advantage
- Good understanding of native Android code/interface with Android apps, if preferable
- Must be an expert in managing data at the application database level - Transaction and Reporting database design
- Must be good in coding & product development
- Experienced with monitoring and observability
- Experienced with production support and debugging (heap dumps, thread dumps etc)
- Experience in Logistics domain is desirable
- Prior startup experience is preferred
Requirements:
- Bachelor’s or equivalent degree in Computer Science / BE
- Ability to handle the team of 7 to 8 members and giving guidance to the team
- Strong problem-solving skills
- Ability to respect deadlines and will to perform in an accelerated learning environment.
Key Skills: Java, J2EE, JSP, Spring, Angular, Apache, MySQL, SQL Queries,LinuxEnviroment, HTML, JQuery, Java Script, Version Control, .Net, Activiti, IBatis, Maven – Build, Git Hub, J Bos, CSS, HTML 5, JIRA, Selenium,AWS – Server Handling, Android, React JS
Note: Hiring from logistics and Ecommerce company only and 30-45 days’ notice is ok and can also buy back notice if required.
We are actively seeking software development engineers who are interested in designing robust trading systems and refining programs to efficiently manage various types of financial market data that facilitate our quantitative investment research. By designing and improving the firm's internal applications, the SDE will play a key role in expanding the firm's trading capabilities.
Responsibilities:
- Management & scaling up existing infrastructure for high-frequency market data capture.
- Develop a scalable and consistent data handling infrastructure for the above data to facilitate efficient backtesting of quantitative investment strategies.
- Perform R& D; to build a software platform in Python for backtesting various kind of investment strategies using the above databases.
- This will involve studying the strategy development process and performance evaluation metrics.
- Develop autopilot risk-management systems to monitor live performance of the Portfolio.
- Improve the existing algorithms to achieve better execution price and reduce the latency.
Requirements:
Our ideal candidate would have graduated with a degree in computer science from a top university with 1-3 years industry experience, along with:
- High Level of proficiency in Python and good knowledge of Matlab/C++/C#.
- Past experience in dealing with large datasets and Knowledge of database administration and network programming will be a plus.
- Well-versed in software engineering principles, frameworks and technologies.
- The ability to manage multiple tasks in a fast-paced environment.
- Excellent analytical and problem solving abilities.
- A keen interest in learning about the financial markets.











![[x]cube LABS](/_next/image?url=https%3A%2F%2Fcdnv2.cutshort.io%2Fcompany-static%2F639877aa0ad87e002533a1c5%2Fuser_uploaded_data%2Flogos%2Fx_whiteB_eeCk0gqs.png&w=256&q=75)
