
About AttainU
About
Company video


Connect with the team
Similar jobs
Procedure is hiring for Drover.
This is not a DevOps/SRE/cloud-migration role — this is a hands-on backend engineering and architecture role where you build the platform powering our hardware at scale.
About Drover
Ranching is getting harder. Increased labor costs and a volatile climate are placing mounting pressure to provide for a growing population. Drover is empowering ranchers to efficiently and sustainably feed the world by making it cheaper and easier to manage livestock, unlock productivity gains, and reduce carbon footprint with rotational grazing. Not only is this a $46B opportunity, you'll be working on a climate solution with the potential for real, meaningful impact.
We use patent-pending low-voltage electrical muscle stimulation (EMS) to steer and contain cows, replacing the need for physical fences or electric shock. We are building something that has never been done before, and we have hundreds of ranches on our waitlist.
Drover is founded by Callum Taylor (ex-Harvard), who comes from 5 generations of ranching, and Samuel Aubin, both of whom grew up in Australian ranching towns and have an intricate understanding of the problem space. We are well-funded and supported by Workshop Ventures, a VC firm with experience in building unicorn IoT companies.
We're looking to assemble a team of exceptional talent with a high eagerness to dive headfirst into understanding the challenges and opportunities within ranching.
About The Role
As our founding cloud engineer, you will be responsible for building and scaling the infrastructure that powers our IoT platform, connecting thousands of devices across ranches nationwide.
Because we are an early-stage startup, you will have high levels of ownership in what you build. You will play a pivotal part in architecting our cloud infrastructure, building robust APIs, and ensuring our systems can scale reliably. We are looking for someone who is excited about solving complex technical challenges at the intersection of IoT, agriculture, and cloud computing.
What You'll Do
- Develop Drover IoT cloud architecture from the ground up (it’s a green field project)
- Design and implement services to support wearable devices, mobile app, and backend API
- Implement data processing and storage pipelines
- Create and maintain Infrastructure-as-Code
- Support the engineering team across all aspects of early-stage development -- after all, this is a startup
Requirements
- 5+ years of experience developing cloud architecture on AWS
- In-depth understanding of various AWS services, especially those related to IoT
- Expertise in cloud-hosted, event-driven, serverless architectures
- Expertise in programming languages suitable for AWS micro-services (eg: TypeScript, Python)
- Experience with networking and socket programming
- Experience with Kubernetes or similar orchestration platforms
- Experience with Infrastructure-as-Code tools (e.g., Terraform, AWS CDK)
- Familiarity with relational databases (PostgreSQL)
- Familiarity with Continuous Integration and Continuous Deployment (CI/CD)
Nice To Have
- Bachelor’s or Master’s degree in Computer Science, Software Engineering, Electrical Engineering, or a related field
Application Deployment Engineers / Deployment Engineer – Video Analytics / CCTV Solutions / Application Implementation Engineer
Company Name
Paralaxiom Technologies Private Limited
Company Website
https://www.linkedin.com/company/paralaxiom
Company details
Paralaxiom Technologies is involved in deep learning algorithms to develop video analytics-based security and compliance applications. They offer OCR products and image classification tools enhanced by machine learning algorithms and robust statistical analysis. We are among the earliest practitioners of AI software and we have world-class credentials in these technologies. They offer products like Paralaxiom VAST(Video Analytics and Surveillance Toolkit) and Paralaxiom AMPLE(Paralaxiom natural language processing platform).
In today's world, all premises be it manufacturing, hospitals, offices, hotels, cities, airports, shops, warehouses etc. are covered with CCTV Cameras. Continuous monitoring through dedicated command center or e-surveillance is proving to be ineffective as well as costly to manage.
We have pioneered the use of AI / ML technologies to headlessly live monitor CCTV cameras to generate very accurate alerts, alarms & insights and deliver them directly to the right stakeholders for quick, proactive action.
We have worked closely with hundreds of customers from diverse industries, AI Hardware Partners, CCTV OEMs, VMSs, System Integrators & Consultants to bring to the world VAST, an enterprise-ready Video Surveillance as a Service (VSaaS) solution.
Location: Pune / Bangalore / Chennai
Mode of Working: Work From Office
Days of Working: 5 Days a week
Responsibilities
Position Overview:
Paralaxiom is a video analytics and machine vision company with its VAST product line a path-breaking product for safety, security, and operations for CCTV operations.
We are looking for application engineers for this product line.
Experience: 1 -2 years
Key Responsibilities:
1) Gathering information from customers on their needs and understanding how VAST software matches their desires
2) Design the solution and installation of the software
3 ) Making sure the VAST software continues to work properly after maintenance and testing
4) Take notes of all aspects of the application for future upgrades and maintenance
5) Troubleshooting the software
6) Training the end users
7) Excellent Knowledge of Python and shell scripting
8) Working knowledge of IP networking and troubleshooting
We need someone with 1-2 years of experience with the following skillset:
Great communication skills
Debugging and analytical skills
Knowledge of hardware and software integration will be a plus
Knowledge of Camera NVR will be a plus
Hands-on system and functional testing will be a plus
Interview process: 3 Video + Final Discussion - F2F
Python Software Engineer (3–5 Years Experience)
Location: [Pune]
Role Overview
We are seeking skilled Python engineers to join our core product team. You will work on backend services, API development, and system integrations, contributing to a codebase of over 250,000 Python lines and collaborating with frontend, DevOps, and native code teams.
Key Responsibilities
· Design, develop, and maintain scalable Python backend services and APIs
· Optimize performance and reliability of large, distributed systems
· Collaborate with frontend (JS/HTML/CSS) and native (C/C++/C#) teams
· Write unit/integration tests and participate in code reviews
· Troubleshoot production issues and implement robust solutions
Required Skills
· 3–5 years of professional Python development experience
· Strong understanding of OOP, design patterns, and modular code structure
· Experience with MongoDB (PyMongo), Mako, RESTful APIs, and asynchronous programming
· Familiarity with code quality tools (flake8, pylint) and test frameworks (pytest, unittest)
· Experience with Git and collaborative development workflows
· Ability to read and refactor large, multi-module codebases
Nice to Have
· Experience with web frameworks (web.py, Flask, Django)
· Knowledge of C/C++ or C# for cross-platform integrations
· Familiarity with CI/CD, Docker, and cloud deployment
· Exposure to security, encryption, or enterprise SaaS products
What We Offer
· Opportunity to work on a mission-critical, enterprise-scale product
· Collaborative, growth-oriented engineering culture
· Flexible work arrangements (remote/hybrid)
· Competitive compensation and benefits
Strong Senior Backend Engineer profiles
Mandatory (Experience 1) – Must have 5+ years of hands-on Backend Engineering experience building scalable, production-grade systems
Mandatory (Experience 2) – Must have strong backend development experience using one or more frameworks (FastAPI / Django (Python), Spring (Java), Express (Node.js).
Mandatory (Experience 3) – Must have deep understanding of relevant libraries, tools, and best practices within the chosen backend framework
Mandatory (Experience 4) – Must have strong experience with databases, including SQL and NoSQL, along with efficient data modeling and performance optimization
Mandatory (Experience 5) – Must have experience designing, building, and maintaining APIs, services, and backend systems, including system design and clean code practices
Mandatory (Domain) – Experience with financial systems, billing platforms, or fintech applications is highly preferred (fintech background is a strong plus)
Mandatory (Company) – Must have worked in product companies / startups, preferably Series A to Series D
Mandatory (Education) – Candidates from top engineering institutes (IITs, BITS, or equivalent Tier-1 colleges) are preferred
CTC: up to 20 LPA
Exp: 4 to 7 Years
Required Qualifications
- Bachelor's degree in Computer Science, Information Technology, or related field
- 4+ years of experience in software development
- Strong proficiency in Java with deep understanding of web technology stack
- Hands-on experience developing applications with Spring Boot framework
- Solid understanding of Python programming language with practical Flask framework experience
- Working knowledge of NATS server for messaging and streaming data
- Experience deploying and managing applications in Kubernetes
- Understanding of microservices architecture and RESTful API design
- Familiarity with containerization technologies (Docker)
- Experience with version control systems (Git)
Skills & Competencies
- Skills Java (Spring Boot, Spring Cloud, Spring Security)
- Python (Flask, SQL Alchemy, REST APIs)
- NATS messaging patterns (pub/sub, request/reply, queue groups)
- Kubernetes (deployments, services, ingress, ConfigMaps, Secrets)
- Web technologies (HTTP, REST, WebSocket, gRPC)
- Container orchestration and management
- Soft Skills Problem-solving and analytical thinking
Have you streamed a program on Disney+, watched your favorite binge-worthy series on Peacock or cheered your favorite team on during the World Cup from one of the 20 top streaming platforms around the globe? If the answer is yes, you’ve already benefitted from Conviva technology, helping the world’s leading streaming publishers deliver exceptional streaming experiences and grow their businesses.
Conviva is the only global streaming analytics platform for big data that collects, standardizes, and puts trillions of cross-screen, streaming data points in context, in real time. The Conviva platform provides comprehensive, continuous, census-level measurement through real-time, server side sessionization at unprecedented scale. If this sounds important, it is! We measure a global footprint of more than 500 million unique viewers in 180 countries watching 220 billion streams per year across 3 billion applications streaming on devices. With Conviva, customers get a unique level of actionability and scale from continuous streaming measurement insights and benchmarking across every stream, every screen, every second.
What you get to do in this role:
Work on extremely high scale RUST web services or backend systems.
Design and develop solutions for highly scalable web and backend systems.
Proactively identify and solve performance issues.
Maintain a high bar on code quality and unit testing.
What you bring to the role:
5+ years of hands-on software development experience.
At least 2+ years of RUST development experience.
Knowledge of cargo packages for kafka, redis etc.
Strong CS fundamentals, including system design, data structures and algorithms.
Expertise in backend and web services development.
Good analytical and troubleshooting skills.
What will help you stand out:
Experience working with large scale web services and applications.
Exposure to Golang, Scala or Java
Exposure to Big data systems like Kafka, Spark, Hadoop etc.
Underpinning the Conviva platform is a rich history of innovation. More than 60 patents represent award-winning technologies and standards, including first-of-its kind-innovations like time-state analytics and AI-automated data modeling, that surfaces actionable insights. By understanding real-world human experiences and having the ability to act within seconds of observation, our customers can solve business-critical issues and focus on growing their business ahead of the competition. Examples of the brands Conviva has helped fuel streaming growth for include: DAZN, Disney+, HBO, Hulu, NBCUniversal, Paramount+, Peacock, Sky, Sling TV, Univision and Warner Bros Discovery.
Privately held, Conviva is headquartered in Silicon Valley, California with offices and people around the globe. For more information, visit us at www.conviva.com. Join us to help extend our leadership position in big data streaming analytics to new audiences and markets!
● Follow due SDLC process for different phases of
development like design, development, unit testing,
maintenance
● Understand the requirements clearly from Functional
specs and future roadmap
● Write and update software design specs for the
solution
● Participate and invite peer review for design and
coding
● Support the automation requirements of continuous
integration and continuous deployments
● Handle customer escalations and special requests in a
timely manner
● Work with the rest of the team in a professional manner
and show ownership and commitment.
Qualifications and Required Skills:
● 2+ years experience in Software development
● In-depth knowledge of database technologies (SQL and
NoSQL)
● Proficiency in Python/Django framework
● Experience building RESTful APIs and knowledge of
RESTful design patterns
● Well versed with the complete Software Engineering Lifecycle.
● Strong Software development fundamentals including OOPS,
MVC, General Algorithms, Design Patterns etc.
● Good debugging and problem solving skills
● Analytical skills.
● Algorithmic and structured thinking.
● Works effectively within a collaborative team environment
● Good communication skills and positive attitude
Other Benefits:
● Insurance - Medical Insurance Coverage of 3,00,000 for
individual, spouse and kids. *Option of including parents and
higher coverage amount of 5,00,000
● Professional Development - Reimbursement towards relevant
courses that can help you grow your career and build your
knowledge level.
● Food - Lunch, Tea and Snacks. All you can eat!
● Work Culture- Flexible Result Oriented Work Environment
● Dress - We aren’t buttoned down. Wear anything you’re
comfortable in.Wear any comfortable dress you like to the office
– and be as comfortable at work
● Foosball - Whether you want to pick up the game or master
your skills, you’ll find plenty of people to play with
● TGIF - We’ll find something to do at the end of each week. Work
hard, party harder!
Roles & Responsibilities:
- Excellent problem-solving & analytical skills & OO principles
- Solid understanding of the data structure & algorithms.
- Expert knowledge with any one of the Javascript Framework such as Node Js.
- Any additional experience with Java or python, GoLang would be nice to have.
- Experience working with SQL database like MySQL or PostgreSQL
- Experience working with NoSQL databases like mongoDB, couchbase would be nice to have
- Working experience with cache like redis, mem cached or elasticache would be nice to have
- Working experience with Message Queues would be a great advantage.
- Experience with deployment & AWS would be nice to have
- Ability to do in depth analysis or problem statement or use case & Build robust & scalable architecture
- Logging & monitoring of the production system in real time
- Lead a team of 2-4 team members & make sure team practices defined coding practices across the project.
Company Introduction –
- Information Security & Data Analytics Series A funded company.
- Working in cutting edge technologies - Using AI for predictive intelligence and Facial Biometrics.
- Among Top 5 Cyber excellence companies globally (Holger Schulze awards)
- Bronze award for best startup of the year (Indian Express IT awards), only cyber Security Company in top 3.
- More than 100+ clients in India.
Job Description:-
Job Title: Python Developer
Key Requirements:-
- Mine data from structured and unstructured data sources.
- Extract data (text, images, and videos) from multiple documents in different formats.
- Extract information and intelligence from data.
- Extract data based on regular expressions.
- Collect data from structured RDBMS databases.
- Work closely with Project/Business/Research teams to provide mined data/intelligence for analysis.
- Should have strong exposure to core python skills like multiprocessing, multithreading, file handling, data structure like JSON, Data frames, and User Defined Data structure.
- Should have excellent knowledge of classes, file handling, memory manipulations.
- Strong Knowledge in Python.
- Strong exposure to frond end languages like CSS, JavaScript, Ajax etc.
- Should have exposure to requests, Frontera, scarpy-cluster, elastic-search, distributed computing tools like Kafka, Hbase, Redis, Zookeeper, restAPI.
- Should be familiar with *nix development environment.
- Knowledge of Django will be added advantage.
- Excellent knowledge on Web Crawling/Web scraping.
- Should have used scraping modules like Selenium, Scrapy, and Beautiful soup.
- Experience with text processing.
- Basics of databases. Good troubleshooting and debugging skills.
Experience : 1-4 Years Experiene
Education
B.Tech, MCA, Computer Engineering










