The ideal candidate is a self-motivated, multi-tasker, and demonstrated team-player. You will be a senior developer responsible for the development of new software products and enhancements to existing products. You should excel in working with large-scale applications and frameworks and have outstanding communication and leadership skills.
Responsibilities:
- Writing clean, high-quality, high-performance, maintainable code
- Develop and support software including applications, database integration, interfaces, and new functionality enhancements
- Coordinate cross-functionally to insure project meets business objectives and compliance standards
- Support test and deployment of new products and features
- Participate in code reviews
Requirements:
- B.Tech from Tier 1 College.
- 3+ yrs of experience, at least 2 yrs of experience with Python and Django.
- Expertise in Object Oriented Design, Database Design, and XML Schema
- Experience with Agile or Scrum software development methodologies
- Ability to multi-task, organise, and prioritise work

About Spinny
About
Connect with the team
Similar jobs
Job Title : Python Django Developer
Experience : 6+ Years
Location : Gurgaon (Work from Office)
Job Type : Full-time
Working Days: Monday to Friday (5 Days)
Timings: 9:30 AM – 6:30 PM
Job Summary :
We are seeking a highly skilled and experienced Python Django Developer to join our dynamic team in Gurgaon.
The ideal candidate will have a strong background in backend development, Django frameworks, and RESTful API integration.
You will be responsible for building and maintaining scalable web applications and collaborating with cross-functional teams.
Key Responsibilities :
- Develop, test, and maintain robust, scalable, high-performance web applications using Django and Python.
- Design and implement RESTful APIs and integrate third-party APIs and services.
- Write reusable, testable, and efficient code following best practices and coding standards.
- Work with frontend developers to integrate user-facing elements using server-side logic.
- Optimize applications for maximum speed and scalability.
- Perform code reviews, troubleshoot issues, and ensure application performance and security.
- Collaborate with product managers, designers, and other developers to deliver high-quality products.
Required Skills & Qualifications :
- 6+ Years of hands-on experience with Python and Django framework.
- Strong experience with RESTful APIs, Django ORM, and PostgreSQL/MySQL.
- Proficiency in version control systems like Git.
- Familiarity with frontend technologies such as HTML, CSS, JavaScript, and AJAX.
- Experience with Docker, Celery, Redis, and cloud platforms (AWS/Azure) is a plus.
- Solid understanding of software development principles and design patterns.
- Excellent problem-solving skills and attention to detail.
- Strong communication and collaboration abilities.
Good to Have :
- Experience with Django REST Framework (DRF).
- Knowledge of unit testing and CI/CD pipelines.
- Exposure to Agile development methodologies.
- Familiarity with containerization and deployment tools (Docker, Kubernetes).
Python. Django Rest Framework experience would be great, but not essential! We prioritise
quality over quantity.
Postgres with InfluxDB, Oracle, MySQL and Redis experience a plus.
Docker, Kubernetes, Helm, OpenShift and associated tooling.
AWS, especially ECS, Lambda, RDS and DynamoDB. Performance and uptime are super
important to us.
The challenge of scaling a global, distributed API to 10,000+ requests per second.
We have SDKs in a bunch of languages, so the more polyglot you are the better.
If you like writing JS and React that would be awesome too.
Job Description :
We are looking for a Backend Developer to join our Software Development team in Gurgaon.
At Zunpulse, you can expect to design flexible and scalable solutions and work on complex
challenges by utilizing skills in data structures, algorithms, and object-oriented programming. You
will be part of the team that develops state-of-the-art products which aim at helping millions of
customers.
Job Description
- Must have worked on Spring, Spring boot framework
- Strong Java developer skills
- Should have strong knowledge of Data Structures and Algorithms
- Should be well versed with Agile Development methodologies
- Prior experience in handling backend
- Should have strong debugging and communication skill
- Microservices experience would be a plus
- Built backend for Android, iOS applications in past
- Well versed with AWS deployments
- Excellent problem-solving skills
- Experience in SQL or Mongo Database
- Perform UI tests to optimize performance
- Write clean reusable code
Essential Qualification
• Bachelor’s degree in Computer Science/ Electronics or equivalent.
• Graduated in 2020/21 having 1-2 years of industrial experience
Skill-set
Spring boot, Java, Flask, Python, Data Structures, Algorithms, Basic knowledge of AWS/GCP
What we offer
● Growth-oriented culture
● Limitless learning opportunities by working with cutting edge tech stacks and with a
diverse and talented team
● Excellent compensation with competitive benefits
● Flexible work environment
● A chance to make an impact on our business strategy using your work
About zunpulse:
Zunpulse is India's leading smart electricity brand. Backed by Godrej, we are a 150+ member
engineering team led by co-founders from IIT Kharagpur, IIT Delhi & IIT Kanpur. We have developed
the largest range of IoT and AI-enabled smart home products with one single offering in mind
#phonesekarocontrol. Founded in June 2016, we have impacted 1,00,000 + homes to date.
Customers are looking for comfort, convenience, and security, all of which zunpulse provides with a
single app, to provide it all with a hyper-personalized experience. With the zunpulse app, the team is
creating a new way for the customers to manage their lives while keeping them cool, fun, and trendy.
With an all-around IoT tech solution, zunpulse is set to create a future smart home suite that can
transform how you live and interact with your home.
To know more about us refer:
https://www.zunpulse.com/
About the role
Checking quality is one of the most important tasks at Anakin. Our clients are pricing their products based on our data, and minor errors on our end can lead to our client's losses of millions of dollars. You would work with multiple tools and with people across various departments to ensure the accuracy of the data being crawled. You would setup manual and automated processes and make sure they run to ensure the highest possible data quality.
You are the engineer other engineers can count on. You embrace every problem with enthusiasm. You remove hurdles, are a self-starter, team player. You have the hunger to venture into unknown areas and make the system work.
Your Responsibilities would be to:
- Understand customer web scraping and data requirements; translate these into test approaches that include exploratory manual/visual testing and any additional automated tests deemed appropriate
- Take ownership of the end-to-end QA process in newly-started projects
- Draw conclusions about data quality by producing basic descriptive statistics, summaries, and visualisations
- Proactively suggest and take ownership of improvements to QA processes and methodologies by employing other technologies and tools, including but not limited to: browser add-ons, Excel add-ons, UI-based test automation tools etc.
- Ensure that project requirements are testable; work with project managers and/or clients to clarify ambiguities before QA begins
- Drive innovation and advanced validation and analytics techniques to ensure data quality for Anakin's customers
- Optimize data quality codebases and systems to monitor the Anakin family of app crawlers
- Configure and optimize the automated and manual testing and deployment systems used to check the quality of billions of data points of over 1000+ crawlers across the company
- Analyze data and bugs that require in-depth investigations
- Interface directly with external customers including managing relationships and steering requirements
Basic Qualifications:
- 2+ years of experience as a backend or a full-stack software engineer
- Web scraping experience with Python or Node.js
- 2+ years of experience with AWS services such as EC2, S3, Lambda, etc.
- Should have managed a team of software engineers
- Should be paranoid about data quality
Preferred Skills and Experience:
- Deep experience with network debugging across all OSI layers (Wireshark)
- Knowledge of networks or/and cybersecurity
- Broad understanding of the landscape of software engineering design patterns and principles
- Ability to work quickly and accurately in a highly stressful environment during removing bugs in run-time within minutes
- Excellent communicator, both written and verbal
Additional Requirements:
- Must be available to work extended hours and weekends when needed to meet critical deadlines
- Must have an aversion to politics and BS. Should let his/her work speak for him/her.
- Must be comfortable with uncertainty. In almost all the cases, your job will be to figure it out.
- Must not be bounded to comfort zone. Often, you will need to challenge yourself to go above and beyond.
- Experience in HTML, CSS and Bootstrap library is required.
- Experience in Python Programming. Experience in Pandas and Numpy Libraries.
- Experience with Django Framework & REST API.
- Experience Flask, Microservices
- Experience in FAST API would be added advantage.
- Exposure in converting ML models to Microservices and exposing them as API
- GRPC development
- Experience with Django/Flask Framworks
- Worked on web app and mobile app
- Understanding of DevOps tools - Jekins/Kubernetes/Azure/AWS/Docker will be a plus
Hi All,
We are hiring!!
Company: SpringML India Pvt Ltd.
Role:Lead Data Engineer
Location: Hyderabad
Website: https://springml.com/">https://springml.com/
About Company:
At SpringML, we are all about empowering the 'doers' in companies to make smarter decisions with their data. Our predictive analytics products and solutions apply machine learning to today's most pressing business problems so customers get insights they can trust to drive business growth.
We are a tight-knit, friendly team of passionate and driven people who are dedicated to learning, get excited to solve tough problems and like seeing results, fast. Our core values include placing our customers first, empathy and transparency, and innovation. We are a team with a focus on individual responsibility, rapid personal growth, and execution. If you share similar traits, we want you on our team.
What's the opportunity?
SpringML is looking to hire a top-notch Lead Data Engineer who is passionate about working with data and using the latest distributed framework to process large dataset.
As a Lead Data Engineer, your primary role will be to design and build data pipelines. You will be focused on helping client projects on data integration, data prep and implementing machine learning on datasets.
In this role, you will work on some of the latest technologies, collaborate with partners on early win, consultative approach with clients, interact daily with executive leadership, and help build a great company. Chosen team members will be part of the core team and play a critical role in scaling up our emerging practice.
Responsibilities:
- Ability to work as a member of a team assigned to design and implement data integration solutions.
- Build Data pipelines using standard frameworks in Hadoop, Apache Beam and other open-source solutions.
- Learn quickly – ability to understand and rapidly comprehend new areas – functional and technical – and apply detailed and critical thinking to customer solutions.
- Propose design solutions and recommend best practices for large scale data analysis
Skills:
- B.tech degree in computer science, mathematics or other relevant fields.
- 6+years of experience in ETL, Data Warehouse, Visualization and building data pipelines.
- Strong Programming skills – experience and expertise in one of the following: Java, Python, Scala, C.
- Proficient in big data/distributed computing frameworks such as Apache Spark, Kafka,
- Experience with Agile implementation methodology
Company Introduction (apna)
We are building a novel platform to create better economic opportunities for the blue/grey
workforce around the world. We are a very product focused and customer obsessed company. Delivering high quality products at lightning speed is our USP. If you are looking for challenging work and rapid merit based growth in your career, look no further and welcome aboard.
Checkout our product on Google Play store apna - Job Search App | Job Groups
Job Description
We are looking for a Senior Back-End Engineer responsible looking for engineering ownership of one/many products at apna. Your primary focus will be development of all server-side logic, definition and maintenance of the central database, and ensuring high performance and responsiveness to requests from the front-end (Android + WebApp). A basic understanding of front-end technologies is therefore necessary as well.
Responsibilities
● Take ownership of end-to-end product development
● Be champion of Test Driven Development methodology
● Building reusable code and libraries for future use
● Optimization of the application for maximum speed and scalability
● Implementation of security and data protection
● Technical Documentation around the owned product
Skills And Qualifications
● Expert level in OOPS concepts and REST API development
● Proficiency in one or more of the following languages: Python, Java, Ruby, Go
● Knowledge around ElasticSearch, Celery, Redis is must
● Integration of multiple data sources and databases into one system
● Implementing automated testing platforms and unit tests
● Proficient understanding of code versioning tools, such as Git
● Knowledge of GCP would be a plus
Perks
● Have autonomy to work from home. Only one rule: Get things Done!
● We promote and celebrate team members contributing at a variety of professional
conferences and events. Be a speaker, contribute to open sourcepublish academic
work, publish an online blog showing your expertise. we are here to encourage you!
● You join us in your current area of expertise and let us know about your personal
professional goals. Looking to start your company? Further studies at the best institutes?
Want to change career track? We make sure you reach there!
● Generous pool of ESOPs
Be Part Of Building The Future
Dremio is the Data Lake Engine company. Our mission is to reshape the world of analytics to deliver on the promise of data with a fundamentally new architecture, purpose-built for the exploding trend towards cloud data lake storage such as AWS S3 and Microsoft ADLS. We dramatically reduce and even eliminate the need for the complex and expensive workarounds that have been in use for decades, such as data warehouses (whether on-premise or cloud-native), structural data prep, ETL, cubes, and extracts. We do this by enabling lightning-fast queries directly against data lake storage, combined with full self-service for data users and full governance and control for IT. The results for enterprises are extremely compelling: 100X faster time to insight; 10X greater efficiency; zero data copies; and game-changing simplicity. And equally compelling is the market opportunity for Dremio, as we are well on our way to disrupting a $25BN+ market.
About the Role
The Dremio India team owns the DataLake Engine along with Cloud Infrastructure and services that power it. With focus on next generation data analytics supporting modern table formats like Iceberg, Deltalake, and open source initiatives such as Apache Arrow, Project Nessie and hybrid-cloud infrastructure, this team provides various opportunities to learn, deliver, and grow in career. We are looking for innovative minds with experience in leading and building high quality distributed systems at massive scale and solving complex problems.
Responsibilities & ownership
- Lead, build, deliver and ensure customer success of next-generation features related to scalability, reliability, robustness, usability, security, and performance of the product.
- Work on distributed systems for data processing with efficient protocols and communication, locking and consensus, schedulers, resource management, low latency access to distributed storage, auto scaling, and self healing.
- Understand and reason about concurrency and parallelization to deliver scalability and performance in a multithreaded and distributed environment.
- Lead the team to solve complex and unknown problems
- Solve technical problems and customer issues with technical expertise
- Design and deliver architectures that run optimally on public clouds like GCP, AWS, and Azure
- Mentor other team members for high quality and design
- Collaborate with Product Management to deliver on customer requirements and innovation
- Collaborate with Support and field teams to ensure that customers are successful with Dremio
Requirements
- B.S./M.S/Equivalent in Computer Science or a related technical field or equivalent experience
- Fluency in Java/C++ with 8+ years of experience developing production-level software
- Strong foundation in data structures, algorithms, multi-threaded and asynchronous programming models, and their use in developing distributed and scalable systems
- 5+ years experience in developing complex and scalable distributed systems and delivering, deploying, and managing microservices successfully
- Hands-on experience in query processing or optimization, distributed systems, concurrency control, data replication, code generation, networking, and storage systems
- Passion for quality, zero downtime upgrades, availability, resiliency, and uptime of the platform
- Passion for learning and delivering using latest technologies
- Ability to solve ambiguous, unexplored, and cross-team problems effectively
- Hands on experience of working projects on AWS, Azure, and Google Cloud Platform
- Experience with containers and Kubernetes for orchestration and container management in private and public clouds (AWS, Azure, and Google Cloud)
- Understanding of distributed file systems such as S3, ADLS, or HDFS
- Excellent communication skills and affinity for collaboration and teamwork
- Ability to work individually and collaboratively with other team members
- Ability to scope and plan solution for big problems and mentors others on the same
- Interested and motivated to be part of a fast-moving startup with a fun and accomplished team










