
autonomous world.Rich data in large volumes is getting collected at the edge (outside a datacenter) in use cases like autonomous vehicles, smart manufacturing, satellite imagery, smart retail, smart agriculture etc.These datasets are characterized by being unstructured
(images/videos), large size (Petabytes per month), distributed (across edge, on-prem and
cloud) and form the input for training AI models to get to higher degrees of automation.
Akridata is engaged with building products that solve these unique challenges and be at the forefront of this edge data revolution.
The company is backed by prominent VCs and has it’s entire software engineering team
based out of India and provides ample opportunities for from-scratch design and
development.
Role:
This role is an individual contributor role with key responsibilities in developing web server
backends for Akridata management plane software that provides a ‘single pane of glass’ for users to manage assets, specify and monitor large volume data pipelines at scale involving 10s of petabytes of data.
This role involves:
1. Working with tech leads and the rest of the team on the feature design activities and
picking appropriate tools and techniques for implementation.
2. Be a hands-on developer able to independently make correct implement choices, follow
sound development practices to ensure an enterprise grade application.
3. Guide and mentor junior team members.
What we are looking for:
1. A Bachelor’s or Master’s degree in computer science with strong CS fundamentals and
problem solving.
2. 5+ years of hands-on experience with software development with 3+ years on web
backend development.
3. A good understanding of backend application interactions with relational databases like
MySQL, Postgres etc
4. Knowledge of web server development frameworks preferably on Python.
5. Enthusiastic to work in a dynamic, fast paced startup environment.
Good to have:
1. Hands-on experience with designing database schema and implementing and debugging SQL queries for optimal performance for large datasets
2. Experience working with applications deployed on Kubernetes clusters.
3. Experience with working on a product from early stages of it’s development typically in a
startup environment.

Similar jobs
🚀 Hiring: Python Developer
⭐ Experience: 2+ Years
📍 Location: Mumbai (Andheri East )
⭐ Work Mode:- 5 Days Work From Office
⏱️ Notice Period: Immediate Joiners
(Only immediate joiners & candidates serving notice period)
Looking for a skilled Python Developer with experience in Django / FastAPI and MongoDB / PostgreSQL.
⭐ Must-Have Skills:-
✅ 2+ years of professional experience as a Python Developer
✅Proficient in Django or FastAPI
✅Hands-on with MongoDB or PostgreSQL
✅Strong understanding of REST APIs & Git
Job Title: Cloud Development & Linux Debugging Engineer
Experience: 5–10 Years
Location: Bangalore / Chennai
Job Summary
We are looking for an experienced Cloud Development & Linux Debugging Engineer with strong expertise in Linux internals, system-level programming, and cloud technologies. The ideal candidate will have hands-on experience in developing, debugging, and optimizing Linux-based systems along with exposure to DevOps tools and containerized environments.
Key Responsibilities
- Develop and debug software at the Linux system level (kernel/user space).
- Work on Linux internals, low-level system components, and performance optimization.
- Design, develop, and maintain applications using Python and C/C++.
- Troubleshoot complex issues in Linux and cloud-based environments.
- Collaborate with cross-functional teams in an Agile/Scrum environment.
- Contribute to automation and infrastructure using DevOps tools.
- Work with containerized and cloud platforms such as Kubernetes and OpenStack.
Required Skills
- Strong experience in Linux software development (Linux internals, system-level programming).
- Proficiency in Python and C/C++.
- Solid debugging and analytical skills.
- Hands-on experience with Ansible, Puppet, and DevOps practices.
- Experience working with OpenStack and Kubernetes.
- Good understanding of Agile/Scrum methodologies.
- Excellent communication and teamwork skills.
Preferred Skills (Good to Have)
- Experience with Go / Golang and Go templating.
- Knowledge of Kubernetes Operators and Helm.
- Exposure to containerization technologies (Docker, Kubernetes).
- Contributions to open-source projects.
- Experience with cloud-native architectures.
Qualifications
- Bachelor’s/Master’s degree in Computer Science, Engineering, or related field.
- Self-driven individual with a strong learning mindset.
- Ability to work independently and in collaborative team environments.
We are seeking a skilled Java Developer with hands-on experience in Java and Spark to build scalable data processing solutions. You'll contribute to high-performance data pipelines and analytics platforms in a dynamic Agile environment.
Key Responsibilities
- Design and develop Java applications integrated with Apache Spark for ETL processes, data transformations, and analytics.
- Build and optimize Spark jobs (Spark SQL, DataFrames, Streaming) for large-scale data processing.
- Collaborate with data engineers and analysts to implement robust data workflows.
- Write clean, maintainable Java code following best practices (Spring Boot, Microservices preferred).
- Perform code reviews, unit testing, and contribute to CI/CD pipelines.
- Troubleshoot and optimize Spark performance for production workloads.
- Document technical solutions and mentor junior developers.
Required Skills & Qualifications
- 4-7 years of hands-on Java development experience.
- Strong expertise in Apache Spark (Spark Core, Spark SQL, PySpark basics).
- Proficiency in Java 8/11+ with multithreading and collections frameworks.
- Experience with data processing (ETL, data pipelines, big data).
- Familiarity with build tools (Maven/Gradle) and version control (Git).
- Strong problem-solving skills and Bangalore location availability.
- Excellent communication skills for cross-team collaboration.
Good to Have
- Experience with Snowflake for cloud data warehousing.
- Knowledge of DBT (Data Build Tool) for analytics engineering.
- Python scripting for data manipulation and automation.
- Exposure to AWS/GCP/Azure cloud platforms.
- Familiarity with Kafka, Airflow, or containerization (Docker/Kubernetes).
Genspark is hiring Professionals for C Development for there Premium Client
Work Location- Chennai
Entry Criteria
Graduate from Any Engineering Background /BSc/MSc /MCA with specialization(Computer/Electronics/IT )
Minimum 1 year experience in Industry
Working Knowledge of C/Embedded/C++/DSA
Programming Aptitude (Any Language)
Basic understanding of programming constructs: variables, loops, conditionals, functions
Logical thinking and algorithmic approach
Computer Science Fundamentals:
Data structures basics: arrays, stacks, queues, linked lists
Operating System basics: what is a process/thread, memory, file system, etc.
Basic understanding of compilation, runtime, networking and sockets etc.
Problem Solving & Logical Reasoning
Ability to trace logic, find errors, and reason through pseudocode
Analytical and debugging capabilities
Learning Attitude & Communication
Demonstrated interest in low-level or systems programming (even if no experience)
Willingness to learn C and work close to the OS level
Clarity of thought and ability to explain what they do know
Soft Skills :
Able to explain and communicate the thoughts clearly in English
Confident in solving new problems independently or with guidance
Willingness to take feedback and iterate
Evaluation Process
Candidates will be assigned an online test, followed by Technical Screening.
Shortlisted Candidates will have to appear for a F2F Interview with the Client, Chennai.









