autonomous world.Rich data in large volumes is getting collected at the edge (outside a datacenter) in use cases like autonomous vehicles, smart manufacturing, satellite imagery, smart retail, smart agriculture etc.These datasets are characterized by being unstructured
(images/videos), large size (Petabytes per month), distributed (across edge, on-prem and
cloud) and form the input for training AI models to get to higher degrees of automation.
Akridata is engaged with building products that solve these unique challenges and be at the forefront of this edge data revolution.
The company is backed by prominent VCs and has it’s entire software engineering team
based out of India and provides ample opportunities for from-scratch design and
development.
Role:
This role is an individual contributor role with key responsibilities in developing web server
backends for Akridata management plane software that provides a ‘single pane of glass’ for users to manage assets, specify and monitor large volume data pipelines at scale involving 10s of petabytes of data.
This role involves:
1. Working with tech leads and the rest of the team on the feature design activities and
picking appropriate tools and techniques for implementation.
2. Be a hands-on developer able to independently make correct implement choices, follow
sound development practices to ensure an enterprise grade application.
3. Guide and mentor junior team members.
What we are looking for:
1. A Bachelor’s or Master’s degree in computer science with strong CS fundamentals and
problem solving.
2. 5+ years of hands-on experience with software development with 3+ years on web
backend development.
3. A good understanding of backend application interactions with relational databases like
MySQL, Postgres etc
4. Knowledge of web server development frameworks preferably on Python.
5. Enthusiastic to work in a dynamic, fast paced startup environment.
Good to have:
1. Hands-on experience with designing database schema and implementing and debugging SQL queries for optimal performance for large datasets
2. Experience working with applications deployed on Kubernetes clusters.
3. Experience with working on a product from early stages of it’s development typically in a
startup environment.
Similar jobs
Key Responsibilities:
● Work closely with product managers, designers, frontend developers, and other
cross-functional teams to ensure the seamless integration and alignment of frontend and
backend technologies, driving cohesive and high-quality product delivery.
● Develop and implement coding standards and best practices for the backend team.
● Document technical specifications and procedures.
● Stay up-to-date with the latest backend technologies, trends, and best practices.
● Collaborate with other departments to identify and address backend-related issues.
● Conduct code reviews and ensure code quality and consistency across the backend team.
● Create technical documentation, ensuring clarity for future development and
maintenance.
Requirements;
● Experience: 4-6 years of hands-on experience in backend development, with a strong
background in product-based companies or startups.
● Education: Bachelor’s degree or above in Computer Science or a related field.
● Programming skills: Proficient in Python and software development principles, with a
focus on clean, maintainable code, and industry best practices. Experienced in unit
testing, AI-driven code reviews, version control with Git, CI/CD pipelines using GitHub
Actions, and integrating New Relic for logging and APM into backend systems.
● Database Development: Proficiency in developing and optimizing backend systems in
both relational and non-relational database environments, such as MySQL and NoSQL
databases.
● GraphQL: Proven experience in developing and managing robust GraphQL APIs,
preferably using Apollo Server. Ability to design type-safe GraphQL schemas and
resolvers, ensuring seamless integration and high performance.
● Cloud Platforms: Familiar with AWS and experienced in Docker containerization and
orchestrating containerized systems.
● System Architecture: Proficient in system design and architecture with experience in
developing multi-tenant platforms, including security implementation, user onboarding,
payment integration, and scalable architecture.
● Linux Systems: Familiarity with Linux systems is mandatory, including deployment and
management.
● Continuous Learning: Stay current with industry trends and emerging technologies to
influence architectural decisions and drive continuous improvement.
Benefits:
● Competitive salary.
● Health insurance.
● Casual dress code.
● Dynamic & Collaboration friendly office.
● Hybrid work schedule.
Industry
- IT Services and IT Consulting
Employment Type
Full-time
The Sr. Analytics Engineer would provide technical expertise in needs identification, data modeling, data movement, and transformation mapping (source to target), automation and testing strategies, translating business needs into technical solutions with adherence to established data guidelines and approaches from a business unit or project perspective.
Understands and leverages best-fit technologies (e.g., traditional star schema structures, cloud, Hadoop, NoSQL, etc.) and approaches to address business and environmental challenges.
Provides data understanding and coordinates data-related activities with other data management groups such as master data management, data governance, and metadata management.
Actively participates with other consultants in problem-solving and approach development.
Responsibilities :
Provide a consultative approach with business users, asking questions to understand the business need and deriving the data flow, conceptual, logical, and physical data models based on those needs.
Perform data analysis to validate data models and to confirm the ability to meet business needs.
Assist with and support setting the data architecture direction, ensuring data architecture deliverables are developed, ensuring compliance to standards and guidelines, implementing the data architecture, and supporting technical developers at a project or business unit level.
Coordinate and consult with the Data Architect, project manager, client business staff, client technical staff and project developers in data architecture best practices and anything else that is data related at the project or business unit levels.
Work closely with Business Analysts and Solution Architects to design the data model satisfying the business needs and adhering to Enterprise Architecture.
Coordinate with Data Architects, Program Managers and participate in recurring meetings.
Help and mentor team members to understand the data model and subject areas.
Ensure that the team adheres to best practices and guidelines.
Requirements :
- Strong working knowledge of at least 3 years of Spark, Java/Scala/Pyspark, Kafka, Git, Unix / Linux, and ETL pipeline designing.
- Experience with Spark optimization/tuning/resource allocations
- Excellent understanding of IN memory distributed computing frameworks like Spark and its parameter tuning, writing optimized workflow sequences.
- Experience of relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., Redshift, Bigquery, Cassandra, etc).
- Familiarity with Docker, Kubernetes, Azure Data Lake/Blob storage, AWS S3, Google Cloud storage, etc.
- Have a deep understanding of the various stacks and components of the Big Data ecosystem.
- Hands-on experience with Python is a huge plus
Able to build API'S
Should have production knowledge
Have deployed solutions in production.
Responsibilities:
- Develop cloud-based Python Django software products
- Working closely with UX and Front-end Developers
- Participating in architectural, design, and product discussions
- Designing and creating RESTful APIs for internal and partner consumption
- Working in an agile environment with an excellent team of engineers
Skills:
- Highly proficient with Python
- Hands-on experience with the Django framework
- Solid database skills in a relational database (i.e. PostgresSQL, MySQL, etc.)
- Strongly preferred exposure with Elasticsearch
- Knowledge of how to build and use RESTful APIs
- Strong knowledge of version control (i.e. git, svn, etc.)
- Experience deploying Python applications into production
- Amazon Web Services (AWS) infrastructure knowledge is a plus
Position description:
- Architecture & Design systems for Predictive analysis and writing algorithms to deal with financial data
- Must have experience on web services and APIs (REST, JSON, and similar) and creation and consumption of RESTful APIs
- Proficiently writing algorithms with Python/Pandas/Numpy; Jupyter/PyCharm
- Experience with relational and NoSQL databases (Eg. MSSQL, MongoDB, Redshift, PostgreSQL, Redis)
- Implementing Machine Learning Models using Python/R for best performance
- Working with Time Series Data & analyzing large data sets.
- Implementing financial strategies in python and generating reports to analyze the strategy results.
Primary Responsibilities:
- Writing algorithms to deal with financial data and Implementing financial strategies in (Python, SQL) and generating reports to analyze the strategy results.
Educational qualifications preferred Degree: Bachelors degree
Required Knowledge:
- Highly skilled in SQL, Python, Pandas, Numpy, Machine Learning, Predictive Modelling, Algorithm designing, OOPS concepts
- 2 - 7 years Full-Time working experience on core SQL, Python role (Non-Support)
- Bachelor’s Degree in Engineering, equivalent or higher education.
- Writing algorithms to deal with financial data and Implementing financial strategies in (Python, SQL) and generating reports to analyze the strategy results.
Who We are
At e6data (https://e6x.io/">https://e6data.io) we’re building core algorithms that makes querying 25 - 200X more performant (i.e. faster and/or cheaper) than every other analytics platform in the world. This results in billions of dollars of value unlocked across faster analytics, savings on compute / infra, and new use cases becoming possible.
Our core team has 100+ years of combined experience and 10+ patents across diverse but complementary disciplines like real time databases, time series databases, SQL OLAP engines, distributed computing platforms, stream processing, and log analytics. Two of the founding team are second time founders with past exits.
We are passionate about solving the deepest and most important problems in computer science, and believe in making our technology available to customers regardless of their location, size, or budget. Our vision is for e6data to power every analytical workload in the world - regardless of deployment scenario (cloud, on-premise, edge), across every use case (end user queries, ETL / data pipelines), and every latency requirement (batch, streaming / real time).
What you'll do?
- Write awesome code
- The focus will be on performance of our data platform
- Identify tactical and long-term product improvements (could be code, script, etc.)
- Responsible for Planning and executing Metadata and Governance tool supports
- Build our internal Data Platform using exdata
- Adding scalable functionality to the system for new features like UDF's, sql functions etc
What are we looking for?
- Worked with Python for a minimum of 2 years
- Well versed with Django REST framework
- Knowledge of document based data stores like MongoDB
- Strong team player
- A passion for continuous learning and growth (both at a personal and team level)
- Experience (or enjoys) working in a startup will be an added advantage
Strong expertise in AWS architecture, deployment and coding using Python and AWS Lambda functions. She/ he should also have experience with NOSQL database like MongoDB/ DynamoDB and exposure to Machine Learning. She/ he should be familiar with Ajax calls, interaction with API using JSON files and xml.
Knowledge of testing tools like JMeter and Selenium and programming languages like Javascript and Dot Net will be a big advantage.
The person should be able to guide junior resources, perform code reviews and solve performance issues as required.
Roles & Responsibilities -
This position is a hands-on Python / SQL software developer role, candidate needs to have exposure to electronic trading business, proficient in python and experience building systems for data processing. The candidate will join the front office development team
- Build and maintain infrastructure for data retrieval, processing and storage
- Build strong working relationships with international teams
- Be willing and able to adapt to changes in priorities
- Ability to learn and apply new technologies to deliver added business value
- Maintain a strong focus on quality
Skill Sets & Prerequisites -
- Proficient in Python / SQL
- Good understanding of data base management systems
- Experience developing and maintaining systems that handle large amounts of data
- Understanding of Electronic Trading Systems
- Attention to detail and code quality
- Excellent problem solving and analytical skills in a high-pressure environment
- Strong communication skills and an ability to convey ideas and concepts with clarity