Role and Responsibilities Build a low latency serving layer that powers DataWeave's Dashboards, Reports, and Analytics functionality Build robust RESTful APIs that serve data and insights to DataWeave and other products Design user interaction workflows on our products and integrating them with data APIs Help stabilize and scale our existing systems. Help design the next generation systems. Scale our back end data and analytics pipeline to handle increasingly large amounts of data. Work closely with the Head of Products and UX designers to understand the product vision and design philosophy Lead/be a part of all major tech decisions. Bring in best practices. Mentor younger team members and interns. Constantly think scale, think automation. Measure everything. Optimize proactively. Be a tech thought leader. Add passion and vibrance to the team. Push the envelope. Skills and Requirements 8- 15 years of experience building and scaling APIs and web applications. Experience building and managing large scale data/analytics systems. Have a strong grasp of CS fundamentals and excellent problem solving abilities. Have a good understanding of software design principles and architectural best practices. Be passionate about writing code and have experience coding in multiple languages, including at least one scripting language, preferably Python. Be able to argue convincingly why feature X of language Y rocks/sucks, or why a certain design decision is right/wrong, and so on. Be a self-starter—someone who thrives in fast paced environments with minimal ‘management’. Have experience working with multiple storage and indexing technologies such as MySQL, Redis, MongoDB, Cassandra, Elastic. Good knowledge (including internals) of messaging systems such as Kafka and RabbitMQ. Use the command line like a pro. Be proficient in Git and other essential software development tools. Working knowledge of large-scale computational models such as MapReduce and Spark is a bonus. Exposure to one or more centralized logging, monitoring, and instrumentation tools, such as Kibana, Graylog, StatsD, Datadog etc. Working knowledge of building websites and apps. Good understanding of integration complexities and dependencies. Working knowledge linux server administration as well as the AWS ecosystem is desirable. It's a huge bonus if you have some personal projects (including open source contributions) that you work on during your spare time. Show off some of your projects you have hosted on GitHub.
Who are we and Why Yulu for you?Yulu's vision is to reduce traffic congestion by providing a scalable, affordable, efficient and clean solution for the first mile, last mile and short distance commute. We strongly believe that short commute mode can be made more Efficient, Affordable and GREEN, thereby ensuring that future generations will have a healthy environment to live. We want to replace big personal vehicles with a smaller form factor that does not require fossil fuel. Yulu uses IoT technology to create a vast network of shared bikes that can be rented easily by user-friendly app in pay per use business model.We are a mission driven company based out of Bangalore with a solid founding team and big ambitions. At Yulu we believe that every day is a mission and success is exciting when its attained with fun and as a team. If you want to solve the problem of traffic congestion, reduce pollution, positively impact the society and be part of an international dynamic team, join us in our journey to shape better India for our future generations.Responsibility Working with Big Data tools and frameworks to provide requested capabilities Identify development needs in order to improve and streamline operations Develop and manage BI solutions Implementing ETL process and Data Warehousing Monitoring performance and managing infrastructure Skills required 2+ years of experience in relevent field Proficient understanding of distributed computing principles Proficiency with Hadoop and Spark Experience with building stream-processing systems, using solutions such as Kafka and Spark-Streaming Good knowledge of Data querying tools SQL and Hive Experience with integration of data from multiple data sources Knowledge of various ETL techniques and frameworks Experience with Python/Java/Scala (at least one) Experience with cloud services such as AWS or GCP Experience with NoSQL databases, such as DynamoDB,MongoDB will be an advantage