TetherBox Technologies | Empowering global water businesses with an Internet of Things toolset to help reduce inefficiencies
We are a start-up in India seeking excellence in everything we do with an unwavering curiosity and enthusiasm. We build simplified new-age AI driven Big Data Analytics platform for Global Enterprises and solve their biggest business challenges. Our Engineers develop fresh intuitive solutions keeping the user in the center of everything. As a Cloud-ML Engineer, you will design and implement ML solutions for customer use cases and problem solve complex technical customer challenges. Expectations and Tasks - Total of 7+ years of experience with minimum of 2 years in Hadoop technologies like HDFS, Hive, MapReduce - Experience working with recommendation engines, data pipelines, or distributed machine learning and experience with data analytics and data visualization techniques and software. - Experience with core Data Science techniques such as regression, classification or clustering, and experience with deep learning frameworks - Experience in NLP, R and Python - Experience in performance tuning and optimization techniques to process big data from heterogeneous sources. - Ability to communicate clearly and concisely across technology and the business teams. - Excellent Problem solving and Technical troubleshooting skills. - Ability to handle multiple projects and prioritize tasks in a rapidly changing environment. Technical Skills Core Java, Multithreading, Collections, OOPS, Python, R, Apache Spark, MapReduce, Hive, HDFS, Hadoop, MongoDB, Scala We are a retained Search Firm employed by our client - Technology Start-up @ Bangalore. Interested candidates can share their resumes with me - Jia@TalentSculpt.com. I will respond to you within 24 hours. Online assessments and pre-employment screening are part of the selection process.
Should take responsibility for the planning and execution of technical projects Strong grasp of CS fundamentals and excellent problem solving abilities to ensure that systems are high in quality with maintainable and reusable code Experience with hands on coding and independent ownership of software components Deep understanding of software engineering practices, Object Oriented Analysis & Design, Design Patterns, Data Structure, Algorithms Ensure best engineering practices in performance, scalability, enterprise system
Experience - 2-5 years Location- Bangalore About us: Zilingo is an online marketplace established in 2015 with a focus in fashion and lifestyle. Founded by former Mckinsey consultant Ankiti Bose, and ex-Yahoo engineer Dhruv Kapoor, Zilingo hosts over 4,000 SMEs across Indonesia, Singapore, Thailand, China, Hong Kong, Vietnam and Taiwan with a reach of more than a million customers per month. Zilingo is built on the core principle of creating value in the long-tail fashion and lifestyle market which is enormous - yet largely unorganised. With Zilingo, you can discover fashion from Chatuchak, Haji Lane and Pasar Baru and find local designers and indie labels, all in one place. Zilingo is focussed on bringing together long-tail vendors that sell offline within markets and malls across Asia and provide them with a platform to grow their business. Passionate about building a culture of micro entrepreneurship, the company offer sellers unprecedented access to buyers, free-of-cost access to analytics, speedy logistics and secure payments. Zilingo currently has offices in 5 countries (Thailand, Singapore, Indonesia, Hong Kong & India). The young dynamic team, led by co-founders Ankiti Bose and Dhruv Kapoor, is looking for driven self-motivated individuals who are excited to build the next big thing in mobile shopping. We are looking for awesome Backend Developers who can add value to the organisation. Job Description: Zilingo is an exciting, Sequoia-backed marketplace connecting sellers and consumers in South-East Asia to each other, across national, language, and currency boundaries. Zilingo's wholly in-house platform is key to enabling small sellers to sell unique products to consumers across the region. Zilingo's backend platform is a network of microservices written in Scala on top of the Play framework and Akka. As a backend developer at Zilingo, you'll get to develop tons of our core libraries which help us develop better-scaling services faster, work on difficult problems, including recommendations, search, and fulfillment, and DevOps. We have a passionate engineering culture that encourages solving difficult problems and engaging closely with unique challenges to add value to hundreds of thousands of customers and sellers. We encourage ownership and innovation and love to work as a team.
Work on backend of a Saas platform and other cool things.
Equal Experts makes simple solutions to big business problems. We provide tailored, end-to-end services in software development and delivery – from user research and design, to technical architecture, development and QA, all the way to devops, continuous delivery, hosting and support. With offices in the UK, US, Portugal, India and Canada, our network of over 700 experienced software consultants – a blend of permanent employees and associates – has created software for a wide range of public and private sector clients. These include organisations as diverse as HMRC, the Home Office, O2, Camelot and major institutions in the publishing, financial and retail sectors. Continuing growth saw our total sales reaching £42 million in 2015/16. Everyone at Equal Experts is committed to using technology and modern agile practices to deliver measurable business value. Our people typically have at least 10 years’ experience in delivering valuable, working software, and this focus on experience sets us apart – it’s what allows us to develop high-quality software faster, and for lower cost. Recent updateSee all See all recent updates Image accompanying recent update 10 years together: Associate model - Equal Experts equalexperts.com 11 Likes 5d Company details Website http://www.equalexperts.com Headquarters London Year founded 2007 Company type Privately Held Company size 501-1,000 employees Specialties Application Development, Offshore & Distributed delivery, System Integration, Iterative Delivery, Digital Transformation, User Centred Design, Continuous Delivery, E-Commerce, DevOps, Big Data, Mobile (Web and Native), Open Source, Java, Scala, NoSQL, .Net, REST, and Cloud
At Equal Experts we are a network of talented experienced software consultants specialising in Agile Delivery. So, what do we do in our regular day at EE? We indulge in all things that would excite you! Like; ● Work on large-scale, custom distributed software systems using Java, Scala, C#/.NET, MongoDB, Neo4j, Groovy, Angular JS, ReactJS, .Net, Cucumber and the likes ● Be responsible for the quality of software and resolving any issues regarding client satisfaction ● Employ Agile development including task estimation, test automation, deployment automation and Continuous Integration to improve overall execution speed and product quality ● Work in a dynamic, collaborative, transparent, non-hierarchal, and ego-free culture where your talent is valued over a role title ● Spread the word about best practices in software development inside and outside Equal Experts community ● Speak at conferences like Experts Talk and others ● Learn something new everyday, write blogs ● We work almost exclusively on customer site providing a mix of delivery and consulting services, so you'll be flexible about travel. Here is what we would like you to bring: ● Development and delivery experience with Java, .NET, Scala and the likes ● Passion for software engineering and craftsman-like coding prowess ● Great OO skills, including strong design patterns knowledge ● Experience working with Agile, Lean and/or Continuous Delivery approaches and best practices, such as Extreme Programming (XP) ● Keen to work collaboratively with people, sharing your ideas to solve real business problems.
Job Title: Distributed Systems Engineer - SDET Job Location: Pune, India Job Description: Are you looking to put your computer science skills to use? Are you looking to work for one of the hottest start-ups in Silicon Valley? Are you looking to define the next generation data management platform based on Apache Spark? Are you excited by the idea of being a Spark committer? If you answered yes to all of the questions above, we definitely want to talk to you. We are looking to add highly motivated engineers to work as a QE software engineer in our product development team in Pune. We work on cutting edge data management products that transform the way businesses operate. As a distributed systems engineer (if you are good) , you will get to work on defining key elements of our real time analytics platform, including 1. Distributed in memory data management 2. OLTP and OLAP querying in a single platform 3. Approximate Query Processing over large data sets 4. Online machine learning algorithms applied to streaming data sets 5. Streaming and continuous querying Requirements: 1. Experience in testing modern SQL, NewSQL products highly desirable 2. Experience with SQL language, JDBC, end to end testing of databases 3. Hands on Experience in writing SQL queries 4. Experience on database performance benchmarks like TPC-H, TPC-C and TPC-E a plus 5. Prior experience in benchmarking against Cassandra or MemSQL is a big plus 6. You should be able to program either in Java or have some exposure to functional programming in Scala 7. You should care about performance, and by that, we mean performance optimizations in a JVM 8. You should be self motivated and driven to succeed 9. If you are an open source committer on any project, especially an Apache project, you will fit right in 10. Experience working with Spark, SparkSQL, Spark Streaming is a BIG plus 11. Plans & authors Test plans and ensure testability is considered by development in all stages of the life cycle. 12. Plans, schedules and tracks the creations of Test plans / automation scripts using defined methodologies for manual and/or automated tests 13. Work as QE team member in troubleshooting, isolating, reproducing, tracking bugs and verifying fixes. 14. Analyze test results to ensure existing functionality and recommends corrective action. Documents test results, manages and maintains defect & test case databases to assist in process improvement and estimation of future releases. 15. Performs the assessment and planning of test efforts required for automation of new functions/features under development. Influences design changes to improve quality and feature testability. 16. If you have solved big complex problems, we want to talk to you 17. If you are a math geek, with a background in statistics, mathematics and you know what a linear regression is, this just might be the place for you 18. Exposure to stream data processing Storm, Samza is a plus Open source contributors: Send us your Github id Product: SnappyData is a new real-time analytics platform that combines probabilistic data structures, approximate query processing and in memory distributed data management to deliver powerful analytic querying and alerting capabilities on Apache Spark at a fraction of the cost of traditional big data analytics platforms. SnappyData fuses the Spark computational engine with a highly available, multi-tenanted in-memory database to execute OLAP and OLTP queries on streaming data. Further, SnappyData can store data in a variety of synopsis data structures to provide extremely fast responses on less resources. Finally, applications can either submit Spark programs or connect using JDBC/ODBC to run interactive or continuous SQL queries. Skills: 1. Distributed Systems, 2. Scala, 3. Apache Spark, 4. Spark SQL, 5. Spark Streaming, 6. Java, 7. YARN/Mesos What's in it for you: 1. Cutting edge work that is ultra meaningful 2. Colleagues who are the best of the best 3. Meaningful startup equity 4. Competitive base salary 5. Full benefits 6. Casual, Fun Office Company Overview: SnappyData is a Silicon Valley funded startup founded by engineers who pioneered the distributed in memory data business. It is advised by some of the legends of the computing industry who have been instrumental in creating multiple disruptions that have defined computing over the past 40 years. The engineering team that powers SnappyData built GemFire, one of the industry leading in memory data grids, which is used worldwide in mission critical applications ranging from finance to retail.
Racetrack is India’s 1st sales & support focused AI driven organization. Racetrack fuels organizations’ sales efforts and power sales teams with genuinely intelligent tools to effectively organize and expand the business horizon. Racetrack.ai works on the AI technologies like NLP, ML & DL in an integrated/inter-disciplinary manner to create the products to be as close to the human mind and intellect. With a passion to create a platform for businesses which can not only transform customer experience for higher customer acquisition and retention rates but also cuts down on expenses borne by businesses to achieve these goals racetrack.ai’s Marvin & Turing aims at covering various Processes of Sales and Support driven by AI technologies. Marvin- Is an AI Driven Communication BoT who acts like a Virtual intelligent sales agent who not only engages a website’s visitors intelligently but also make the conversations fruitful and more intellectual with every chat. Turning: Is a product recommender to up sell and cross sell for businesses. It understands the customer needs and help organizations to get a bigger share of wallet of the consumer. Our current industry focuses are BFSI, Health Care, Education, E-commerce, Telecom, Travel & Tourism, Real Estate etc.
Our client is a Gurgaon based software startup in Artificial Intelligence, Big Data and Data Science domain. Our clients, have built a data scientist, a virtual one. It's an Artificial Intelligence powered agent who can learn & work 24x7 to deliver business insights that matter the most. ● Working on a Unique concept ● Recognized by Indian Angel Network (IAN), the biggest network in India along with DIPP (Govt. Of India) and NASSCOM. ● Winner of $120K credits as part of Microsoft Bizspark Plus program ● Raised two professional rounds of funding ● Alumni of premier institutes (like IIT Bombay, IIT Delhi) on our advisory panel ● The current hiring is for the core team expansion. It will be under 10. I.e. candidate will be part of the core founding team and will get tremendous exposure. Core founding team focuses on invention and gets huge opportunities to file patents. ● Days: Monday to Saturdays ● One weekday off per month as per employee's choice. This is on top of earned/privilege leaves and bank holidays. ● Location: Gurgaon ● Line Management: Directly reporting to CxO Team Position: - Sr. Product Developer Job Description: Sr. Product Developer will be part of the client’s Lab. As a Sr. Product Developer, the candidate will be working very closely with Product Management, AI Research and Data Scientist team. Key responsibilities include, ● Design and Development of product in Big Data architecture and framework - Apache Spark, HDFS, Flume, Kafka with Scala/Java ● Development of Machine Learning Algorithms in Apache Spark framework ● Design and Development of integration connectors with external data sources like RDBMS (MySQL, Oracle etc.) and other products ● Lead, mentor and coach team members Skills ● 4+ Years of product development experience in Scala / Java ● Must possess in-depth knowledge of core design/architectural concerns like design patterns, performance, code re-usability and quality ● Should have good understanding of RDBMS and EA diagrams ● Experience on development (or upgrades) of Apache Spark libraries and contributions to Apache Spark or other open source frameworks is an added advantage ● Understanding on data security or statistics (like probability distribution etc.) is an added advantage ● Ability to take initiatives, self-motivated and a learning attitude is must. Experience & Qualification Required - B.Tech from Tier 1.5 (NIT/IIIT/DCE) with 4+ years of Experience
Looking for a technically sound and excellent trainer on big data technologies. Get an opportunity to become popular in the industry and get visibility. Host regular sessions on Big data related technologies and get paid to learn.