At least 10 years of hands-on experience in migration of complex software packages and products to Azure (Cloud Service Providers CSP) IaaS and PaaS At least 7 years of hands-on experience on programming and scripting languages (.Net, C#, WCF, MVC Web API, SQL Server, SQL Azure, Powershell). Good to have experience in IT systems, operations, automation and configuration tools to enable continuous-integration and deployment (Jenkins) Solid understanding of database management systems–traditional RDBMS ( MS SQL) Ability to wear multiple hats spanning the software-development- life-cycle across Requirements, Design, Code Development, QA, Testing, and Deployment –experience working in an Agile/Scrum methodology Analytical and Communication skills
Responsibilities: Lead backend team and mentor junior engineers Design and implement REST-based APIs and microservices in Node.js Write and maintain scalable, performant code that can be shared across platforms Work and communicate with our mobile client developers and project managers, managing priorities and giving input on future features About You : You attended a top university, studying computer science or similar You have experience or interest in writing applications in Node.js You have strong server-side development and experience with databases You understand the ins and outs of RESTful web services You know your way around the UNIX command line You have great communication skills and ability to work with others You are a strong team player, with a do-whatever-it-takes attitude
Job Title: Distributed Systems Engineer - SDET Job Location: Pune, India Job Description: Are you looking to put your computer science skills to use? Are you looking to work for one of the hottest start-ups in Silicon Valley? Are you looking to define the next generation data management platform based on Apache Spark? Are you excited by the idea of being a Spark committer? If you answered yes to all of the questions above, we definitely want to talk to you. We are looking to add highly motivated engineers to work as a QE software engineer in our product development team in Pune. We work on cutting edge data management products that transform the way businesses operate. As a distributed systems engineer (if you are good) , you will get to work on defining key elements of our real time analytics platform, including 1. Distributed in memory data management 2. OLTP and OLAP querying in a single platform 3. Approximate Query Processing over large data sets 4. Online machine learning algorithms applied to streaming data sets 5. Streaming and continuous querying Requirements: 1. Experience in testing modern SQL, NewSQL products highly desirable 2. Experience with SQL language, JDBC, end to end testing of databases 3. Hands on Experience in writing SQL queries 4. Experience on database performance benchmarks like TPC-H, TPC-C and TPC-E a plus 5. Prior experience in benchmarking against Cassandra or MemSQL is a big plus 6. You should be able to program either in Java or have some exposure to functional programming in Scala 7. You should care about performance, and by that, we mean performance optimizations in a JVM 8. You should be self motivated and driven to succeed 9. If you are an open source committer on any project, especially an Apache project, you will fit right in 10. Experience working with Spark, SparkSQL, Spark Streaming is a BIG plus 11. Plans & authors Test plans and ensure testability is considered by development in all stages of the life cycle. 12. Plans, schedules and tracks the creations of Test plans / automation scripts using defined methodologies for manual and/or automated tests 13. Work as QE team member in troubleshooting, isolating, reproducing, tracking bugs and verifying fixes. 14. Analyze test results to ensure existing functionality and recommends corrective action. Documents test results, manages and maintains defect & test case databases to assist in process improvement and estimation of future releases. 15. Performs the assessment and planning of test efforts required for automation of new functions/features under development. Influences design changes to improve quality and feature testability. 16. If you have solved big complex problems, we want to talk to you 17. If you are a math geek, with a background in statistics, mathematics and you know what a linear regression is, this just might be the place for you 18. Exposure to stream data processing Storm, Samza is a plus Open source contributors: Send us your Github id Product: SnappyData is a new real-time analytics platform that combines probabilistic data structures, approximate query processing and in memory distributed data management to deliver powerful analytic querying and alerting capabilities on Apache Spark at a fraction of the cost of traditional big data analytics platforms. SnappyData fuses the Spark computational engine with a highly available, multi-tenanted in-memory database to execute OLAP and OLTP queries on streaming data. Further, SnappyData can store data in a variety of synopsis data structures to provide extremely fast responses on less resources. Finally, applications can either submit Spark programs or connect using JDBC/ODBC to run interactive or continuous SQL queries. Skills: 1. Distributed Systems, 2. Scala, 3. Apache Spark, 4. Spark SQL, 5. Spark Streaming, 6. Java, 7. YARN/Mesos What's in it for you: 1. Cutting edge work that is ultra meaningful 2. Colleagues who are the best of the best 3. Meaningful startup equity 4. Competitive base salary 5. Full benefits 6. Casual, Fun Office Company Overview: SnappyData is a Silicon Valley funded startup founded by engineers who pioneered the distributed in memory data business. It is advised by some of the legends of the computing industry who have been instrumental in creating multiple disruptions that have defined computing over the past 40 years. The engineering team that powers SnappyData built GemFire, one of the industry leading in memory data grids, which is used worldwide in mission critical applications ranging from finance to retail.
Responsibilities Exp 3~5 years Build up a strong and scalable crawler system for leveraging external user & content data source from Facebook, Youtube and others internet products or service. Getting top trending keywords & topic from social media. Design and build initial version of the real-time analytics product from Machine Learning Models to recommend video contents in real time to 10M+ User Profiles independently. Architect and build Big Data infrastructures using Java, Kafka, Storm, Hadoop, Spark and other related frameworks, experience with Elastic search is a plus Excellent Analytical, Research and Problem Solving skills, in-depth knowledge of Data Structure Desired Skills and Experience B.S./M.S. degree in computer science, mathematics, statistics or a similar quantitative field with good college background 3+ years of work experience in relevant field (Data Engineer, R&D engineer, etc) Experience in Machine Learning and Prediction & Recommendation techniques Experience with Hadoop/MapReduce/Elastic-Stack/ELK and Big Data querying tools, such as Pig, Hive, and Impala Proficiency in a major programming language (e.g. Java/C/Scala) and/or a scripting language (Python) Experience with one or more NoSQL databases, such as MongoDB, Cassandra, HBase, Hive, Vertica, Elastic Search Experience with cloud solutions/AWS, strong knowledge in Linux and Apache Experience with any map-reduce SPARK/EMR Experience in building reports and/or data visualization Strong communication skills and ability to discuss the product with PMs and business owners
Our client is a Gurgaon based software startup in Artificial Intelligence, Big Data and Data Science domain. Our clients, have built a data scientist, a virtual one. It's an Artificial Intelligence powered agent who can learn & work 24x7 to deliver business insights that matter the most. ● Working on a Unique concept ● Recognized by Indian Angel Network (IAN), the biggest network in India along with DIPP (Govt. Of India) and NASSCOM. ● Winner of $120K credits as part of Microsoft Bizspark Plus program ● Raised two professional rounds of funding ● Alumni of premier institutes (like IIT Bombay, IIT Delhi) on our advisory panel ● The current hiring is for the core team expansion. It will be under 10. I.e. candidate will be part of the core founding team and will get tremendous exposure. Core founding team focuses on invention and gets huge opportunities to file patents. ● Days: Monday to Saturdays ● One weekday off per month as per employee's choice. This is on top of earned/privilege leaves and bank holidays. ● Location: Gurgaon ● Line Management: Directly reporting to CxO Team Position: - Sr. Product Developer Job Description: Sr. Product Developer will be part of the client’s Lab. As a Sr. Product Developer, the candidate will be working very closely with Product Management, AI Research and Data Scientist team. Key responsibilities include, ● Design and Development of product in Big Data architecture and framework - Apache Spark, HDFS, Flume, Kafka with Scala/Java ● Development of Machine Learning Algorithms in Apache Spark framework ● Design and Development of integration connectors with external data sources like RDBMS (MySQL, Oracle etc.) and other products ● Lead, mentor and coach team members Skills ● 4+ Years of product development experience in Scala / Java ● Must possess in-depth knowledge of core design/architectural concerns like design patterns, performance, code re-usability and quality ● Should have good understanding of RDBMS and EA diagrams ● Experience on development (or upgrades) of Apache Spark libraries and contributions to Apache Spark or other open source frameworks is an added advantage ● Understanding on data security or statistics (like probability distribution etc.) is an added advantage ● Ability to take initiatives, self-motivated and a learning attitude is must. Experience & Qualification Required - B.Tech from Tier 1.5 (NIT/IIIT/DCE) with 4+ years of Experience
Looking for a technically sound and excellent trainer on big data technologies. Get an opportunity to become popular in the industry and get visibility. Host regular sessions on Big data related technologies and get paid to learn.
Desired Skills and Experience Strong platform / infrastructure automation expertise or DevOps experience. Experience in Cloud Platforms – Amazon AWS, Cloudstack, Openstack would be a huge advantage. Working with automation technologies (Puppet, Chef, Ansible) is highly desirable. Experience with Git/Github, Gerrit/ReviewBoard/Phabricator, Jenkins/Hudson and/or other build and continuous integration systems are highly desirable. Experience with scripting languages – Bash, Python, Ruby, Powershell. Ability to build, create, version control and deploy software repositoriesExperience in proposal writing, presales, architecture, designing code is a must. Strong problem solving and debugging skills. Open source contributions in cloud domain will be a huge plus. Experience with virtualization technologies – VMware, Xenserver, Microsoft Hyper-V. Prior experience with server automation products – BMC Bladelogic, Microsoft System Configuration Manager would be a huge plus. Knowledge of ITIL is an added bonus.
Sigmoid is a fast growing Product Based BIG DATA startup. Sequoia Funded & Backed by experienced Professionals & Advisors. Sigmoid is revolutionizing business intelligence and analytics by providing unified tools for historical and real time analysis on Apache Spark. With their suite of products, Sigmoid is democratizing streaming use-cases like RTB Data Analytics, Log Analytics, Fraud Detection, Sensor Data Analytics etc. Sigmoid can enable the customers’ engineering team to set up their infrastructure on Spark and ramp up their development timelines, or enable the analytics team to derive insights from their data. Sigmoid has created a real time exploratory analytics tool using on Apache SPARK which not only vastly improves performance but also reduces the cost. A user can quickly analyse huge volumes of data, filter through multiple dimensions, compare results across time periods and carry out root cause analysis in a matter of seconds. Leading organisations across industry verticals are currently using Sigmoid’s platform in production to create success stories. ------------------------------------ What Sigmoid offers you: Work in a well-funded (Sequoia Capital) Big Data company. Deal with Terabytes of data on a regular basis. Opportunity to contribute to top big data projects. Work on complex problems faced by leading global companies in multiple areas such as fraud detection, real-time analytics, pricing modeling and so on ------------------------------------------------------ We are looking for Someone who has: 6+ years of demonstrable experience designing technological solutions to complex data problems, developing efficient and scalable code. Experience in Design, Architecture, Development of Big Data Technologies. Provides Technical leadership in Big Data space (Apache Spark, Kafka, Flink, Hadoop, MapReduce, HDFS, Hive, HBase, Flume, Sqoop, NoSQL, Cassandra, HBase) Strong understanding of databases and SQL. Defines and Drives best practices in Big Data stack. Drives operational excellence through root cause analysis and continuous improvement for Big Data technologies and processes. Operating knowledge of cloud computing platforms (AWS and/ or Azure or Google Cloud). Mentors/coaches engineers to facilitate their development and provide technical leadership to them A technologist who Loves to code and design and have great problem-solving skills, and the ability & confidence to hack their way out of tight corners. In addition, the ideal candidate would have great problem-solving skills, and the ability & confidence to hack their way out of tight corners. ------------------------------------ Preferred Qualifications: Engineering Bachelors/Masters in Computer Science/IT. Top Tier Colleges (IIT, NIT, IIIT, etc) will be preferred. Salary is not a constraint for the right talent.
Experience : Minimum of 3 years of relevant development experience Qualification : BS in Computer Science or equivalent Skills Required: • Server side developers with good server side development experience in Java AND/OR Python • Exposure to Data Platforms (Cassandra, Spark, Kafka) will be a plus • Interested in Machine Learning will be a plus • Good to great problem solving and communication skill • Ability to deliver in an extremely fast paced development environment • Ability to handle ambiguity • Should be a good team player Job Responsibilities : • Learn the technology area where you are going to work • Develop bug free, unit tested and well documented code as per requirements • Stringently adhere to delivery timelines • Provide mentoring support to Software Engineer AND/ OR Associate Software Engineers • Any other as specified by the reporting authority
Data Scientist - We are looking for a candidate to build great recommendation engines and power an intelligent m.Paani user journey Responsibilities : - Data Mining using methods like associations, correlations, inferences, clustering, graph analysis etc. - Scale machine learning algorithm that powers our platform to support our growing customer base and increasing data volume - Design and implement machine learning, information extraction, probabilistic matching algorithms and models - Care about designing the full machine learning pipeline. - Extending company's data with 3rd party sources. - Enhancing data collection procedures. - Processing, cleaning and verifying data collected. - Ad hoc analysis of the data and present clear results. - Creating advanced analytics products that provide actionable insights. The Individual : - We are looking for a candidate with the following skills, experience and attributes: Required : - Someone with 2+ years of work experience in machine learning. - Educational qualification relevant to the role. Degree in Statistics, certificate courses in Big Data, Machine Learning etc. - Knowledge of Machine Learning techniques and algorithms. - Knowledge in languages and toolkits like Python, R, Numpy. - Knowledge of data visualization tools like D3,js, ggplot2. - Knowledge of query languages like SQL, Hive, Pig . - Familiar with Big Data architecture and tools like Hadoop, Spark, Map Reduce. - Familiar with NoSQL databases like MongoDB, Cassandra, HBase. - Good applied statistics skills like distributions, statistical testing, regression etc. Compensation & Logistics : This is a full-time opportunity. Compensation will be in line with startup, and will be based on qualifications and experience. The position is based in Mumbai, India, and the candidate must live in Mumbai or be willing to relocate.