Responsibilities for Data Engineer • Create and maintain optimal data pipeline architecture, • Assemble large, complex data sets that meet functional / non-functional business requirements. • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc. • Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency and other key business performance metrics. • Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs. • Keep our data separated and secure across national boundaries through multiple data centers and AWS regions. • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader. • Work with data and analytics experts to strive for greater functionality in our data systems.
AppLift is a data-driven technology company that empowers mobile app advertisers to acquire and re-engage quality users at scale on a performance basis. AppLift’s programmatic media buying platform DataLift provides access to all automated supply sources in the market, reaching over a billion users. The technology leverages first- and third-party data to optimize media buys across all stages of the conversion funnel and, through its proprietary LTV optimization technology, enables ROI-maximized user acquisition. AppLift is trusted by 500+ leading global advertisers across all verticals, such as King, Zynga, OLX, Glu Mobile, Myntra, Paltalk, Nexon, and Tap4Fun. Experience:4-8yrs Your Responsibilities: You are hands on with data, implementation and methodologies You are able to implement, measure and evaluate different algorithmic approaches, along with great problem solving skills and strong theoretical foundation You have expertise in implementing machine-learning and algorithmic concepts. Proficiency in coding and understanding of engineering trade-offs You innovate and develop approaches to improve click/conversion rate, eliminate impression/click fraud and enhance bidding strategies You perform statistical analysis, data mining to model user behaviour and improve ad-relevance You are able to work independently and deliver practical results on real data with high accountability Our requirements: 4+ years of experience in applied research or industry work experience Degree in statistics, applied mathematics, machine learning, or other highly quantitative experience Experience in working with technologies and tools like R, Graphlab, Hadoop, Hive, Spark, Pig Coding proficiency in at-least one language like python or Java Prior experience in Ad-Tech, a plus What do we offer? You get valuable insights into mobile marketing/entrepreneurship and have a high impact on shaping the expansion and success of AppLift across India Profit from working with European Serial Entrepreneurs who co-founded over 10 successful companies within the last 8 years and get access to a well-established network and be able to build your own top-tier network & reputation Learn and grow in an environment characterized by flat hierarchy, entrepreneurial drive and fun You experience an excellent learning culture Competitive remuneration package and much more! if interested mail your resume at divya.pushpa<at>applift.com Candidates are preferred from Tier 1 colleages
Data Architect who leads a team of 5 numbers. Required skills : Spark ,Scala , hadoop
AlgonoX Technologies is looking for strong Spark developers. Experience : 3 - 10 years Location : Mumbai Candidates who have experience in the following shall be preferred. Hands on development experience with Spark (Scala/Java). Data prep for running analytical programs. Merging data. Fuzzy data match. Transforming complex data forms. Fixing errors in real time. Writing extensive Spark pipeline. Note : Folks with data moving experience may kindly ignore. Interested may forward their resume to email@example.com
We are looking for developers with 4-8 years experience in Java and Big data Work Location - Bangalore Qualification - Any graduate or Post graduate from tier 1 and tier 2 colleges. Strong programming skills in Java and Big Data Hands on experience with Big Data systems like Spark, HDFS, Hive, Kafka Excellent written and verbal communication skills with the ability to communicate the design of algorithms and systems to other members of the group Interested candidates share the resume to Shyam.firstname.lastname@example.org Shyam Sugathan Senior Recruiter- Human Resources Shyam.email@example.com Floor 2nd | Noor Complex | Mavoor Road | Calicut-4 | Kerala | India. | www.hapstive.com
Byte Prophecy is looking for Data Engineers to build a critical piece of the enterprise data pipeline in our platform MonitorFirst. Candidates should - Have at least 1-2 years of relevant experience in any of the following technologies in our Data Pipeline [#ETL Tools, #Kafka, #Spark, #Cassandra, #Scala or #Python] - Be hands-on and proficient in #Java, #Scala and #SQL - Have strong fundamentals in data structures, algorithms and distributed systems - Experienced in product engineering and production-ready data pipelines is preferred Candidates will - Work in an agile environment as small focused teams - Need to be proactive and goal-oriented - Enjoy working as a team-player About Byte Prophecy We are an enterprise analytics platform company that helps some of the largest companies in India make key business decisions every day. As a unique single platform encompassing collection, transformation, processing, augmented analytics and automated alerts, we've been getting great traction from key stakeholders in the enterprise ecosystem. For our next round of growth, we are looking to hire Data Engineers and Product Analysts for our office in Ahmedabad. Please send your CVs to firstname.lastname@example.org Thanks!
Akridata is a US based early stage startup founded by VxTel and Virident founders incorporated in May 2018. The startup is looking into addressing challenges in certain edge data use cases involving extremely high volume and high BW data generation and processing. The team in Bangalore fully owns the SW development for the product and early stage of the startup implies ample opportunities for 'from-scratch' design and development. What we are looking for i. Strong CS fundamentals and algorithms. ii. Hands-on programming experience preferably in high level languages like Scala/Java/Go. iii) Hands-on experience with design and development of scalable distributed systems like distributed file systems, streaming systems, big data infrastructure systems, NoSQL database systems etc iv. Willing and enthusiastic about learning different technologies based on project requirements Technology areas we work i) big data components - Spark, Hadoop, HDFS ii) ML components - Tensorflow, Spark-MLlib iii) Cloud(AWS) hosted scalable data and control path SW iv) High performance data paths(ML related) with usage of GPUs v) Algorithms around efficient techniques for data summarization What we offer i) A startup environment providing opportunities to have a significant influence on the SW road-map with ample learning opportunities. ii) The excitement of a startup with reasonable financial stability from ample amount of seed funding. iii) Competitive compensation and benefits along-with attractive employee stock option plan
We are looking to hire passionate Java techies who will be comfortable learning and working on Java and any open source frameworks & technologies. She/he should be a 100% hands-on person on technology skills and interested in solving complex analytics use cases. We are working on a complete stack platform which has already been adopted by some very large Enterprises across the world. Candidates with prior experience of having worked in typical R&D environment and/or product based companies with dynamic work environment will be have an additional edge. We currently work on some of the latest technologies like Cassandra, Hadoop, Apache Solr, Spark and Lucene, and some core Machine Learning and AI technologies. Even though prior knowledge of these skills is not mandatory at all for selection, you would be expected to learn new skills on the job.
Looking for Big data Developers in Mumbai Location
Position Description Assists in providing guidance to small groups of two to three engineers, including offshore associates, for assigned Engineering projects Demonstrates up-to-date expertise in Software Engineering and applies this to the development, execution, and improvement of action plans Generate weekly, monthly and yearly report using JIRA and Open source tools and provide updates to leadership teams. Proactively identify issues, identify root cause for the critical issues. Work with cross functional teams, Setup KT sessions and mentor the team members. Co-ordinate with Sunnyvale and Bentonville teams. Models compliance with company policies and procedures and supports company mission, values, and standards of ethics and integrity Provides and supports the implementation of business solutions Provides support to the business Troubleshoots business, production issues and on call support. Minimum Qualifications BS/MS in Computer Science or related field 8+ years’ experience building web applications Solid understanding of computer science principles Excellent Soft Skills Understanding the major algorithms like searching and sorting Strong skills in writing clean code using languages like Java and J2EE technologies. Understanding how to engineer the RESTful, Micro services and knowledge of major software patterns like MVC, Singleton, Facade, Business Delegate Deep knowledge of web technologies such as HTML5, CSS, JSON Good understanding of continuous integration tools and frameworks like Jenkins Experience in working with the Agile environments, like Scrum and Kanban. Experience in dealing with the performance tuning for very large-scale apps. Experience in writing scripting using Perl, Python and Shell scripting. Experience in writing jobs using Open source cluster computing frameworks like Spark Relational database design experience- MySQL, Oracle, SOLR, NoSQL - Cassandra, Mango DB and Hive. Aptitude for writing clean, succinct and efficient code. Attitude to thrive in a fun, fast-paced start-up like environment
Couture.ai is building a patent-pending AI platform targeted towards vertical-specific solutions. The platform is already licensed by Reliance Jio and few European retailers, to empower real-time experiences for their combined >200 million end users. For this role, credible display of innovation in past projects (or academia) is a must. We are looking for a candidate who lives and talks Data & Algorithms, love to play with BigData engineering, hands-on with Apache Spark, Kafka, RDBMS/NoSQL DBs, Big Data Analytics and handling Unix & Production Server. Tier-1 college (BE from IITs, BITS-Pilani, top NITs, IIITs or MS in Stanford, Berkley, CMU, UW–Madison) or exceptionally bright work history is a must. Let us know if this interests you to explore the profile further.
Couture.ai is building a patent-pending AI platform targeted towards vertical-specific solutions. The platform is already licensed by Reliance Jio and few European retailers, to empower real-time experiences for their combined >200 million end users. The founding team consists of BITS Pilani alumni with experience of creating global startup success stories. The core team, we are building, consists of some of the best minds in India in artificial intelligence research and data engineering. We are looking for multiple different roles with 2-7 year of research/large-scale production implementation experience with: - Rock-solid algorithmic capabilities. - Production deployments for massively large-scale systems, real-time personalization, big data analytics, and semantic search. - Or credible research experience in innovating new ML algorithms and neural nets. Github profile link is highly valued. For right fit into the Couture.ai family, compensation is no bar.
Couture.ai is building a patent-pending AI platform targeted towards vertical-specific solutions. The platform is already licensed by Reliance Jio and few European retailers, to empower real-time experiences for their combined >200 million end users. For this role, credible display of innovation in past projects is a must. We are looking for hands-on leaders in data engineering with the 5-11 year of research/large-scale production implementation experience with: - Proven expertise in Spark, Kafka, and Hadoop ecosystem. - Rock-solid algorithmic capabilities. - Production deployments for massively large-scale systems, real-time personalization, big data analytics and semantic search. - Expertise in Containerization (Docker, Kubernetes) and Cloud Infra, preferably OpenStack. - Experience with Spark ML, Tensorflow (& TF Serving), MXNet, Scala, Python, NoSQL DBs, Kubernetes, ElasticSearch/Solr in production. Tier-1 college (BE from IITs, BITS-Pilani, IIITs, top NITs, DTU, NSIT or MS in Stanford, UC, MIT, CMU, UW–Madison, ETH, top global schools) or exceptionally bright work history is a must. Let us know if this interests you to explore the profile further.
RESPONSIBILITIES: 1. Full ownership of Tech right from driving product decisions to architect to deployment. 2. Develop cutting edge user experience and build cutting edge technology solutions like instant messaging in poor networks, live-discussions, live-videos optimal matching. 3. Using Billions of Data Points to Build User Personalisation Engine 4. Building Data Network Effects Engine to increase Engagement & Virality 5. Scaling the Systems to Billions of Daily Hits. 6. Deep diving into performance, power management, memory optimisation & network connectivity optimisation for the next Billion Indians 7. Orchestrating complicated workflows, asynchronous actions, and higher order components 8. Work directly with Product and Design teams REQUIREMENTS: 1. Should have Hacked some (computer or non-computer) system to your advantage. 2. Built and managed systems with a scale of 10Mn+ Daily Hits 3. Strong architectural experience 4. Strong experience in memory management, performance tuning and resource optimisations 5. PREFERENCE- If you are a woman or an ex-entrepreneur or having a CS bachelor’s degree from IIT/BITS/NIT P.S. If you don't fulfil one of the requirements, you need to be exceptional in the others to be considered.
ABOUT US: Arque Capital is a FinTech startup working with AI in Finance in domains like Asset Management (Hedge Funds, ETFs and Structured Products), Robo Advisory, Bespoke Research, Alternate Brokerage, and other applications of Technology & Quantitative methods in Big Finance. PROFILE DESCRIPTION: 1. Get the "Tech" in order for the Hedge Fund - Help answer fundamentals of technology blocks to be used, choice of certain platform/tech over other, helping team visualize product with the available resources and assets 2. Build, manage, and validate a Tech Roadmap for our Products 3. Architecture Practices - At startups, the dynamics changes very fast. Making sure that best practices are defined and followed by team is very important. CTO’s may have to garbage guy and clean the code time to time. Making reviews on Code Quality is an important activity that CTO should follow. 4. Build progressive learning culture and establish predictable model of envisioning, designing and developing products 5. Product Innovation through Research and continuous improvement 6. Build out the Technological Infrastructure for the Hedge Fund 7. Hiring and building out the Technology team 8. Setting up and managing the entire IT infrastructure - Hardware as well as Cloud 9. Ensure company-wide security and IP protection REQUIREMENTS: Computer Science Engineer from Tier-I colleges only (IIT, IIIT, NIT, BITS, DHU, Anna University, MU) 5-10 years of relevant Technology experience (no infra or database persons) Expertise in Python and C++ (3+ years minimum) 2+ years experience of building and managing Big Data projects Experience with technical design & architecture (1+ years minimum) Experience with High performance computing - OPTIONAL Experience as a Tech Lead, IT Manager, Director, VP, or CTO 1+ year Experience managing Cloud computing infrastructure (Amazon AWS preferred) - OPTIONAL Ability to work in an unstructured environment Looking to work in a small, start-up type environment based out of Mumbai COMPENSATION: Co-Founder status and Equity partnership
US based Multinational Company Hands on Hadoop
Looking for a technically sound and excellent trainer on big data technologies. Get an opportunity to become popular in the industry and get visibility. Host regular sessions on Big data related technologies and get paid to learn.
candidate will be responsible for all aspects of data acquisition, data transformation, and analytics scheduling and operationalization to drive high-visibility, cross-division outcomes. Expected deliverables will include the development of Big Data ELT jobs using a mix of technologies, stitching together complex and seemingly unrelated data sets for mass consumption, and automating and scaling analytics into the GRAND's Data Lake. Key Responsibilities : - Create a GRAND Data Lake and Warehouse which pools all the data from different regions and stores of GRAND in GCC - Ensure Source Data Quality Measurement, enrichment and reporting of Data Quality - Manage All ETL and Data Model Update Routines - Integrate new data sources into DWH - Manage DWH Cloud (AWS/AZURE/Google) and Infrastructure Skills Needed : - Very strong in SQL. Demonstrated experience with RDBMS, Unix Shell scripting preferred (e.g., SQL, Postgres, Mongo DB etc) - Experience with UNIX and comfortable working with the shell (bash or KRON preferred) - Good understanding of Data warehousing concepts. Big data systems : Hadoop, NoSQL, HBase, HDFS, MapReduce - Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments. - Working with data delivery teams to set up new Hadoop users. This job includes setting up Linux users, setting up and testing HDFS, Hive, Pig and MapReduce access for the new users. - Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, and other tools. - Performance tuning of Hadoop clusters and Hadoop MapReduce routines. - Screen Hadoop cluster job performances and capacity planning - Monitor Hadoop cluster connectivity and security - File system management and monitoring. - HDFS support and maintenance. - Collaborating with application teams to install operating system and - Hadoop updates, patches, version upgrades when required. - Defines, develops, documents and maintains Hive based ETL mappings and scripts
Our company is working on some really interesting projects in Big Data Domain in various fields (Utility, Retail, Finance). We are working with some big corporates and MNCs around the world. While working here as Big Data Engineer, you will be dealing with big data in structured and unstructured form and as well as streaming data from Industrial IOT infrastructure. You will be working on cutting edge technologies and exploring many others while also contributing back to the open-source community. You will get to know and work on end-to-end processing pipeline which deals with all type of work like storing, processing, machine learning, visualization etc.
Full Stack Developer for Big Data Practice. Will include everything from architecture to ETL to model building to visualization.
We at InfoVision Labs, are passionate about technology and what our clients would like to get accomplished. We continuously strive to understand business challenges, changing competitive landscape and how the cutting edge technology can help position our client to the forefront of the competition.We are a fun loving team of Usability Experts and Software Engineers, focused on Mobile Technology, Responsive Web Solutions and Cloud Based Solutions. Job Responsibilities: ◾Minimum 3 years of experience in Big Data skills required. ◾Complete life cycle experience with Big Data is highly preferred ◾Skills – Hadoop, Spark, “R”, Hive, Pig, H-Base and Scala ◾Excellent communication skills ◾Ability to work independently with no-supervision.
zeotap helps telecom operators unlock the potential of their data safely across industries using privacy-by-design technology http://www.zeotap.com
Check our JD: https://www.zeotap.com/job/senior-tech-lead-m-f-for-zeotap/oEQK2fw0
Ixsight Technologies is an innovative IT company with strong Intellectual Property. Ixsight is focused on creating Customer Data Value through its solutions for Identity Management, Locational Analytics, Address Science and Customer Engagement. Ixsight is also adapting its solutions to Big Data and Cloud. We are in the process of creating new solutions across platforms. Ixsight has served over 80+ clients in India – for various end user applications across traditional BFSI and telecom sector. In the recent past we are catering to the new generation verticals – Hospitality, ecommerce etc. Ixsight has been featured in the Gartner’s India Technology Hype Cycle and has been recognised by both clients and peers for pioneering and excellent solutions. If you wish to play a direct part in creating new products, building IP and being part of Product Creation - Ixsight is the place.