By leveraging our cross domain knowledge of Finance & Accounting, Retail and technology industries, we provide cost effective DATA ANALYTICS solutions to clients.
Good Understanding of Data structure and alogorithms. Working knowledge of -Scala/Python /Java Good Understanding of big data domain( Hadoop/map reduce- /ETL Architecture Hands on Experience of Hive/spark Experience:-3+yrs Immediate - 15 days max 30 days joiners.
ob Title/Designation:Mid / Senior Big Data EngineerJob Description:Role: Big Data EngineerNumber of open positions: 5Location: PuneAt Clairvoyant, we're building a thriving big data practice to help enterprises enable and accelerate the adoption of Big data and cloud services. In the big data space, we lead and serve as innovators, troubleshooters, and enablers. Big data practice at Clairvoyant, focuses on solving our customer's business problems by delivering products designed with best in class engineering practices and a commitment to keep the total cost of ownership to a minimum.Must Have: 4-10 years of experience in software development. At least 2 years of relevant work experience on large scale Data applications. Strong coding experience in Java is mandatory Good aptitude, strong problem solving abilities, and analytical skills, ability to take ownership as appropriate Should be able to do coding, debugging, performance tuning and deploying the apps to Prod. Should have good working experience on o Hadoop ecosystem (HDFS, Hive, Yarn, File formats like Avro/Parquet) o Kafka o J2EE Frameworks (Spring/Hibernate/REST) o Spark Streaming or any other streaming technology. Strong coding experience in Java is mandatory Ability to work on the sprint stories to completion along with Unit test case coverage. Experience working in Agile Methodology Excellent communication and coordination skills Knowledgeable (and preferred hands on) - UNIX environments, different continuous integration tools. Must be able to integrate quickly into the team and work independently towards team goals Role & Responsibilities: Take the complete responsibility of the sprint stories' execution Be accountable for the delivery of the tasks in the defined timelines with good quality. Follow the processes for project execution and delivery. Follow agile methodology Work with the team lead closely and contribute to the smooth delivery of the project. Understand/define the architecture and discuss the pros-cons of the same with the team Involve in the brainstorming sessions and suggest improvements in the architecture/design. Work with other team leads to get the architecture/design reviewed. Work with the clients and counter-parts (in US) of the project. Keep all the stakeholders updated about the project/task status/risks/issues if there are any. Education: BE/B.Tech from reputed institute.Experience: 4 to 9 yearsKeywords: java, scala, spark, software development, hadoop, hiveLocations: Pune
Excellent verbal and written communications skills, a knack for problem-solving, and a passion for delivering high-quality solutions. Ownership of the work – make it your mission to see it through. Strong organizational, prioritization, task, and team management skills – with an eye for detail and diligence towards a quality outcome. Strong commercial awareness that leads to driving profitable outcomes for the business. You, Will, Have 6+ years as a developer/team lead in the Business Intelligence field. Business Requirements Analysis. A strong understanding of financial reporting would a big plus. BI Architecture & Dimensional Modeling. Experience with Microsoft Business Intelligence Technologies (SSIS, SSAS, SSRS) Experience in Azure Data Platform services (Data Factory, Data Lake, Databricks, Synapse, etc.) Experience in delivering high-quality solutions using PowerBI. It is an Advantage to Have Strong understanding of ERP systems and business processes. D365/Power Platform. Multi-dimensional Modeling for Financial Planning, Budgeting, and Forecasting. Machine Learning. Relevant Microsoft certifications. R/Python, Statistical Modeling.
The Sr. NLP Text Mining Scientist / Engineer will have the opportunity to lead a team, shape team culture and operating norms as a result of the fast-paced nature of a new, high-growth organization.• 7+ years of Industry experience primarily related to Unstructured Text Data and NLP (PhD work and internships will be considered if they are related to unstructured text in lieu of industry experience but not more than 2 years will be accounted towards industry experience)• Develop Natural Language Medical/Healthcare documents comprehension related products to support Health business objectives, products and improve processing efficiency, reducing overall healthcare costs• Gather external data sets; build synthetic data and label data sets as per the needs for NLP/NLR/NLU • Apply expert software engineering skills to build Natural Language products to improve automation and improve user experiences leveraging unstructured data storage, Entity Recognition, POS Tagging, ontologies, taxonomies, data mining, information retrieval techniques, machine learning approach, distributed and cloud computing platforms• Own the Natural Language and Text Mining products — from platforms to systems for model training, versioning, deploying, storage and testing models with creating real time feedback loops to fully automated services• Work closely and collaborate with Data Scientists, Machine Learning engineers, IT teams and Business stakeholders spread out across various locations in US and India to achieve business goals• Provide mentoring to other Data Scientist and Machine Learning Engineers• Strong understanding of mathematical concepts including but not limited to linear algebra, Advanced calculus, partial differential equations and statistics including Bayesian approaches• Strong programming experience including understanding of concepts in data structures, algorithms, compression techniques, high performance computing, distributed computing, and various computer architecture• Good understanding and experience with traditional data science approaches like sampling techniques, feature engineering, classification and regressions, SVM, trees, model evaluations• Additional course work, projects, research participation and/or publications in Natural Language processing, reasoning and understanding, information retrieval, text mining, search, computational linguistics, ontologies, semantics • Experience with developing and deploying products in production with experience in two or more of the following languages (Python, C++, Java, Scala)• Strong Unix/Linux background and experience with at least one of the following cloud vendors like AWS, Azure, and Google for 2+ years• Hands on experience with one or more of high-performance computing and distributed computing like Spark, Dask, Hadoop, CUDA distributed GPU (2+ years)• Thorough understanding of deep learning architectures and hands on experience with one or more frameworks like tensorflow, pytorch, keras (2+ years)• Hands on experience with libraries and tools like Spacy, NLTK, Stanford core NLP, Genism, johnsnowlabs for 5+ years• Understanding business use cases and be able to translate them to team with a vision on how to implement• Identify enhancements and build best practices that can help to improve the productivity of the team.
Adept at Machine learning techniques and algorithms. Feature selection, dimensionality reduction, building and optimizing classifiers using machine learning techniques Data mining using state-of-the-art methods Doing ad-hoc analysis and presenting results Proficiency in using query languages such as N1QL, SQL Experience with data visualization tools, such as D3.js, GGplot, Plotly, PyPlot, etc. Creating automated anomaly detection systems and constant tracking of its performance Strong in Python is a must. Strong in Data Analysis and mining is a must Deep Learning, Neural Network, CNN, Image Processing (Must) Building analytic systems - data collection, cleansing and integration Experience with NoSQL databases, such as Couchbase, MongoDB, Cassandra, HBase
About slice slice is a fintech startup focused on India’s young population. We aim to build a smart, simple, and transparent platform to redesign the financial experience for millennials and bring success and happiness to people’s lives. Growing with the new generation is what we dream about and all that we want. We believe that personalization combined with an extreme focus on superior customer service is the key to build long-lasting relations with young people. About team/role In this role, you will have the opportunity to create a significant impact on our business & most importantly our customers through your technical expertise on data as we take on challenges that can reshape the financial experience for the next generation. If you are a highly motivated team player with a knack for problem solving through technology, then we have a perfect job for you. What you’ll do Work closely with Engineering and Analytics teams to assist in Schema Designing, Normalization of Databases, Query optimization etc. Work with AWS cloud services: S3, EMR, Glue, RDS Create new and improve existing infrastructure for ETL workflows from a wide variety of data sources using SQL, NoSQL and AWS big data technologies Manage and monitor performance, capacity and security of database systems and regularly perform server tuning and maintenance activities Debug and troubleshoot database errors Identify, design and implement internal process improvements; optimising data delivery, re-designing infrastructure for greater scalability, data archival Qualification: 2+ years experience working as a Data Engineer Experience with a scripting language - PYTHON preferably Experience with Spark and Hadoop technologies. Experience with AWS big data tools is a plus. Experience with SQL and NoSQL databases technologies like Redshift, MongoDB, Postgres/MySQL, bigQuery, Casandra. Experience on Graph DB (Neo4j and OrientDB) and Search DB (Elastic Search) is a plus. Experience in handling ETL JOBS
JD : ML/NLP Tech Lead- We are looking to hire an ML/NLP Tech lead who can own products for a technology perspective and manage a team of up to 10 members. You will play a pivotal role in re-engineering our products, transformation, and scaling of AssessEdWHAT ARE WE BUILDING :- A revolutionary way of providing continuous assessments of a child's skill and learning, pointing the way to the child's potential in the future. This as opposed to the traditional one-time, dipstick methodology of a test that hurriedly bundles the child into a slot, that in-turn - declares- the child to be fit for a career in a specific area or a particular set of courses that would perhaps get him somewhere. At the core of our system is a lot of data - both structured and unstructured. - We have books and questions and web resources and student reports that drive all our machine learning algorithms. Our goal is to not only figure out how a child is coping but to also figure out how to help him by presenting relevant information and questions to him in topics that he is struggling to learn.Required Skill sets :- Wisdom to know when to hustle and when to be calm and dig deep. Strong can do mentality, who is joining us to build on a vision, not to do a job.- A deep hunger to learn, understand, and apply your knowledge to create technology.- Ability and Experience tackling hard Natural Language Processing problems, to separate wheat from the chaff, knowledge of mathematical tools to succinctly describe the ideas to implement them in code.- Very Good understanding of Natural Language Processing and Machine Learning with projects to back the same.- Strong fundamentals in Linear Algebra, Probability and Random Variables, and Algorithms.- Strong Systems experience in Distributed Systems Pipeline: Hadoop, Spark, etc.- Good knowledge of at least one prototyping/scripting language: Python, MATLAB/Octave or R.- Good understanding of Algorithms and Data Structures.- Strong programming experience in C++/Java/Lisp/Haskell.- Good written and verbal communication.Desired Skill sets :- Passion for well-engineered product and you are - ticked off- when something engineered is off and you want to get your hands dirty and fix it.- 3+ yrs of research experience in Machine Learning, Deep Learning and NLP- Top tier peer-reviewed research publication in areas like Algorithms, Computer Vision/Image Processing, Machine Learning or Optimization (CVPR, ICCV, ICML, NIPS, EMNLP, ACL, SODA, FOCS etc)- Open Source Contribution (include the link to your projects, GitHub etc.)- Knowledge of functional programming.- International level participation in ACM ICPC, IOI, TopCoder, etc - International level participation in Physics or Math Olympiad- Intellectual curiosity about advanced math topics like Theoretical Computer Science, Abstract Algebra, Topology, Differential Geometry, Category Theory, etc.What can you expect :- Opportunity to work on the interesting and hard research problem, to see the real application of state-of-the-art research into practice.- Opportunity to work on important problems with big social impact: Massive, and direct impact of the work you do on the lives of students.- An intellectually invigorating, phenomenal work environment, with massive ownership and growth opportunities.- Learn effective engineering habits required to build/deploy large production-ready ML applications.- Ability to do quick iterations and deployments.- We would be excited to see you publish papers (though certain restrictions do apply).Website : http://Digitalaristotle.aiWork Location: - Bangalore
Job Overview :Your mission is to help lead team towards creating solutions that improve the way our business is run. Your knowledge of design, development, coding, testing and application programming will help your team raise their game, meeting your standards, as well as satisfying both business and functional requirements. Your expertise in various technology domains will be counted on to set strategic direction and solve complex and mission critical problems, internally and externally. Your quest to embracing leading-edge technologies and methodologies inspires your team to follow suit.Responsibilities and Duties :- As a Data Engineer you will be responsible for the development of data pipelines for numerous applications handling all kinds of data like structured, semi-structured &unstructured. Having big data knowledge specially in Spark & Hive is highly preferred.- Work in team and provide proactive technical oversight, advice development teams fostering re-use, design for scale, stability, and operational efficiency of data/analytical solutionsEducation level :- Bachelor's degree in Computer Science or equivalentExperience :- Minimum 5+ years relevant experience working on production grade projects experience in hands on, end to end software development- Expertise in application, data and infrastructure architecture disciplines- Expert designing data integrations using ETL and other data integration patterns- Advanced knowledge of architecture, design and business processes Proficiency in :- Modern programming languages like Java, Python, Scala- Big Data technologies Hadoop, Spark, HIVE, Kafka- Writing decently optimized SQL queries- Orchestration and deployment tools like Airflow & Jenkins for CI/CD (Optional)- Responsible for design and development of integration solutions with Hadoop/HDFS, Real-Time Systems, Data Warehouses, and Analytics solutions- Knowledge of system development lifecycle methodologies, such as waterfall and AGILE.- An understanding of data architecture and modeling practices and concepts including entity-relationship diagrams, normalization, abstraction, denormalization, dimensionalmodeling, and Meta data modeling practices.- Experience generating physical data models and the associated DDL from logical data models.- Experience developing data models for operational, transactional, and operational reporting, including the development of or interfacing with data analysis, data mapping,and data rationalization artifacts.- Experience enforcing data modeling standards and procedures.- Knowledge of web technologies, application programming languages, OLTP/OLAP technologies, data strategy disciplines, relational databases, data warehouse development and Big Data solutions.- Ability to work collaboratively in teams and develop meaningful relationships to achieve common goalsSkills :Must Know :- Core big-data concepts- Spark - PySpark/Scala- Data integration tool like Pentaho, Nifi, SSIS, etc (at least 1)- Handling of various file formats- Cloud platform - AWS/Azure/GCP- Orchestration tool - Airflow
About the job: - You will work with data scientists to architect, code and deploy ML models - You will solve problems of storing and analyzing large scale data in milliseconds - architect and develop data processing and warehouse systems - You will code, drink, breathe and live python, sklearn and pandas. It’s good to have experience in these but not a necessity - as long as you’re super comfortable in a language of your choice. - You will develop tools and products that provide analysts ready access to the data About you: - Strong CS fundamentals - You have strong experience in working with production environments - You write code that is clean, readable and tested - Instead of doing it second time, you automate it - You have worked with some of the commonly used databases and computing frameworks (Psql, S3, Hadoop, Hive, Presto, Spark, etc) - It will be great if you have one of the following to share - a kaggle or a github profile - You are an expert in one or more programming languages (Python preferred). Also good to have experience with python-based application development and data science libraries. - Ideally, you have 2+ years of experience in tech and/or data. - Degree in CS/Maths from Tier-1 institutes.
Description Must have Direct Hands- on, 4 years of experience, building complex Data Science solutions Must have fundamental knowledge of Inferential Statistics Should have worked on Predictive Modelling, using Python / R Experience should include the following, File I/ O, Data Harmonization, Data Exploration Machine Learning Techniques (Supervised, Unsupervised) Multi- Dimensional Array Processing Deep Learning NLP, Image Processing Prior experience in Healthcare Domain, is a plus Experience using Big Data, is a plus Should have Excellent Analytical, Problem Solving ability. Should be able to grasp new concepts quickly Should be well familiar with Agile Project Management Methodology Should have excellent written and verbal communication skills Should be a team player with open mind