We are looking for python developers who have good exposure to distributed systems. This will be an IC role, candidates with excellent design and coding skills are preferred. Must be willing to adapt to new languages as the team is very dynamic in nature. ABOUT THE JOB Is Big Data really big? If you want to explore this area, learn what massive data volumes mean and how internet works, join Distributed Data Engineering team – a small group of elite software engineers that analyze, design and implement system software that brings new functionality, increased reliability, and enhanced scalability to Akamai’s high-performance Distributed Data platform. ABOUT THE TEAM The Distributed Data Engineering team (DDE) develops and operates the networks that process aggregate and store data about every transaction that involves Akamai edge network servers. Data owned by DDE is being consumed for the purposes of customers billing, customer analytics, business decisions support, Akamai’s cost structure management and Akamai’s network management. DDE currently receives over 2PB of data/day and maintains a data store that processes 3 trillion records daily. The product development team within DDE has end-to-end responsibility for the design, development and deployment of the platform components that enable one of the world’s largest cloud-based data systems. ELEVATOR PITCH 3 REASONS WHY A GREAT CANDIDATE SHOULD GET ATTRACTED TO THIS OPPORTUNITY. 1. This role plays a very critical role in performance-critical message brokering subsystem 2. This role demands to analyze, design and implement system software that brings new functionality, increased reliability, and enhanced scalability to Akamai’s high-performance Distributed Data platform. 3. The expectation is to take the ownership of the design of the platform components that enable one of the world’s largest cloud-based data systems. RESPONSIBILITIES • Develop new and enhance existing features for DDE's massively distributed system • Work on performance-critical message brokering subsystem • Work on data collection, processing, and access subsystems • Work on projects that focus on system scalability, performance, and security • Drive feature development from idea inception through design and testing to operational deployment • Follow SW development methodology best practices, including collaboration with QA departments to successfully deploy high-quality new system components BASIC QUALIFICATIONS • BS in Computer Science or equivalent, MS preferred • 6+ years of experience developing SW on Python • 3+ years of experience with Linux and distributed systems • Knowledge of networking principles, including TCP/IP, SSH, SSL and HTTP protocols • Knowledge of software development and design principles • Ability to troubleshoot complex network problems and customer issues DESIRED QUALIFICATIONS • Proven track record of delivering large amounts of high quality, complex code • Highly responsible, motivated, able to work with little supervision • Experience with BigData systems (Hadoop, Spark, Kafka etc.) and principles (Map/Reduce, etc) • Experience with scripting, e.g. Perl, Python, bash and API's such as SOAP and/or RESTful • Experience with DBMS, e.g. PostGRE SQL, MySQL, etc
Please find the below JD: Minimum 4-7 years’ experience with strong C/C++ & socket programming skills in Linux operating system environment. Have a good exposure or experience to signaling protocols - MAP/TCAP,SCCP,M3UA, SCTP protocols Should have good 2G domain knowledge. Good in Operating System, IPC communication. Hands on with software development/test tools like: gdb, valgrind. Networking tools like Wireshark Very good exposure towards Network Concepts, trouble sheetiing and experiencing Network programming . Well versed with TCP,UDP communications. Hands on in Data structures and problem solving skills.
ProMobi Technologies Pvt. Ltd. System Administrator (Linux) Experience - 2 to 4 years Location - Pune We are looking for a Smart and Hands-on Systems Administrator, who would be responsible for managing upkeep, configuration, installing, supporting, and maintaining servers and reliable operation of servers. Ensure performance, resources, and security of the computers and meet the needs of the users Key Responsibilities Ensure that systems are running smoothly, are being monitored, and respond to issues as they arise. Troubleshoot issues and outages Create solutions (tools/scripts) for recurring problems. Monitor the system daily and respond immediately to security or usability concerns Upgrade systems and processes as required for enhanced functionality and security issue resolution Must Have: Good understanding of Linux/Unix Operating System and Networking Technologies Experience with Linux based systems administration skills Knowledge of system security (e.g. intrusion detection systems) and data backup/recovery Experience working with Nagios or similar central monitoring system Experience working with ELK or similar central logging system Strong Troubleshooting and problem solving skills Good to have: Working knowledge of either Chef, Puppet or Ansible Experience with Docker and Kubernetes Expertise in at least one scripting language (Python/Perl/PHP or Ruby) Hands on experience of automating repetitive tasks About Promobi Technologies: ProMobi Technologies is based in Pune, Maharashtra. We are a Product and Services company providing top-notch Mobile and Web Application services to our clients and customers across the globe. We are one of the emerging new startups in India who are providing solutions for Enterprise Mobility Management (EMM). The role is to be part of our “core team” of awesome IT Team, Mobile, Backend engineers, and Sales - Marketing team. This is a very exciting opportunity as we are a fast growing company so you’ll be able to see your hard work and ideas going live to customers quickly.
ABOUT US: Arque Capital is a FinTech startup working with AI in Finance in domains like Asset Management (Hedge Funds, ETFs and Structured Products), Robo Advisory, Bespoke Research, Alternate Brokerage, and other applications of Technology & Quantitative methods in Big Finance. PROFILE DESCRIPTION: 1. Get the "Tech" in order for the Hedge Fund - Help answer fundamentals of technology blocks to be used, choice of certain platform/tech over other, helping team visualize product with the available resources and assets 2. Build, manage, and validate a Tech Roadmap for our Products 3. Architecture Practices - At startups, the dynamics changes very fast. Making sure that best practices are defined and followed by team is very important. CTO’s may have to garbage guy and clean the code time to time. Making reviews on Code Quality is an important activity that CTO should follow. 4. Build progressive learning culture and establish predictable model of envisioning, designing and developing products 5. Product Innovation through Research and continuous improvement 6. Build out the Technological Infrastructure for the Hedge Fund 7. Hiring and building out the Technology team 8. Setting up and managing the entire IT infrastructure - Hardware as well as Cloud 9. Ensure company-wide security and IP protection REQUIREMENTS: Computer Science Engineer from Tier-I colleges only (IIT, IIIT, NIT, BITS, DHU, Anna University, MU) 5-10 years of relevant Technology experience (no infra or database persons) Expertise in Python and C++ (3+ years minimum) 2+ years experience of building and managing Big Data projects Experience with technical design & architecture (1+ years minimum) Experience with High performance computing - OPTIONAL Experience as a Tech Lead, IT Manager, Director, VP, or CTO 1+ year Experience managing Cloud computing infrastructure (Amazon AWS preferred) - OPTIONAL Ability to work in an unstructured environment Looking to work in a small, start-up type environment based out of Mumbai COMPENSATION: Co-Founder status and Equity partnership
Immediate Onsite for DevOps Engg with 2+ years of working experience in Distributed Systems Linux/ Unix Operating systems Experience in one or more scripting languages (e.g. bash, perl, python) Jenkins (including plugins) or other build automation tools, Git, Experience with one or more of the following: Artifactory, Nexus, or other artifact storage tool Experience with Maven and/or other build tools Applied experience with modern application technologies and design patterns including: Cloud infrastructure, distributed computing, horizontal scaling, and database technologies Experience tuning the performance of Java applications in virtual environments Embraces technology trends that drive intelligent automation
What is Artivatic : Artivatic is AI enterprise tech platform built on genomic science, psychology and neuroscience capabilities to automate the decision making, prediction, personalization & recommendation in real time. Artivatic is building its own propitiatory algorithms 'connected-data-genome-mapping' & 'cross-sector-connected-intelligence'. It is available in form of API, SDK & SaaS platform for enterprises & developers to build intelligent systems & solutions. Artivatic utilizes neuroscience-based proprietary artificial intelligence systems to solve business problems and speed up processes that are usually conducted manually. Job description : We are looking for someone with an experience in Scala/Java and Play framework. Restful API development experience is mandatory. Must have experience in Database management systems. Preference will be given to profiles with experience in Scala Bonus points if they have prior experience in Cassandra Database. Location: Koramangala, Bangalore Roles and Responsibilities : - Building server-side logic that powers our APIs, in effect deploying machine learning models in production system that can scale to billions of API calls - Scaling and performance tuning of database to handle billions of API calls and thousands of concurrent requests - Collaborate with data science team to build effective solutions for data collection, pre-processing and integrating machine learning into the workflow - Collaborate, provide technical guidance, and engage in design and code review for other team members. - Excellent Scala, Cassandra, architect, API, software, Python, Java programming and software design skills, including debugging, performance analysis and test design - Proficiency with at least one Scala, GoLang, Python micro-frameworks like Flask, Tornado, Play, Spring etc. with experience in building REST APIs - Experience or understanding in building web crawlers, data fetching bots etc. - Experience with design and optimisation of Neo4j, Cassandra, NoSQL databases, PostgreSQL, Redis, Elastic Search - Familiarity with one of the cloud service providers, AWS or Google Compute Engine - Computer Science degree with 5+ years of backend programming experience
If you love puppet, you have come to the right place. We are automating from things from scratch to fully grown up production servers. Interested. Apply here
Hi , We are Civil Maps . Civil Maps, provider of 3D mapping technology for fully autonomous vehicles, has raised a $6.6 million seed funding round from Ford Motor Company, Motus Ventures, Wicklow Capital, StartX Stanford and Yahoo cofounder Jerry Yang’s AME Cloud Ventures. Civil Maps’ mission is to make it possible for fully autonomous vehicles (SAE Levels 4-5) to drive anywhere smoothly and safely. Through artificial intelligence and vehicle-based local processing, Civil Maps converts sensor data into meaningful map information built specifically to direct fully autonomous vehicles. The company will use the seed investment to accelerate product development and deployment with a number of leading automotive companies and technology partners. Civil Maps’ artificial intelligence software aggregates raw 3D data from LiDAR (high-resolution laser imaging), camera and other sensors onboard autonomous vehicles and organizes the information into machine-readable maps. The information is vastly more actionable information than today’s mapping systems and requires a fraction of the data storage and transmission for existing technologies. Thanks to this light data footprint, Civil Maps’ spatial information is far less costly to transmit over cellular networks, enabling the company to more easily crowdsource, update and share road data in real time—a major improvement over the lengthy processes that require human annotation in current use. As a result, the company can quickly generate and maintain maps that give fully autonomous vehicles the comprehensive and precise knowledge to autonomous operate safely and smoothly on all roads. About the Opportunity : This is an opportunity to work on some of the most exciting problems in the autonomous vehicle industry. Civil Maps is looking for a Devops engineer to grow its overall infrastructure. You will be interacting with a completely custom-made distributed computing framework, handling terabytes and petabytes of machine vision data, and millions of requests from robotic platforms. In our interdisciplinary teams, you will get acquainted with the many facets of 3D map creation: data collection, geospatial registration, feature extraction, auditing. Come work on novel challenges, learn and grow while helping Civil Maps achieve on-demand perception of the world around us. We are hiring the professionals for the below skills: Work Location : Hyderabad Job Description for DevOps Enginner: Experience : 7-10 yrs Responsibilities : Deliver next generation distributed applications using a variety of tools to include Kubernetes, Ansible, RHEL, Openshift, Atomic, Mesos, Openstack and Docker. Build solutions and provide a framework for Microservices and DevOps. Bring knowledge and help with continuous integration, deployment and security. Automating the Build & Deployment environment with relative tools. Troubleshooting Build & Deployment issues. Experience with the following : Deploy, manage Continuous Integration tools (Jenkins, Travis etc..). Cloud-based deployments (AWS, Azure, OpenStack). Linux administration, usage, and bash scripting (Bash, Batch, Powershell) Deploying MEAN stack applications. Security tools like Fortify, Nessus. Network architecture (routing, load balancing, firewalls, VLANs). Virtualization (VMWare, XEN, KVM, Openstack). Container technologies (Mesos, Docker, Rocket, etc.). Experience in setting up backup and restore mechanism for the DB and other relevant systems.
Technically Hands-on, prior experience with scalable Architecture -Bring 4-8 years of software engineering and product delivery experience, with strong background in algorithms -Excellent command over Data Structures and Algorithms -Exceptional coding skills in an Object Oriented programming language (Java preferred) -Strong problem solving and analytical skills -Experience with web technologies, PHP/ Java, Python, Linux, Apache, MySQL, solr, memcache, redis -Experience in architecting & building real-time, large scale e-commerce applications - Experience with high performance websites catering to millions of daily traffic is a plus