Cutshort logo
Teradata logo
Azure Specialist
Azure Specialist
Teradata's logo

Azure Specialist

Anand Jain's profile picture
Posted by Anand Jain
2 - 6 yrs
₹6L - ₹11L / yr
Hyderabad
Skills
Azure IaaS
PaaS
Azure deployment automation
Databases
Storage & Networking
Windows Azure
Experience required: 1. Experience in migrating/deploying 3rd party applications to Azure (specially Linux based applications) 2. Experience in PowerShell 3. Very good knowledge Azure storage solutions 4. Linux image building/conversion for Azure 5. Working experience in Azure IaaS, PaaS, storage, network and database 6. Experience in Azure deployment automation and DevOps (preferred) General Cloud knowledge to look for 1. Good experience database as a service in cloud 2. Very Good knowledge of Cloud storage. 3. Good understanding of HA/DR Setups in Cloud 4. Must have knowledge of Virtual machines (IAAS) and Cloud services (PAAS). 5. Understanding of Virtualization concepts and virtual system administration 6. Experience with Hyper-V configuration and administration Qualifications · Education: Bachelor's in computer science, Master's degree (M. Tech, MS) · Experience: >2 years, Hands on development experience in Azure components · Location: Hyderabad
Read more
Users love Cutshort
Read about what our users have to say about finding their next opportunity on Cutshort.
Shubham Vishwakarma's profile image

Shubham Vishwakarma

Full Stack Developer - Averlon
I had an amazing experience. It was a delight getting interviewed via Cutshort. The entire end to end process was amazing. I would like to mention Reshika, she was just amazing wrt guiding me through the process. Thank you team.
Companies hiring on Cutshort
companies logos

About Teradata

Founded :
1979
Type :
Products & Services
Size :
100-1000
Stage :
Bootstrapped

About

Teradata is a leading provider of Pervasive Data Intelligence solutions that enable businesses to get real-time, intelligent answers by leveraging 100% of their relevant data, regardless of scale or volume of query. Teradata offers industry-leading products, expertise, and services that help businesses solve complex data challenges and drive innovation. Teradata operates in various industries, including finance, healthcare, retail, telecommunications, and more.
Read more

Connect with the team

Profile picture
Anand Jain

Company social profiles

blogtwitterfacebook

Similar jobs

Fountane inc
HR Fountane
Posted by HR Fountane
Remote only
5 - 8 yrs
₹18L - ₹28L / yr
data architect
Large Language Models (LLM)
ETL
Windows Azure
Google Cloud Platform (GCP)
+5 more

Position Overview: We are looking for an experienced and highly skilled Data Architect to join our team and help design, implement, and optimize data systems that support high-end analytical solutions for our clients. As a customer-centric Data Architect, you will work closely with clients to understand their business needs and translate them into robust, scalable, and efficient technical solutions. You will be responsible for end-to-end data modelling, integration workflows, and data transformation processes while ensuring security, privacy, and compliance.In this role, you will also leverage the latest advancements in artificial intelligence, machine learning, and large language models (LLMs) to deliver high-impact solutions that drive business success. The ideal candidate will have a deep understanding of data infrastructure, optimization techniques, and cost-effective data management


Key Responsibilities:


• Customer Collaboration:

– Partner with clients to gather and understand their business

requirements, translating them into actionable technical specifications.

– Act as the primary technical consultant to guide clients through data challenges and deliver tailored solutions that drive value.


•Data Modeling & Integration:

– Design and implement scalable, efficient, and optimized data models to support business operations and analytical needs.

– Develop and maintain data integration workflows to seamlessly extract, transform, and load (ETL) data from various sources into data repositories.

– Ensure smooth integration between multiple data sources and platforms, including cloud and on-premise systems


Data Processing & Optimization:

– Develop, optimize, and manage data processing pipelines to enable real-time and batch data processing at scale.

– Continuously evaluate and improve data processing performance, optimizing for throughput while minimizing infrastructure costs.


• Data Governance & Security:

–Implement and enforce data governance policies and best practices, ensuring data security, privacy, and compliance with relevant industry regulations (e.g., GDPR, HIPAA).

–Collaborate with security teams to safeguard sensitive data and maintain privacy controls across data environments.


• Cross-Functional Collaboration:

– Work closely with data engineers, data scientists, and business

analysts to ensure that the data architecture aligns with organizational objectives and delivers actionable insights.

– Foster collaboration across teams to streamline data workflows and optimize solution delivery.


• Leveraging Advanced Technologies:

– Utilize AI, machine learning models, and large language models (LLMs) to automate processes, accelerate delivery, and provide

smart, data-driven solutions to business challenges.

– Identify opportunities to apply cutting-edge technologies to improve the efficiency, speed, and quality of data processing and analytics.


• Cost Optimization:

–Proactively manage infrastructure and cloud resources to optimize throughput while minimizing operational costs.

–Make data-driven recommendations to reduce infrastructure overhead and increase efficiency without sacrificing performance.


Qualifications:


• Experience:

– Proven experience (5+ years) as a Data Architect or similar role, designing and implementing data solutions at scale.

– Strong expertise in data modelling, data integration (ETL), and data transformation processes.

– Experience with cloud platforms (AWS, Azure, Google Cloud) and big data technologies (e.g., Hadoop, Spark).


• Technical Skills:

– Advanced proficiency in SQL, data modelling tools (e.g., Erwin,PowerDesigner), and data integration frameworks (e.g., Apache

NiFi, Talend).

– Strong understanding of data security protocols, privacy regulations, and compliance requirements.

– Experience with data storage solutions (e.g., data lakes, data warehouses, NoSQL, relational databases).


• AI & Machine Learning Exposure:

– Familiarity with leveraging AI and machine learning technologies (e.g., TensorFlow, PyTorch, scikit-learn) to optimize data processing and analytical tasks.

–Ability to apply advanced algorithms and automation techniques to improve business processes.


• Soft Skills:

– Excellent communication skills to collaborate with clients, stakeholders, and cross-functional teams.

– Strong problem-solving ability with a customer-centric approach to solution design.

– Ability to translate complex technical concepts into clear, understandable terms for non-technical audiences.


• Education:

– Bachelor’s or Master’s degree in Computer Science, Information Systems, Data Science, or a related field (or equivalent practical experience).


LIFE AT FOUNTANE:

  • Fountane offers an environment where all members are supported, challenged, recognized & given opportunities to grow to their fullest potential.
  • Competitive pay
  • Health insurance for spouses, kids, and parents.
  • PF/ESI or equivalent
  • Individual/team bonuses
  • Employee stock ownership plan
  • Fun/challenging variety of projects/industries
  • Flexible workplace policy - remote/physical
  • Flat organization - no micromanagement
  • Individual contribution - set your deadlines
  • Above all - culture that helps you grow exponentially!


A LITTLE BIT ABOUT THE COMPANY:

Established in 2017, Fountane Inc is a Ventures Lab incubating and investing in new competitive technology businesses from scratch. Thus far, we’ve created half a dozen multi-million valuation companies in the US and a handful of sister ventures for large corporations, including Target, US Ventures, and Imprint Engine.

We’re a team of 120+ strong from around the world that are radically open-minded and believes in excellence, respecting one another, and pushing our boundaries to the furthest it's ever been.

Read more
TVARIT GmbH
at TVARIT GmbH
2 candid answers
Shivani Kawade
Posted by Shivani Kawade
Remote, Pune
2 - 4 yrs
₹8L - ₹20L / yr
skill iconPython
PySpark
ETL
databricks
Azure
+6 more

TVARIT GmbH develops and delivers solutions in the field of artificial intelligence (AI) for the Manufacturing, automotive, and process industries. With its software products, TVARIT makes it possible for its customers to make intelligent and well-founded decisions, e.g., in forward-looking Maintenance, increasing the OEE and predictive quality. We have renowned reference customers, competent technology, a good research team from renowned Universities, and the award of a renowned AI prize (e.g., EU Horizon 2020) which makes Tvarit one of the most innovative AI companies in Germany and Europe. 

 

 

We are looking for a self-motivated person with a positive "can-do" attitude and excellent oral and written communication skills in English. 

 

 

We are seeking a skilled and motivated Data Engineer from the manufacturing Industry with over two years of experience to join our team. As a data engineer, you will be responsible for designing, building, and maintaining the infrastructure required for the collection, storage, processing, and analysis of large and complex data sets. The ideal candidate will have a strong foundation in ETL pipelines and Python, with additional experience in Azure and Terraform being a plus. This role requires a proactive individual who can contribute to our data infrastructure and support our analytics and data science initiatives. 

 

 

Skills Required 

  • Experience in the manufacturing industry (metal industry is a plus)  
  • 2+ years of experience as a Data Engineer 
  • Experience in data cleaning & structuring and data manipulation 
  • ETL Pipelines: Proven experience in designing, building, and maintaining ETL pipelines. 
  • Python: Strong proficiency in Python programming for data manipulation, transformation, and automation. 
  • Experience in SQL and data structures  
  • Knowledge in big data technologies such as Spark, Flink, Hadoop, Apache and NoSQL databases. 
  • Knowledge of cloud technologies (at least one) such as AWS, Azure, and Google Cloud Platform. 
  • Proficient in data management and data governance  
  • Strong analytical and problem-solving skills. 
  • Excellent communication and teamwork abilities. 

 


Nice To Have 

  • Azure: Experience with Azure data services (e.g., Azure Data Factory, Azure Databricks, Azure SQL Database). 
  • Terraform: Knowledge of Terraform for infrastructure as code (IaC) to manage cloud. 


Read more
Bengaluru (Bangalore)
2 - 6 yrs
₹8L - ₹20L / yr
skill iconDocker
skill iconKubernetes
DevOps
skill iconAmazon Web Services (AWS)
Windows Azure
+3 more
Job Description:

About BootLabs

https://www.google.com/url?q=https://www.bootlabs.in/&;sa=D&source=calendar&ust=1667803146567128&usg=AOvVaw1r5g0R_vYM07k6qpoNvvh6" target="_blank">https://www.bootlabs.in/

-We are a Boutique Tech Consulting partner, specializing in Cloud Native Solutions. 
-We are obsessed with anything “CLOUD”. Our goal is to seamlessly automate the development lifecycle, and modernize infrastructure and its associated applications.
-With a product mindset, we enable start-ups and enterprises on the cloud
transformation, cloud migration, end-to-end automation and managed cloud services. 
-We are eager to research, discover, automate, adapt, empower and deliver quality solutions on time.
-We are passionate about customer success. With the right blend of experience and exuberant youth in our in-house team, we have significantly impacted customers.




Technical Skills:


Expertise in any one hyper scaler (AWS/AZURE/GCP), including basic services like networking,
data and workload management.
  • AWS 

              Networking: VPC, VPC Peering, Transit Gateway, Route Tables, SecuritGroups, etc.
              Data: RDS, DynamoDB, Elastic Search
Workload: EC2, EKS, Lambda, etc.
  •  Azure
                Networking: VNET, VNET Peering,
               Data: Azure MySQL, Azure MSSQL, etc.
               Workload: AKS, Virtual Machines, Azure Functions
  • GCP
               Networking: VPC, VPC Peering, Firewall, Flowlogs, Routes, Static and External IP Addresses
                Data: Cloud Storage, DataFlow, Cloud SQL, Firestore, BigTable, BigQuery
               Workload: GKE, Instances, App Engine, Batch, etc.

Experience in any one of the CI/CD tools (Gitlab/Github/Jenkins) including runner setup,
templating and configuration.
Kubernetes experience or Ansible Experience (EKS/AKS/GKE), basics like pod, deployment,
networking, service mesh. Used any package manager like helm.
Scripting experience (Bash/python), automation in pipelines when required, system service.
Infrastructure automation (Terraform/pulumi/cloud formation), write modules, setup pipeline and version the code.

Optional:

Experience in any programming language is not required but is appreciated.
Good experience in GIT, SVN or any other code management tool is required.
DevSecops tools like (Qualys/SonarQube/BlackDuck) for security scanning of artifacts, infrastructure and code.
Observability tools (Opensource: Prometheus, Elasticsearch, Open Telemetry; Paid: Datadog,
24/7, etc)
Read more
DigilyticsAI
at DigilyticsAI
1 video
2 recruiters
Komal Rathore
Posted by Komal Rathore
Gurugram
3 - 9 yrs
₹6L - ₹15L / yr
Linux/Unix
skill iconDocker
skill iconKubernetes
DevOps
skill iconAmazon Web Services (AWS)
+2 more

About Us

At Digilytics™, we build and deliver easy to use AI products to the secured lending and consumer industry sectors. In an ever-crowded world of clever technology solutions looking for a problem to solve, our solutions start with a keen understanding of what creates and what destroys value in our clients’ business.

Founded by Arindom Basu (Founding member of Infosys Consulting), the leadership of Digilytics™ is deeply rooted in leveraging disruptive technology to drive profitable business growth. With over 50 years of combined experience in technology-enabled change, the Digilytics™ leadership is focused on building a values-first firm that will stand the test of time.

We are currently focused on developing a product, Revel FS, to revolutionise loan origination for mortgages and secured lending. We are also developing a second product, Revel CI, focused on improving trade (secondary) sales to consumer industry clients like auto and FMCG players.

The leadership strongly believes in the ethos of enabling intelligence across the organization. Digiliytics AI is headquartered in London, with a presence across India.

Website: http://www.digilytics.ai">www.digilytics.ai

 

 Responsibilities

  • Experience with Azure services (Virtual machines, Containers, Databases, Security/Firewall, Function Apps etc)
  • Hands-on experience on Kubernetes/Docker/helm.
  • Deployment of java Builds & administration/configuration of Nginx/Reverse Proxy, Load balancer, Ms-SQL, Github, Disaster Recovery,
  • Linux – Must have basic knowledge- User creation/deletion, ACL, LVM etc.
  • CI/CD - Azure DevOps or any other automation tool like Terraform, Jenkins etc.
  • Experience with SharePoint and O365 administration
  • Azure/Kubernetes certification will be preferred.
  • Microsoft Partnership experience is good to have. 
  • Excellent understanding of required technologies
  • Good interpersonal skills and the ability to communicate ideas clearly at all levels
  • Ability to work in unfamiliar business areas and to use your skills to create solutions
  • Ability to both work in and lead a team and to deliver and accept peer review
  • Flexible approach to working environment and hours to meet the needs of the business and clients

 

Must Haves:

  • Hands-on experience on Kubernetes/Docker/helm.
  • Experience on Azure/Aws or any other cloud provider.
  • Linux & CI/CD tools knowledge.

 

Experience & Education:

  • A start up mindset with proven experience working in both smaller and larger organizations having multicultural exposure
  • Between 4-9 years of experience working closely with the relevant technologies, and developing world-class software and solutions
  • Domain and industry experience by serving customers in one or more of these industries - Financial Services, Professional Services, other Retail Consumer Services
  • A bachelor's degree, or equivalent, in Software Engineering and Computer Science

 

Read more
IT Infrastructure Service Co
IT Infrastructure Service Co
Agency job
via Unnati by Veena Salian
Delhi, Gurugram, Noida, Ghaziabad, Faridabad
5 - 10 yrs
₹18L - ₹22L / yr
DevOps
Linux/Unix
Shell Scripting
Google Cloud Platform (GCP)
Databases
+4 more
This is a great opportunity for young and ambitious people looking to make a career in IT infrastructure solution sales!!

Our Client is an IT infrastructure services company, focused and specialized in delivering solutions and services on Microsoft products and technologies. They are a Microsoft partner and cloud solution provider. Our Client's objective is to help small, mid-sized as well as global enterprises to transform their business by using innovation in IT, adapting to the latest technologies and using IT as an enabler for business to meet business goals and continuous growth.

With focused and experienced management and a strong team of IT Infrastructure professionals, they are adding value by making IT Infrastructure a robust, agile, secure and cost-effective service to the business. As an independent IT Infrastructure company, they provide their clients with unbiased advice on how to successfully implement and manage technology to complement their business requirements.
 
As a Systems Engineer/DevOps Engineer, you will work in a proactive environment, ensuring availability and reliability of existing systems, deploying changes, and acting as a technical point of escalation for user incidents.
 
What you will do:
  • Providing on-call support within a high availability production environment
  • Logging issues
  • Providing Complex problem analysis and resolution for technical and application issues
  • Supporting and collaborating with team members
  • Running system updates
  • Monitoring and responding to system alerts
  • Developing and running system health checks
  • Applying industry standard practices across the technology estate
  • Performing system reviews
  • Reviewing and maintaining infrastructure configuration
  • Diagnosing performance issues and network bottlenecks
  • Collaborating within geographically distributed teams
  • Supporting software development infrastructure by continuous integration and delivery standards
  • Working closely with developers and QA teams as part of a customer support centre
  • Projecting delivery work, either individually or in conjunction with other teams, external suppliers or contractors
  • Ensuring maintenance of the technical environments to meet current standards
  • Ensuring compliance with appropriate industry and security regulations
  • Providing support to Development and Customer Support teams
  • Managing the hosted infrastructure through vendor engagement
  • Managing 3rd party software licensing ensuring compliance
  • Delivering new technologies as agreed by the business

 


Candidate Profile:

What you need to have:

  • Experience working within a technical operations environment relevant to associated skills stated.
  • Be proficient in:
  1. Linux, zsh/ bash/ similar
  2. ssh, tmux/ screen/ similar
  3. vim/ emacs/ similar
  4. Computer networking
  • Have a reasonable working knowledge of:
  1. Cloud infrastructure, Preferably GCP
  2. One or more programming/ scripting languages
  3. Git
  4. Docker
  5. Web services and web servers
  6. Databases, relational and NoSQL
  • Some familiarity with:
  1. Puppet, ansible
  2. Terraform
  3. GitHub, CircleCI , Kubernetes
  4. Scripting language- Shell
  5. Databases: Cassandra, Postgres, MySQL or CloudSQL
  6. Agile working practices including scrum and Kanban
  7. Private & public cloud hosting environments
  • Strong technology interests with a positive ‘can do’ attitude
  • Be flexible and adaptable to changing priorities
  • Be good at planning and organising their own time and able to meet targets and deadlines without supervision
  • Excellent written and verbal communication skills.
  • Approachable with both colleagues and team members
  • Be resourceful and practical with an ability to respond positively and quickly to technical and business challenges
  • Be persuasive, articulate and influential, but down to earth and friendly with own team and colleagues
  • Have an ability to establish relationships quickly and to work effectively either as part of a team or singularly
  • Be customer focused with both internal and external customers
  • Be capable of remaining calm under pressure
  • Technically minded with good problem resolution skills and systematic manner
  • Excellent documentation skills
  • Prepared to participate in out of hours support rota

 

Read more
Loonycorn
at Loonycorn
1 recruiter
Vitthal Srinivasan
Posted by Vitthal Srinivasan
Bengaluru (Bangalore)
0 - 7 yrs
₹3L - ₹5L / yr
Data Structures
Big Data
Cloud Computing
skill iconKubernetes
skill iconJava
+1 more
About Loonycorn:

Founded by Google, Stanford, and Columbia alumni, Loonycorn is a leading studio for e-learning content on machine learning, cloud computing, blockchain, and other emerging technologies.

About the Role:

We are looking for folks to build software projects which will be part of technical content similar to what you'd find at the links below:

https://www.pluralsight.com/search?q=janani+ravi
https://www.pluralsight.com/search?q=vitthal+srinivasan
https://www.udemy.com/u/janani-ravi-2/

This involves:
- learning a new technology from scratch
- coding real-world projects which use that technology (you will be a software engineer writing code, not making slides or talking about code)

What is important to us:
- Grit - Perseverance in working on hard problems. Technical video-making is difficult and detail-oriented (that's why it is a highly profitable business)
- Craftsmanship - Our video-making is quite artisanal - lots of hard work and small details. There are many excellent roles where doing smart 80-20 trade-offs is the way to succeed - this is not one of them.
- Clarity - Talking and thinking in direct, clear ways is super-important in what we do. Folks who use a lot of jargon or cliches, or work on writing code without understanding what goes on underneath will not be a fit
- Creativity - Analogies, technical metaphors, and other artistic elements are an important part of what we do.

What is not all that important to us:
- Your school or labels: Perfectly fine whatever college or company you are applying from
- English vocabulary or pronunciation: You don't need to 'talk well' or be flashy to build good projects
Read more
APT Portfolio
at APT Portfolio
1 recruiter
Ankita  Pachauri
Posted by Ankita Pachauri
Delhi, Gurugram, Bengaluru (Bangalore)
10 - 15 yrs
₹50L - ₹70L / yr
DevOps
skill iconKubernetes
skill iconDocker
skill iconAmazon Web Services (AWS)
Windows Azure
+13 more

A.P.T Portfolio, a high frequency trading firm that specialises in Quantitative Trading & Investment Strategies.Founded in November 2009, it has been a major liquidity provider in global Stock markets. 


As a manager, you would be incharge of managing the devops team and your remit shall include the following

  • Private Cloud - Design & maintain a high performance and reliable network architecture to support  HPC applications
  • Scheduling Tool - Implement and maintain a HPC scheduling technology like Kubernetes, Hadoop YARN  Mesos, HTCondor or Nomad for processing & scheduling analytical jobs. Implement controls which allow analytical jobs to seamlessly utilize ideal capacity on the private cloud. 
  • Security - Implementing best security practices and implementing data isolation policy between different divisions internally. 
  • Capacity Sizing - Monitor private cloud usage and share details with different teams. Plan capacity enhancements on a quarterly basis. 
  • Storage solution - Optimize storage solutions like NetApp, EMC, Quobyte for analytical jobs. Monitor their performance on a daily basis to identify issues early.
  • NFS - Implement and optimize latest version of NFS for our use case. 
  • Public Cloud - Drive AWS/Google-Cloud utilization in the firm for increasing efficiency, improving collaboration and for reducing cost. Maintain the environment for our existing use cases. Further explore potential areas of using public cloud within the firm. 
  • BackUps  - Identify and automate  back up of all crucial data/binary/code etc in a secured manner at such duration warranted by the use case. Ensure that recovery from back-up is tested and seamless. 
  •  Access Control  - Maintain password less access control and improve security over time. Minimize failures for automated job due to unsuccessful logins. 
  •  Operating System  -Plan, test and roll out new operating system for all production, simulation and desktop environments. Work closely with developers to highlight new performance enhancements capabilities of new versions. 
  •  Configuration management  -Work closely with DevOps/ development team to freeze configurations/playbook for various teams & internal applications. Deploy and maintain standard tools such as Ansible, Puppet, chef etc for the same. 
  •  Data Storage & Security Planning  - Maintain a tight control of root access on various devices. Ensure root access is rolled back as soon the desired objective is achieved.
  • Audit access logs on devices. Use third party tools to put in a monitoring mechanism for early detection of any suspicious activity. 
  • Maintaining all third party tools used for development and collaboration - This shall include maintaining a fault tolerant   environment for GIT/Perforce, productivity tools such as Slack/Microsoft team, build tools like Jenkins/Bamboo etc


Qualifications 

  • Bachelors or Masters Level Degree, preferably in CSE/IT
  • 10+ years of relevant experience in sys-admin function
  • Must have strong knowledge of IT Infrastructure, Linux, Networking and grid.
  • Must have strong grasp of automation & Data management tools.
  • Efficient in scripting languages and python


Desirables

  • Professional attitude, co-operative and mature approach to work, must be focused, structured and well considered, troubleshooting skills.
  •  Exhibit a high level of individual initiative and ownership, effectively collaborate with other team members.

 

APT Portfolio is an equal opportunity employer

Read more
Google of biomedicine
Google of biomedicine
Agency job
via Pinnacle Consultz by SPP Praveen
Bengaluru (Bangalore), Gurugram
8 - 16 yrs
₹50L - ₹70L / yr
skill iconC++
skill iconPython
skill iconGo Programming (Golang)
DevOps
Databases
+1 more
Staff Software Engineer

We are looking for Principal Engineers, who are strong individual contributors with
expertise and passion in solving difficult problems in many areas.
Your day at nference,
• Acting as an entrepreneur - taking ownership the problem statement end-to-end
• Delivering direct value to the customer - and not just stop with delivery
• Estimate, plan, divide and conquer customer problem statements - through sturdily
developed & performant technical solutions
• Handle multiple competing priorities and ambiguity - all in a fast-paced, high growth
environment Qualities Which We Look For In The Ideal Candidate
• 6-8 years of experience in building High Performance Distributed Systems
• Proven track record in building backend systems from scratch
• Excellent coding skills (preferably any two of C/C++/Python and Go)
• Good depth in Algorithms & Data Structures
• Good understanding of OS level concepts
• Experience working on DevOps tools for deployment, monitoring etc. like Ansible, ELK
Prometheus etc
• Wide knowledge of different Technologies like Databases, Messaging Systems etc
• Experience building complex technical solutions - highly scalable service-oriented
architectures, distributed cloud-based systems - which power our products

Benefits:

• Be a part of “Google of biomedicine” as recognized by the Washington Post
• Work with some of the brilliant minds of the world solving exciting real-world
problems through Artificial Intelligence, Machine Learning, analytics and insights
through triangulating unstructured and structured information from the biomedical
literature as well as from large-scale molecular and real-world datasets.
• Our benefits package includes the best of what leading organizations provide, such as
stock options, paid time off, healthcare insurance, gym/broadband reimbursement
Read more
ITS
at ITS
1 recruiter
Shruti Pandit
Posted by Shruti Pandit
Mumbai
3 - 7 yrs
₹10L - ₹15L / yr
skill icon.NET
skill iconC#
skill iconJavascript
skill iconNodeJS (Node.js)
NOSQL Databases
+2 more

Requisite Skills and Experience                                                                                                          

  • 5+ years of object-oriented software development of web applications
  • 5+ years of C# in a .NET environment
  • Preferable Experience in Cloud Technologies (AWS, AZURE)
  • Experience with MS SQL stored procedures, database design and optimization
  • Experience developing and utilizing web services
  • Experience with high volume consumer oriented web sites
  • Experience working in an Agile Scrum environment
  • Experience with formal test procedures, release management
  • Preferable Travel industry experience preferred, especially programming associated with GDS’s - SABRE, Farelogix, and Travelport
  • Ability to distill complex business needs into functional requirements
  • Excellent follow-up, verbal and written communication skills
  • Excellent problem solving skills and an ability to make good decisions
  • Team player with a professional attitude
  • Ability to manage multiple tasks and maintain tight schedules
  • Ability to work effectively with a geographically dispersed team in India

 

Strategic and Development Responsibilities

  • Ability to "think outside the box"
  • Ability to provide accurate and timely project estimates.

 

 

Pluses:

  • Experience with Angular JS frameworks and Mobile Development.
  • Experience with developing machine learning algorithms.
  • Experience on projects that required localization/globalization
  • Experience with MS SQL Reporting Services
  • Knowledge of web farm configuration and best practices

 

Software Tools/Languages

Required:        MS Visual Studio, MS SQL Server Management Studio, IIS, AWS

C#, Web Services, MS SQL (stored procedures), Jenkins, DevOps Tools.

 

Education

Bachelor of Science in CS, MIS, or equivalent
Read more
Mobile Programming LLC
at Mobile Programming LLC
1 video
34 recruiters
Nivedita Kumari
Posted by Nivedita Kumari
Gurugram, Bengaluru (Bangalore), Chennai, Pune, Mohali, Panchkula, Dehradun
4 - 10 yrs
₹7L - ₹14L / yr
skill iconNodeJS (Node.js)
skill iconDocker
skill iconKubernetes
skill iconJavascript
Databases
Mandatory Skills:

- Backend server development & support with Node, JavaScript, JSON, REST, NoSQL, Cloud Native technologies like Docker & Registry, Kubernetes & Helm

- Skilled in data structures, algorithms, modularization, OOP, microservices, and design patterns

- Skilled in coding best practices using containers, packaging (npm, yarn), agility (with Git, Jira), unit testing (JEST), CI/CD (Jenkins), debugging, and ensuring high productivity & quality

- Exposure to security (OIDC/JWT, RBAC, monitoring, auditing)

- Good with learning, problem solving & innovation

- Good written & verbal communications skill; team player with good logical thinking, EQ, and resilience
Read more
Why apply to jobs via Cutshort
people_solving_puzzle
Personalized job matches
Stop wasting time. Get matched with jobs that meet your skills, aspirations and preferences.
people_verifying_people
Verified hiring teams
See actual hiring teams, find common social connections or connect with them directly. No 3rd party agencies here.
ai_chip
Move faster with AI
We use AI to get you faster responses, recommendations and unmatched user experience.
21,01,133
Matches delivered
37,12,187
Network size
15,000
Companies hiring
Did not find a job you were looking for?
icon
Search for relevant jobs from 10000+ companies such as Google, Amazon & Uber actively hiring on Cutshort.
companies logo
companies logo
companies logo
companies logo
companies logo
Get to hear about interesting companies hiring right now
Company logo
Company logo
Company logo
Company logo
Company logo
Linkedin iconFollow Cutshort
Users love Cutshort
Read about what our users have to say about finding their next opportunity on Cutshort.
Shubham Vishwakarma's profile image

Shubham Vishwakarma

Full Stack Developer - Averlon
I had an amazing experience. It was a delight getting interviewed via Cutshort. The entire end to end process was amazing. I would like to mention Reshika, she was just amazing wrt guiding me through the process. Thank you team.
Companies hiring on Cutshort
companies logos