ansible-jobs-in-hyderabad, Hyderabad

108 Ansible Jobs in Hyderabad

Toggle to save search
posted 1 week ago
experience10 to 20 Yrs
Salary6 - 14 LPA
location
Hyderabad
skills
  • gcp
  • ansible
  • fusion
  • data engineering
  • data
  • sql
  • bigquery
Job Description
Job ID: ITC/G/20251029/22301 Role: GCP Data Engineer Location: Hyderabad Status: Open   Role Overview This role focuses on designing, developing, testing, and deploying data models and transformations on Google Cloud Platform, primarily using BigQuery and Data Fusion. The engineer will be responsible for building scalable data pipelines, optimizing data structures for analytics and visualization, and supporting CI/CD automation. Key Responsibilities Design and develop GCP-based data models and ETL/ELT pipelines. Build and deploy transformations in BigQuery and Data Fusion. Optimize data views and datasets for reporting and visualization use cases. Manage and maintain code repositories, artifacts, and deployment workflows. Implement CI/CD pipelines using tools like Git, Jenkins, Ansible, etc. Perform data preparation, wrangling, and quality checks. Work with relational, dimensional, and preferably Data Vault data models. Collaborate in a hybrid work environment with cross-functional teams. Required Skills Google Cloud Platform (GCP) BigQuery Data Fusion SQL / T-SQL Ansible CI/CD Tools (Git, Jenkins) Data Modeling (Relational, Dimensional; Data Vault preferred) Data Wrangling & Preparation Qualification B.Tech (Computer Science, IT, or related field preferred) Experience 6-15 years in Data Engineering / Cloud Data Platforms Salary Range 10,00,000 - 27,00,000 (based on experience and skill level)
INTERVIEW ASSURED IN 15 MINS

Top Companies are Hiring in Your City

For Multiple Roles

Jio Platforms Ltd
Jio Platforms Ltdslide-preview-Genpact
posted 2 months ago

Senior Automation Engineer

LTIMindtree Limited
LTIMindtree Limited
experience7 to 12 Yrs
location
Hyderabad, Bangalore+5

Bangalore, Noida, Chennai, Kolkata, Pune, Coimbatore

skills
  • ansible
  • cloud
  • automation
  • engineer
Job Description
We are looking for an experienced Network Automation Engineer to design, implement, and optimize automation solutions for our Private Cloud datacenter network, which underpins large-scale AI/ML GPU and TPU workloads. This role focuses on automating configuration, provisioning, and monitoring of high-performance networking devices to ensure low latency, high throughput, and reliability in a mission-critical environment. This role involves automating network device management as well as OS-level network configurations on servers. Expertise in Ansible and Python is essential, and experience with GoLang is a strong plus.  Key Responsibilities: Develop and maintain network automation frameworks for large-scale datacenter environments supporting AI/ML workloads. Build Ansible playbooks, roles, and modules to automate device configurations, software upgrades, and compliance checks across multi-vendor environments. Design and implement Python-based automation scripts and tools to integrate with APIs, orchestration platforms, and monitoring systems. Automate OS core networking configurations on servers (Linux / Windows / Hypervisor) including bonding, VLANs, routing tables, kernel network parameters, MTU tuning, and NIC performance optimization.
INTERVIEW ASSURED IN 15 MINS
posted 3 weeks ago
experience3 to 7 Yrs
location
Hyderabad, Telangana
skills
  • software
  • virtualization
  • hardware
  • databases
  • storage engineering
  • integration
  • automation
  • scaling
  • scripting languages
  • Python
  • software development
  • Ansible
  • applications
  • technical processes
  • infrastructure engineering
  • networking terminology
  • deployment practices
  • resilience
  • performance assessments
  • VMware VCF
  • Terraform
  • storage clusters
  • blueprints
  • cloud infrastructure
  • cloud technologies
Job Description
As an Infrastructure Engineer III at JPMorganChase within the Infrastructure Platforms team, you have the opportunity to leverage your strong knowledge of software, applications, and technical processes to play a pivotal role in the migration of applications from legacy infrastructure to the modern private cloud platform. Your role involves coordinating and collaborating across teams, driving infrastructure engineering technologies, and ensuring the delivery of infrastructure in accordance with business requirements. You will be responsible for analyzing complex systems, mitigating risks, implementing changes, and overseeing infrastructure upgrades to modernize the organization's technology processes. **Key Responsibilities:** - Coordinate and collaborate across teams for end-to-end migration of applications to the modern private cloud platform - Drive workstreams or projects involving infrastructure engineering technologies - Apply deep technical expertise to analyze complex systems and anticipate issues - Architect and implement changes to resolve issues and modernize technology processes - Ensure infrastructure engineering and delivery align with business requirements - Execute work according to compliance standards, risk, security, and business objectives - Monitor project progress, report status to stakeholders, and support troubleshooting during migration efforts **Qualifications Required:** - Formal training or certification in infrastructure engineering concepts and 3+ years of applied experience in virtualization - Deep knowledge in areas such as hardware, networking, databases, storage engineering, deployment practices, automation, and performance assessments - Bachelor's degree in IT, Computer Science, or related field - Experience with VMware VCF, Terraform, storage clusters, and blueprints - Ability to work in a fast-paced, collaborative environment - Basic working knowledge of scripting languages (e.g., Python) - Experience in software development (Python and Ansible) with 3+ years of experience - Knowledge of cloud infrastructure and multiple cloud technologies - Ability to drive projects towards completion and effectively communicate with senior leaders The company prefers candidates who have demonstrated experience in leading infrastructure migration projects, collaborating with cross-functional teams to resolve technical challenges, implementing automation solutions, and effectively communicating project updates and technical recommendations to senior leadership.,
ACTIVELY HIRING
question

Are these jobs relevant for you?

posted 2 weeks ago
experience3 to 7 Yrs
location
Hyderabad, All India
skills
  • automation
  • GCP
  • Azure
  • Ansible
  • Python
  • Bash
  • AIML
  • cloudnative technologies
  • Terraform
  • SRE practices
  • GitOps workflows
Job Description
Role Overview: You will be joining Deutsche Brse Group as a PaaS, AI & Automation Engineer in the Corporate IT Cloud Infrastructure Operations team. This role offers you the opportunity to shape the cloud-native platforms of the organization and work at the intersection of development and infrastructure operations. Your focus will be on automation, reliability, and modern cloud technologies to ensure performance, resilience, and security in a highly regulated financial environment. Key Responsibilities: - Design, implement, and maintain automated deployment pipelines to modernize the infrastructure landscape through automation. - Integrate AI/ML solutions to optimize ticket resolution, automate request handling, facilitate customer self-service, and enhance system observability. - Adhere to security best practices, ensure compliance with internal and external standards, and implement AI-assisted vulnerability verification, mitigation, and resolution. - Collaborate cross-functionally with engineering, security, and operations teams to integrate tools and processes effectively. - Create comprehensive documentation and knowledge base articles related to the implementation and application of solutions. Qualifications Required: - Experience integrating AI/ML, preferably into infrastructure operations. - 3+ years of experience in automation in a mission-critical environment. - Hands-on experience with GCP, Vertex AI, and familiarity with cloud-native serverless technologies. Azure experience is a plus. - Proficiency in automation platforms such as Ansible, GCP Workflows. - Solid understanding of API integrations, authentication, and event-driven architecture. - Familiarity with monitoring and observability (OTel, Prometheus integration). - Minimum 2 years of scripting and programming experience (Python, Bash, .). - Experience with Infrastructure as Code (IaC) in Terraform for declarative infrastructure provisioning. - Exposure to SRE practices and GitOps workflows. - Ability to self-motivate and possess strong analytical, communication, and problem-solving skills. - Fluent in written and spoken English. Role Overview: You will be joining Deutsche Brse Group as a PaaS, AI & Automation Engineer in the Corporate IT Cloud Infrastructure Operations team. This role offers you the opportunity to shape the cloud-native platforms of the organization and work at the intersection of development and infrastructure operations. Your focus will be on automation, reliability, and modern cloud technologies to ensure performance, resilience, and security in a highly regulated financial environment. Key Responsibilities: - Design, implement, and maintain automated deployment pipelines to modernize the infrastructure landscape through automation. - Integrate AI/ML solutions to optimize ticket resolution, automate request handling, facilitate customer self-service, and enhance system observability. - Adhere to security best practices, ensure compliance with internal and external standards, and implement AI-assisted vulnerability verification, mitigation, and resolution. - Collaborate cross-functionally with engineering, security, and operations teams to integrate tools and processes effectively. - Create comprehensive documentation and knowledge base articles related to the implementation and application of solutions. Qualifications Required: - Experience integrating AI/ML, preferably into infrastructure operations. - 3+ years of experience in automation in a mission-critical environment. - Hands-on experience with GCP, Vertex AI, and familiarity with cloud-native serverless technologies. Azure experience is a plus. - Proficiency in automation platforms such as Ansible, GCP Workflows. - Solid understanding of API integrations, authentication, and event-driven architecture. - Familiarity with monitoring and observability (OTel, Prometheus integration). - Minimum 2 years of scripting and programming experience (Python, Bash, .). - Experience with Infrastructure as Code (IaC) in Terraform for declarative infrastructure provisioning. - Exposure to SRE practices and GitOps workflows. - Ability to self-motivate and possess strong analytical, communication, and problem-solving skills. - Fluent in written and spoken English.
ACTIVELY HIRING
posted 3 weeks ago
experience3 to 7 Yrs
location
Hyderabad, Telangana
skills
  • automation
  • GCP
  • Azure
  • Ansible
  • Python
  • Bash
  • AIML
  • cloudnative technologies
  • Terraform
  • SRE practices
  • GitOps workflows
Job Description
Role Overview: As a PaaS, AI & Automation Engineer at Deutsche Brse Group, you will be part of the Corporate IT Cloud Infrastructure Operations team focused on building the next generation of cloud-native operations. Your role will involve modernizing infrastructure through automation, cloud-native technologies, and AI-driven enhancements. You will contribute to designing and implementing solutions to enhance operational efficiency and ensure performance, resilience, and security. Key Responsibilities: - Design, implement, and maintain automated deployment pipelines for infrastructure automation & deployment. - Integrate AI/ML solutions to optimize ticket resolution, automate request handling, facilitate customer self-service, and enhance system observability. - Adhere to security best practices, ensure compliance with internal and external standards, and implement AI assisted vulnerability verification, mitigation, and resolution. - Collaborate with engineering, security, and operations teams to integrate tools and processes for innovation. - Create comprehensive documentation and knowledge base articles related to the implementation and application of solutions. Qualifications Required: - Experience integrating AI/ML, preferably into infrastructure operations. - 3+ years of experience in automation in a mission-critical environment. - Hands-on experience with GCP, Vertex AI, familiarity with cloud native serverless technologies, and Azure experience is a plus. - Proficiency in automation platforms such as Ansible, GCP Workflows. - Solid understanding of API integrations, authentication, event-driven architecture, monitoring, and observability (OTel, Prometheus integration). - 2+ years of scripting and programming experience in Python, Bash, etc. - Experience with Infrastructure as Code (IaC) in Terraform for declarative infrastructure provisioning. - Exposure to SRE practices and GitOps workflows. - Self-motivated with strong analytical, communication, and problem-solving skills. - Fluent in written and spoken English.,
ACTIVELY HIRING
posted 1 week ago
experience3 to 7 Yrs
location
Hyderabad, Telangana
skills
  • Linux system administration
  • Jenkins
  • Git
  • Docker
  • Ansible
  • Kubernetes
  • DevOps tools
  • Terraform
  • CICD pipelines
  • Bitbucket pipelines
  • AWS services
  • EKS Kubernetes Pods
  • Jenkins Administrative
Job Description
Role Overview: As a DevOps Platform Engineer at Capgemini, you will be responsible for handling DevOps tools installation and management on Linux systems and cloud. You will play a key role in designing, configuring, and managing end-to-end CI/CD pipelines, utilizing tools like Jenkins, Git, Terraform, Docker, and Ansible for automation and deployment. Your expertise in Linux system administration and experience in managing complete CI/CD workflows will be essential in this role. Key Responsibilities: - Install and manage DevOps tools on Linux systems and cloud platforms - Design, configure, and manage CI/CD pipelines using Jenkins, Git, Terraform, Docker, and Ansible - Implement and manage complete CI/CD workflows using Bitbucket pipelines - Deploy and manage workloads on Kubernetes via EKS - Manage EKS Kubernetes Pods, including deployment, monitoring, and troubleshooting of containerized applications - Collaborate with the team to ensure efficient operation of DevOps processes Qualifications Required: - Experience in Linux system administration, including installation, configuration, and file system management - Proficiency in using key DevOps tools like Jenkins, Git, Terraform, Docker, and Ansible - Strong understanding of the DevOps lifecycle and experience in CI/CD pipeline management - Knowledge of AWS services and experience in deploying workloads on Kubernetes via EKS - Hands-on experience in Jenkins administration and managing CI/CD workflows - Ability to work effectively in a collaborative team environment and adapt to changing priorities Note: Capgemini is a global business and technology transformation partner, focusing on accelerating the transition to a digital and sustainable world. With a diverse team of over 340,000 members in more than 50 countries, Capgemini offers a range of career paths, comprehensive wellness benefits, and opportunities to work on cutting-edge projects in tech and engineering. Join Capgemini to unlock the value of technology and contribute to creating a more sustainable and inclusive world.,
ACTIVELY HIRING
posted 2 weeks ago
experience5 to 9 Yrs
location
Hyderabad, All India
skills
  • GCP
  • Kubernetes
  • Ansible
  • Python
  • Bash
  • Security
  • Jenkins
  • Harness
  • Docker
  • Containerization
  • Cloud Engineering
  • Compute Engine
  • Cloud Functions
  • PubSub
  • BigQuery
  • Terraform
  • API Gateways
  • OAuth 20
  • JWT
  • CICD
  • Google Cloud Professional Cloud Architect
  • Associate Cloud Engineer
Job Description
As a Cloud Engineer at HCL Technologies, you will play a key role in leveraging your expertise in cloud engineering, specifically focused on Google Cloud Platform (GCP). Your responsibilities will include: - Demonstrating proficiency in various GCP services such as Compute Engine, Kubernetes, Cloud Functions, Pub/Sub, and BigQuery. - Utilizing your skills in Terraform, Ansible, Python, and Bash to automate and streamline cloud operations. - Ensuring API Gateways & Security by implementing OAuth 2.0, JWT to maintain a secure cloud environment. - Hands-on experience with CI/CD tools like Jenkins for continuous integration and Harness for continuous delivery. - Proficiency in Docker and containerization to efficiently deploy and manage applications. To excel in this role, you are required to hold a certification as a Google Cloud Professional Cloud Architect or Associate Cloud Engineer. Additionally, with at least 5 years of experience in cloud engineering, you will have the opportunity to grow and evolve in your career at HCLTech. As a Cloud Engineer at HCL Technologies, you will play a key role in leveraging your expertise in cloud engineering, specifically focused on Google Cloud Platform (GCP). Your responsibilities will include: - Demonstrating proficiency in various GCP services such as Compute Engine, Kubernetes, Cloud Functions, Pub/Sub, and BigQuery. - Utilizing your skills in Terraform, Ansible, Python, and Bash to automate and streamline cloud operations. - Ensuring API Gateways & Security by implementing OAuth 2.0, JWT to maintain a secure cloud environment. - Hands-on experience with CI/CD tools like Jenkins for continuous integration and Harness for continuous delivery. - Proficiency in Docker and containerization to efficiently deploy and manage applications. To excel in this role, you are required to hold a certification as a Google Cloud Professional Cloud Architect or Associate Cloud Engineer. Additionally, with at least 5 years of experience in cloud engineering, you will have the opportunity to grow and evolve in your career at HCLTech.
ACTIVELY HIRING
posted 3 weeks ago
experience2 to 6 Yrs
location
Hyderabad, Telangana
skills
  • Go
  • C
  • C
  • Docker
  • Kubernetes
  • Relational database
  • Jenkins
  • Groovy
  • Ansible
  • Cassandra
  • Apache Spark
  • AWS EC2
  • ElastiCache
  • ELB
  • Terraform
  • NoSQL database
  • Google gRPC
  • protobuf
  • AWS S3
Job Description
As a Software Engineer II - Online Backend for Gameplay Services at Electronic Arts, you will be part of a team that powers online features for EA's games, aiming to make every player's multiplayer experience memorable. Your primary mission will be to optimize the matchmaking process to enhance the gaming encounters for players worldwide. **Key Responsibilities:** - Design brand new services covering all aspects from storage to application logic to management console - Enhance and add features to existing systems - Research and select new best-of-breed technologies to meet challenging requirements - Communicate with engineers from across the company to deliver the next generation of online features for both established and not-yet-released games - Aim to optimize performance and scalability of server systems - Participate in the full product cycle from design and testing to deployment and support for LIVE environments and game team customers - Maintain a suite of automated tests to validate the correctness of backend services - Advocate for best practices within the engineering team - Collaborate with product managers to enhance new features supporting EA's business **Qualifications:** - Bachelor/Master's degree in Computer Science, Computer Engineering, or related field - 2+ years professional programming experience with Go/C#/C++ - Experience with cloud computing products such as AWS EC2, ElastiCache, and ELB - Experience with technologies like Docker, Kubernetes, and Terraform - Experience with relational or NoSQL databases - Familiarity with all phases of the product development lifecycle - Strong problem-solving skills - Excellent collaboration abilities - Effective written and verbal communication skills - Self-motivated and capable of working autonomously **Bonus Qualifications:** - Experience with Jenkins and Groovy - Experience with Ansible - Knowledge of Google gRPC and protobuf - Experience with high traffic services and highly scalable, distributed systems - Familiarity with scalable data storage and processing technologies such as Cassandra, Apache Spark, and AWS S3 - Experience with stress testing, performance tuning, and optimization - Previous work experience in the games industry Join Electronic Arts, where creativity thrives, and ideas matter. Benefit from a holistic approach to wellness and a supportive work environment that encourages growth and development.,
ACTIVELY HIRING
posted 1 week ago

DevOps Engineer

Hitachi Careers
experience3 to 7 Yrs
location
Hyderabad, Telangana
skills
  • DevOps
  • Networking
  • IAM
  • Storage
  • Ansible
  • Jenkins
  • System Administration
  • Scripting
  • Bash
  • Python
  • Go
  • Configuration Management
  • Containerization
  • Orchestration
  • Docker
  • Kubernetes
  • Helm
  • Infrastructure Engineering
  • Systems Administration
  • Platform Engineering
  • Cloud Platforms
  • VPCs
  • Compute Services
  • Infrastructure as Code
  • Terraform
  • CICD Pipelines
  • GitLab CI
  • CircleCI
  • ArgoCD
  • LinuxUnix Systems
  • Observability Tools
  • Logging Frameworks
  • ELK
  • Datadog
  • Fluentd
  • Prometheus
  • Grafana
  • DevOps Principles
  • AgileLean Methodologies
  • SDLC Practices
Job Description
Role Overview: As a DevOps Engineer at our organization, you will play a crucial role in driving the strategic direction, implementation, and continuous improvement of our DevOps practices. Your responsibilities will include developing scalable CI/CD pipelines, enhancing cloud infrastructure reliability, and ensuring secure and efficient software delivery. The ideal candidate is expected to be a technical expert with deep expertise in automation, cloud operations, configuration management, and infrastructure-as-code (IaC). Collaboration across engineering, security, product, and QA teams is essential to enable a culture of continuous delivery, operational excellence, and system reliability. Key Responsibilities: - Develop scalable CI/CD pipelines - Enhance cloud infrastructure reliability - Ensure secure and efficient software delivery - Collaborate across engineering, security, product, and QA teams - Enable a culture of continuous delivery, operational excellence, and system reliability Qualifications Required: - Bachelor's or Master's degree in Computer Science, Engineering, or a related technical discipline - 3 to 5 years of overall experience in infrastructure engineering, DevOps, systems administration, or platform engineering - Hands-on expertise in cloud platforms (AWS, Azure, or GCP) with deep knowledge of networking, IAM, VPCs, storage, and compute services - Strong proficiency in Infrastructure as Code (IaC) using Terraform, Ansible, or equivalent - Experience building and managing CI/CD pipelines using tools such as Jenkins, GitLab CI, CircleCI, or ArgoCD - Strong background in Linux/Unix systems, system administration, scripting (e.g., Bash, Python, Go), and configuration management - Experience implementing containerization and orchestration using Docker, Kubernetes, Helm - Familiarity with observability tools and logging frameworks (e.g., ELK, Datadog, Fluentd, Prometheus, Grafana) - Solid understanding of DevOps principles, Agile/Lean methodologies, and modern SDLC practices Additional Company Details: At GlobalLogic, we prioritize a culture of caring, where we consistently put people first. You'll experience an inclusive culture of acceptance and belonging, with opportunities for continuous learning and development. We offer interesting and meaningful work that allows you to make an impact while maintaining balance and flexibility in your work-life integration. As a high-trust organization, integrity is a cornerstone of our value proposition to our employees and clients. (Note: The above information is based on the provided Job Description),
ACTIVELY HIRING
posted 1 month ago
experience2 to 6 Yrs
location
Hyderabad, Telangana
skills
  • Java
  • AWS
  • GIT
  • GitHub
  • Docker
  • Ansible
  • Kubernetes
  • CICD
Job Description
As a Software Engineer for developer productivity at Pega, you will work in a collaborative Agile team following the SCRUM model alongside talented engineers. Your responsibilities will include: - Collaborating with Product owners and stakeholders to design and implement exciting features into the product - Participating actively in technical discussions to enhance product capabilities and resolve complex issues - Ensuring that the delivered work meets functional and technical requirements as well as quality standards - Debugging and maintaining the existing code base - Documenting and adhering to development standards for all coding - Unit testing code to ensure technical requirements are met - Ensuring proper documentation and knowledge transfer of information - Working on complex issues to design and develop high-quality software Qualifications required for this role: - 2-4 years of software development experience, preferably in a product development company - Good understanding of programming languages, design, continuous integration, and delivery (CI/CD) - Knowledge of programming languages such as Java and proficiency in writing code and scripts - Hands-on experience in Java and AWS is preferred, along with experience in building CI/CD orchestration - Experience in code, build, and release management using GIT, GitHub workflows - Familiarity with container and automation tools like Docker, Ansible, and Kubernetes - Strong interpersonal, communication, presentation, analytical, and problem-solving skills At Pega, you will benefit from: - Gartner Analyst acclaimed technology leadership across product categories - Continuous learning and development opportunities - An innovative, inclusive, agile, flexible, and fun work environment - Unique team dynamics that set this position apart within the organization Please note, the specific Job ID for this role is 22675.,
ACTIVELY HIRING
posted 1 week ago
experience6 to 10 Yrs
location
Hyderabad, Telangana
skills
  • Java
  • RESTful APIs
  • SQL
  • data structures
  • algorithms
  • automated testing
  • deployment
  • ansible
  • Jenkins
  • Angular
  • Node
  • memory management
  • Spring Boot Framework
  • microservices architecture
  • API management platforms
  • Azure Cloud
  • NoSQL databases
  • Linux based infrastructure
  • CICD pipelines
  • Terraform
  • AzureDevOps
  • GitHub Actions
  • UIUX framework
  • concurrency
  • database query optimization
Job Description
Job Description: As a Software Engineer III at NCR Atleos, you will be expected to have expertise in Java and Spring Boot Framework with Java 11 or higher. Your responsibilities will include working with microservices architecture, RESTful APIs, and API management platforms. It is essential for you to have experience with Azure Cloud, SQL & NoSQL databases, and designing high-volume web services using API protocols and data formats. Strong understanding of data structures and algorithms will be crucial for designing efficient and scalable solutions. Moreover, you will work on Linux-based infrastructure and have experience with CI/CD pipelines, automated testing, and deployment tools like Terraform, Ansible, Jenkins, and AzureDevOps/GitHub Actions. Experience with UI/UX frameworks such as Angular and Node would be beneficial. You will also be responsible for driving code reviews, design reviews, and architecture discussions. Additionally, optimizing Java applications for performance and leading small teams for substantial projects will be part of your role. Your strong analytical and problem-solving abilities will help in addressing complex technical challenges, and excellent written and verbal communication skills are necessary for effective interaction with stakeholders, project managers, and other developers. Qualifications Required: - Expertise in Java and Spring Boot Framework - Experience with microservices architecture, RESTful APIs, and API management platforms - Experience in Azure Cloud, SQL & NoSQL databases - Experience in designing high-volume web services using API protocols and data formats - Strong understanding of data structures and algorithms - Experience working on Linux-based infrastructure - Experience with CI/CD pipelines and tools for automated testing and deployment (e.g., Terraform, Ansible, Jenkins, AzureDevOps/GitHub Actions) - Ability to create/configure IAC pipeline - Experience with UI/UX frameworks (e.g., Angular, Node) - Experience in optimizing Java applications for performance, including memory management, concurrency, and database query optimization - Capable of running and directing small teams for substantial projects - Strong analytical and problem-solving abilities - Excellent written and verbal communication skills About NCR Atleos: NCR Atleos is a global technology company headquartered in Atlanta, Georgia, creating exceptional self-service banking experiences. With a comprehensive self-service channel, NCR Atleos offers services, software, and hardware solutions to enhance financial access worldwide.,
ACTIVELY HIRING
posted 2 weeks ago
experience5 to 10 Yrs
location
Hyderabad, Telangana
skills
  • Windows
  • SCCM
  • Active Directory
  • Ansible
  • HyperV
  • Terraform
  • Trend Micro
  • Tenable
Job Description
As a Windows Server Admin at our company, your primary role will be to provide 24/7 L2/L3 support for Windows server infrastructure. You will be responsible for responding to monitoring alerts, troubleshooting OS-level issues, resolving performance problems related to CPU, memory, and disk I/O, managing Active Directory users, groups, and policies, fixing authentication and account lockout issues, applying and troubleshooting Windows Updates and patches, monitoring and restoring backups, supporting Hyper-V and virtual machine issues, reviewing logs, performing root cause analysis, maintaining documentation, and updating the knowledge base. Additionally, you will need to identify vulnerabilities, perform remediation, build and configure Windows Servers following CIS benchmarks, act as an escalation point for the IT Service Desk, automate provisioning and decommissioning of servers using tools like Ansible and Terraform, ensure server builds include Tronox security tools (SCCM, Trend Micro, Tenable), conduct regular backup and restore drills, collaborate with various teams, handle JIRA ticket lifecycle, support scheduled change windows and execute planned upgrades, and manage major business outages during weekends and holidays. Qualifications required for this role include a Bachelor's degree in Computer Science, Information Technology, or related field, 5+ years of experience in Windows Server administration and operations, strong knowledge of Active Directory, Hyper-V, and backup solutions, experience with automation tools such as Ansible and Terraform, familiarity with SCCM, Intune, Trend Micro, and Tenable, excellent troubleshooting and analytical skills, and relevant certifications (e.g., Microsoft Certified: Windows Server, ITIL) are a plus. It is essential that you are willing to work in 24/7 rotational shifts, including weekends and holidays. If you are interested in this position and meet the qualifications, please share your resume to bhavana.madugula@wisseninfotech.com.,
ACTIVELY HIRING
posted 5 days ago
experience8 to 12 Yrs
location
Hyderabad, Telangana
skills
  • Configuration management
  • Automation
  • Ansible
  • Python
  • Patching
  • Red Hat Satellite
  • DevOps
  • High Availability
  • Docker
  • VMWare
  • Linux operating systems
  • OS installations
  • Upgrades
  • Server OS maintenance
  • SRE methodologies
  • Disaster Recovery environments
  • Virtual container applications
  • Deployment automation tools
  • Storage technologies
  • Change management processes
  • Incident management processes
  • Compliance frameworks
Job Description
As a Linux Engineer, your role involves designing, implementing, and evaluating current and future infrastructure initiatives to ensure the highest level of efficiency and performance. You will collaborate closely with various IT groups to contribute to overall architectural design, implementation, and troubleshooting. Your key responsibilities include: - Analyzing, acquiring, installing, modifying, and supporting Linux operating systems, utilities, and Internet/intranet-related tools - Implementing configuration management and automation using tools like Ansible and Python - Conducting problem diagnosis, troubleshooting, and resolution - Performing Linux OS installations, patching, and upgrades - Maintaining operating systems and associated infrastructure software - Designing and documenting SOPs for system administration activities - Installing and maintaining Server OS, configuration, and patching levels using tools like Red Hat Satellite - Utilizing your expertise in DevOps and SRE methodologies - Being able to work night shifts (IST) as required In terms of qualifications and skills, you should have: - 8+ years of experience in a level 2/3 enterprise server systems support role and/or Linux-based administrator position - Expertise in on-premise and cloud provider systems administration - Proficiency in working with Linux Server in a virtualized environment - Skill in integrating various Linux distributions (particularly Red Hat Enterprise and Ubuntu) in heterogeneous environments - Proficiency in applying Red Hat Satellite to manage all Linux distributions - Experience with High Availability and Disaster Recovery environments, virtual container applications (Docker), deployment automation tools (Ansible), virtualized environments (VMWare), and storage technologies (RAID, CIFS/SMB, NFS, NAS, and SAN) - Familiarity with change and incident management processes - Self-motivation to resolve a wide range of issues creatively - Networking with other senior personnel in your area of expertise - Holding a Bachelor's Degree in Computer Science, MIS, or a related field - Preferably having RHCSA and/or RHCSE certifications - Experience with compliance frameworks such as CIS, NIST, HIPPA, ISO, etc. would be a plus Your expertise as a Linux Engineer will be crucial in ensuring the smooth operation of infrastructure and systems within the organization.,
ACTIVELY HIRING
posted 2 months ago
experience3 to 7 Yrs
location
Hyderabad, Telangana
skills
  • Python
  • Java
  • Operating Systems
  • Networking
  • DevOps
  • Ansible
  • Kubernetes
  • Machine Learning
  • Security Principles
  • Terraform
  • AWS Cloud
  • GenAI
  • Ray
  • Ray Serve
Job Description
As a member of the Site Reliability Engineering (SRE) team in the Applied Machine Learning team within AI and Data Platform org at Apple, you will play a crucial role in designing, developing, and maintaining core platform components that support the fraud decisioning and solutioning infrastructure. Your responsibilities will include developing automation and tooling to enhance operational efficiency, collaborating closely with engineering teams to ensure high availability, resiliency, and security of critical systems, and monitoring and optimizing production platforms for performance and reliability. Joining our team requires a passion for building and operating large-scale platforms and distributed systems using cutting-edge open source technologies in hybrid cloud environments. **Key Responsibilities:** - Build, enhance, and maintain multi-tenant systems employing diverse technologies. - Collaborate with cross-functional teams to deliver impactful customer features. - Lead projects through the full lifecycle, from design discussions to release delivery. - Operate, scale, and optimize high-throughput and highly concurrent services. - Diagnose, resolve, and prevent production and operational challenges. **Qualifications Required:** - Bachelors Degree in Computer Science, Computer Engineering, or equivalent technical degree. - Proficient programming knowledge in Python or Java and ability to read and explain open source codebase. - Good foundation of Operating Systems, Networking, and Security Principles. - Exposure to DevOps tools such as Ansible and Terraform, with experience integrating platform components into Kubernetes and AWS Cloud environments. We are looking for enthusiastic engineers who are passionate about building and maintaining solutioning platform components on cloud and Kubernetes infrastructure. The ideal candidate will collaborate with stakeholders, understand the applications hosted on the platform, and design automation solutions that enhance platform efficiency, reliability, and value. If you are excited to work in a fast-paced, dynamic, and collaborative environment, possess strong coding skills, have a computer science foundation, and are eager to contribute to the development of resilient and highly performant distributed systems, we encourage you to apply.,
ACTIVELY HIRING
posted 2 months ago

Senior VMware Engineer

Gkhr consultancy
experience7 to 11 Yrs
location
Hyderabad, Telangana
skills
  • Operations
  • vCenter
  • Ansible
  • Python
  • VMware Cloud Foundation VCF Design Deployment
  • NSXT NSXV Network Virtualization Security
  • vSAN Implementation
  • Storage Policies
  • Capacity Planning
  • VMware Aria Suite Automation
  • Orchestrator
  • Lifecycle
  • vSphere 6x8x
  • ESXi
  • VMware HCX
  • SRM
  • migration projects
  • PowerCLI
Job Description
As a Senior VMware Engineer at our top MNC service based company, your role will involve leading and executing complex deployments and migrations across VMware Cloud Foundation (VCF), NSX, vSAN, and the VMware Aria Suite. You will focus on designing, implementing, and modernizing enterprise virtualization environments into software-defined data centers. Key Responsibilities: - Design, deploy, and manage VMware Cloud Foundation (VCF) environments. - Configure and manage NSX-T / NSX-V, including switching, routing, firewalls, and microsegmentation. - Set up and maintain vSAN clusters with stretched and fault-tolerant configurations. - Deploy and integrate VMware Aria Suite (Operations, Automation, Orchestrator, Lifecycle). - Lead migration projects using VMware HCX, SRM, and vMotion, ensuring smooth transitions from legacy systems. - Automate infrastructure provisioning via Aria Automation (vRA) and PowerCLI / Ansible / Python. - Perform troubleshooting, performance optimization, and post-deployment support. - Prepare HLDs, LLDs, SOPs, and conduct knowledge transfer sessions. Qualifications Required: - 7+ years of experience in enterprise virtualization and cloud infrastructure. - Expertise in VMware Cloud Foundation (VCF) Design & Deployment, NSX-T / NSX-V Network Virtualization & Security, vSAN Implementation, Storage Policies, and Capacity Planning, VMware Aria Suite Automation, Operations, Orchestrator, Lifecycle. - Deep understanding of vSphere 6.x8.x, vCenter, and ESXi. - Hands-on experience with VMware HCX, SRM, and migration projects. - Scripting knowledge in PowerCLI, Ansible, or Python (preferred). Please note that this role offers a hybrid work mode with 3 days work from the office and is open for candidates across PAN India.,
ACTIVELY HIRING
posted 3 days ago
experience4 to 8 Yrs
location
Hyderabad, Telangana
skills
  • Java
  • Python
  • Docker
  • Ansible
  • Spring Boot
  • Golang
  • NodeJS
  • Linux debugging
  • Terraform
  • Bash programming
Job Description
As a Senior Software Engineer, your primary responsibilities will include: - Creating microservices, components, and tools to automate the life cycle of highly available (HA), distributed services and applications across various platforms. - Incorporating security, quality, scalability, and availability aspects into your code. - Implementing automated testing to ensure product protection. - Utilizing Continuous Integration (CI) and Continuous Deployment to facilitate the deployment of containerized applications on AWS. - Collaborating with the operations team, product managers, and architects to design and develop infrastructure and software solutions. The ideal candidate should have the following skills: - Proficiency in at least one of the following languages: Java, Golang, NodeJS, Python, or equivalent. - Familiarity with container technologies such as Docker and Linux debugging. - Strong interpersonal skills and the ability to work effectively in a fast-paced, collaborative environment. It would be advantageous if you have: - Experience with cloud platforms like Kubernetes, Mesos, Cloud Foundry, OpenShift, Docker, AWS, GCP, Azure, or Serverless. - Previous knowledge of configuration management and code infrastructure using tools like Terraform and Ansible. - Familiarity with Spring Boot and Bash programming skills. - A BS/MS degree in Computer Science or Mathematics.,
ACTIVELY HIRING
posted 2 weeks ago
experience8 to 12 Yrs
location
Hyderabad, Telangana
skills
  • Python
  • JavaScript
  • AWS
  • Ansible
  • PostgreSQL
  • MySQL
  • MongoDB
  • Redis
  • Kafka
  • RabbitMQ
  • Docker
  • Kubernetes
  • RESTful APIs
  • DICOM
  • HL7
  • Active Directory
  • SSO
  • LDAP
  • OAuth
  • TypeScript
  • Terraform
  • GraphQL
  • FHIR
  • OpenID Connect
Job Description
As a Senior Software Engineer at our company, you will play a crucial role in leading the design and development of scalable, cloud-based healthcare SaaS applications. Your responsibilities will be hands-on, with a focus on coding, code reviews, technical guidance, and other key areas such as architecture, system design, production support, and team mentorship. Key Responsibilities: - Build high-quality, maintainable backend services and conduct thorough code reviews - Design and implement fault-tolerant distributed systems for healthcare SaaS products - Translate product requirements into technical specifications and architectural designs - Lead incident response, troubleshoot complex production issues, and drive performance improvements - Ensure that solutions comply with HIPAA, security best practices, and relevant medical device/QMS regulations - Implement integrations with EMRs, medical imaging systems, and healthcare data standards (DICOM, HL7, FHIR) - Recommend technologies, frameworks, and architectural patterns that align with product and scaling goals - Collaborate closely with Product, Project Management, and Support Engineering teams Qualifications Required: - Bachelors degree in Computer Science, Software Engineering, or a related field - 8+ years of professional software development experience, with a focus on senior-level contributions - 5+ years of experience in building and scaling cloud-based SaaS applications - 3+ years of experience working with distributed systems and microservice architectures - Expert proficiency in Python and JavaScript/TypeScript - Deep experience with AWS (EC2, S3, RDS, Lambda, EKS, etc.) - Proficiency with Terraform and Ansible for automation and CI/CD - Strong understanding of microservices, event-driven architectures, and distributed data management - Experience with relational and NoSQL databases (PostgreSQL/MySQL, MongoDB/Redis) - Experience with message brokers (Kafka, RabbitMQ, SQS/SNS) and containerization (Docker, Kubernetes) - Knowledge of RESTful APIs, GraphQL, and API gateway patterns - Experience with healthcare standards (DICOM, HL7, FHIR) and EMR integrations (preferred) - Experience with authentication systems (Active Directory, SSO, LDAP, OAuth, OpenID Connect) (preferred) In addition to the outlined job responsibilities and qualifications, the company offers health insurance, paid time off, and a provident fund as part of the benefits package. This is a full-time position with the requirement to work in person at the specified location.,
ACTIVELY HIRING
posted 2 months ago

Senior Python Developer

Blueberry Unicorn Services
experience5 to 9 Yrs
location
Hyderabad, Telangana
skills
  • Python
  • AWS
  • GCP
  • OCI
  • Flask
  • SQLAlchemy
  • Async
  • Windows
  • Linux
  • Ansible
  • Chef
  • Eventbased programming
  • Terraform
  • Cloudnative templates
  • Container technology
  • CICD
Job Description
Job Description: As a Senior Python Developer in this hybrid role with work hours from 2:00 PM to 11:00 PM, you will be expected to have experience working on Python SDKs for AWS, GCP, and OCI, which will be considered a plus. Your responsibilities will include demonstrating a strong knowledge of Python development, with practical experience in API and ORM frameworks such as Flask and SQLAlchemy. Moreover, your expertise in Async and Event-based task execution programming will be essential for this role. Key Responsibilities: - Demonstrate a strong knowledge of Python development - Work on Python SDKs for AWS, GCP, and OCI - Practical experience in API and ORM frameworks such as Flask and SQLAlchemy - Possess a strong understanding of both Windows and Linux environments - Practical experience with automation tools like Ansible or Chef - Hands-on experience with at least one cloud provider - Proficiency in writing Terraform or cloud-native templates - Knowledge of container technology - Experience with CI/CD practices Qualifications Required: - Strong understanding of Python development - Experience with Python SDKs for AWS, GCP, and OCI - Practical experience in API and ORM frameworks - Familiarity with Windows and Linux environments - Hands-on experience with automation tools like Ansible or Chef - Knowledge of cloud providers and writing Terraform or cloud-native templates - Understanding of container technology - Experience with CI/CD practices,
ACTIVELY HIRING
posted 1 week ago
experience5 to 9 Yrs
location
Hyderabad, Telangana
skills
  • VMware vSphere
  • vCenter
  • AWS
  • Azure
  • Bash
  • Python
  • Ansible
  • PowerCLI
  • Terraform
  • Microsoft Power Platform
  • PowerBI
Job Description
As a Private Cloud Infrastructure Engineer at Zurich, your role will involve leading the evolution of hybrid and private cloud environments. You will utilize your expertise in VMware, automation, and public cloud integration (AWS, Azure) to engineer, script, optimize platforms, and drive migration to modern cloud platforms. Key Responsibilities: - Design, deploy, and manage private and hybrid cloud infrastructure with VMware vSphere/vCenter integration to AWS and Azure. - Engineer and automate provisioning, monitoring, and management tasks using scripting (PowerCLI, Bash, Python) and tools like Ansible, Terraform, Azure DevOps, AWS CloudFormation. - Lead migration of virtualized workloads to public cloud, optimizing for cost, scalability, and performance. - Continuously monitor and optimize platform usage for efficiency and resource consumption transparency. - Collaborate with networking, security, and application teams to ensure secure and robust cloud operations. - Provide actionable insights on cost, efficiency, and platform improvements. - Participate in occasional out-of-hours works. Your Skills & Experience: - Deep expertise in VMware vSphere, vCenter, and related private/hybrid cloud technologies. - Good working knowledge of enterprise-level storage platforms like Dell EMC (Powerflex, PowerMax, Data Domain) and NetApp. - Experience in integrating and migrating workloads to AWS and Azure. - Strong skills in scripting and automation (PowerCLI, Bash, Python, Ansible, Terraform). - Proven ability to optimize cloud and virtualization environments for cost, performance, and efficiency. - Experience with Microsoft Power Platform apps, specifically PowerBI for visuals and interactive dashboards. - Strong problem-solving skills and ability to communicate technical concepts clearly. About NTT DATA: NTT DATA is a $30 billion business and technology services leader, serving 75% of the Fortune Global 100. Committed to accelerating client success and responsible innovation, NTT DATA is a global AI and digital infrastructure provider. With expertise in enterprise-scale AI, cloud, security, connectivity, and more, NTT DATA helps organizations transition confidently into the digital future. As a Global Top Employer, NTT DATA operates in more than 50 countries and offers access to innovation centers and a network of partners. NTT DATA is part of the NTT Group, which invests over $3 billion annually in R&D.,
ACTIVELY HIRING
posted 1 week ago
experience0 to 3 Yrs
location
Hyderabad, Telangana
skills
  • Ansible
  • Kubernetes
  • OpenShift
  • Zabbix
  • Virtual Machines
  • RH Satellite
  • Containers
  • Cloud AWS Azure GCP
  • CI CD Jenkins
Job Description
As a member of the Micron Technology team, you will play a crucial role in maintaining essential IT operations, ensuring system uptime, and providing technical support for issue resolution. Your responsibilities will include: - Having 0-2 years of experience in maintaining various IT operations such as operating systems, security tools, applications, servers, laptops, desktops, software, and hardware while collaborating with different Lab teams. - Providing 24/7 on-call support in rotational shifts and ensuring SLA compliance for all tickets. - Following up on escalated issues and classifying incidents to support system uptime. - Collaborating with engineers to resolve or escalate issues and delivering technical support for issue resolution. - Communicating ticket status to engineers and managers and utilizing the IT knowledge base and best practices. - Monitoring systems to prevent downtime, improving support processes for recurring issues, and providing day-to-day technology direction across Linux and Windows hosts in Lab, Data Center, and Cloud environments. - Troubleshooting operational issues, handling escalations, and resolving business partner issues with strong collaboration and attention to business priorities. - Maintaining global collaboration with other Micron IT and Engineering teams and continuously improving operations through routine monitoring of system performance metrics. Technical Skills & Tools: - Ansible, Kubernetes / OpenShift, RH Satellite, Zabbix - Virtual Machines, Containers, Cloud (AWS / Azure / GCP) - CI / CD (Jenkins) - Ability to learn and use multiple utilities and tools that support operations monitoring and alerting - Participate in and learn new solutions through condensed knowledge transfer sessions About Micron Technology, Inc.: Micron Technology is an industry leader in innovative memory and storage solutions, transforming how the world uses information to enrich life for all. With a focus on technology leadership, Micron delivers a rich portfolio of high-performance DRAM, NAND, and NOR memory and storage products through its Micron and Crucial brands. The innovations created by Micron's people fuel the data economy, enabling advances in artificial intelligence and 5G applications. To learn more about Micron Technology and explore career opportunities, please visit micron.com/careers. Micron prohibits the use of child labor and complies with all applicable laws, rules, regulations, and industry labor standards. Candidates are encouraged to use AI tools to enhance their application materials, ensuring accuracy and reflecting true skills and experiences. Micron advises job seekers to be cautious of unsolicited job offers and verify the authenticity of any communication claiming to be from Micron by checking the official Micron careers website.,
ACTIVELY HIRING
logo

@ 2025 Shine.com | All Right Reserved

Connect with us:
  • LinkedIn
  • Instagram
  • Facebook
  • YouTube
  • Twitter