release-engineer-jobs-in-noida, Noida

3,865 Release Engineer Jobs in Noida

Toggle to save search
posted 2 months ago

Backend Engineer

NetSysCon Consulting LLP
experience3 to 6 Yrs
Salary10 - 12 LPA
location
Gurugram
skills
  • java
  • development
  • agile
  • scrum
  • api development
  • spring boot
  • microservices
  • backend
  • aws
Job Description
Job Title: Backend Engineer FinTech Startup Location: Gurugram Experience: Minimum 3 years Salary: 10 to 12 LPA Industry: FinTech / Financial Services Functional Area: Software Development / Engineering About the Company: Join a fast-growing FinTech startup backed by top-tier venture capitalists, marquee angel investors, and industry mentors. The company is building innovative financial technology solutions that redefine how people interact with money and financial services. About the Role: We are seeking a skilled Backend Engineer who is passionate about building scalable, high-performance systems. You will be responsible for developing backend modules, enhancing existing APIs, and ensuring smooth integrations across the platform. Key Responsibilities: Design, develop, and deploy scalable backend modules and RESTful APIs. Enhance and maintain existing APIs and services. Collaborate with cross-functional teams to conceptualize and release new features. Troubleshoot and resolve technical issues in software design and maintenance. Write clean, testable, and efficient code following best practices. Implement unit and integration tests to ensure reliability. Own assigned modules end-to-end with guidance from senior engineers. Participate in code reviews, sprint planning, and team discussions. Requirements: Education: Bachelor's degree in Computer Science, Information Technology, or related field. Experience & Skills: Minimum 3 years of experience as a Backend Developer. Strong proficiency in Java and Spring Boot framework. Hands-on experience with microservices architecture and API development. Sound understanding of data structures, algorithms, and multi-threading concepts. Familiarity with reactive programming. Experience with AWS services (S3, EKS, MSK, ELB) is a strong plus. Knowledge of Scrum / Agile methodologies. Strong problem-solving and analytical skills. Excellent communication and teamwork abilities. Tech Stack: Core: Java, Spring Boot, Microservices Cloud: AWS (S3, EKS, MSK, ELB) Testing: Unit and Functional Testing frameworks Why Join: Opportunity to work with an experienced founding team and top-tier investors. Ownership and autonomy in building scalable systems from the ground up. Fast-paced, collaborative, and growth-oriented culture.
INTERVIEW ASSURED IN 15 MINS

Top Companies are Hiring in Your City

For Multiple Roles

Jio Platforms Ltd
Jio Platforms Ltdslide-preview-Genpact
posted 2 months ago
experience8 to 13 Yrs
location
Noida, Delhi+6

Delhi, Bangalore, Chennai, Hyderabad, Kolkata, Pune, Coimbatore

skills
  • python
  • kubernetes
  • terraform
Job Description
We are looking for a Senior Kubernetes Platform Engineer with 10+ years of infrastructure experience to design and implement the Zero-Touch Build, Upgrade, and Certification pipeline for our on-premises GPU cloud platform. This role focuses on automating the Kubernetes layer and its dependencies (e.g., GPU drivers, networking, runtime) using 100% GitOps workflows. You will work across teams to deliver a fully declarative, scalable, and reproducible infrastructure stackfrom hardware to Kubernetes and platform services.  Key Responsibilities Architect and implement GitOps-driven Kubernetes cluster lifecycle automation using tools like kubeadm, ClusterAPI, Helm, and Argo CD. Develop and manage declarative infrastructure components for: o GPU stack deployment (e.g., NVIDIA GPU Operator) o Container runtime configuration (Containerd) o Networking layers (CNI plugins like Calico, Cilium, etc.)
INTERVIEW ASSURED IN 15 MINS
posted 2 months ago

Epic Software Engineer Remote

Venpa Global Technologies Private Limited
experience4 to 7 Yrs
WorkRemote
location
Delhi, Bangalore+3

Bangalore, Chennai, Hyderabad, Gurugram

skills
  • javascript
  • node.js
  • epic systems
  • hipaa
  • fhir
  • epic apis
  • react
Job Description
Epic Software Engineer Budget : 150000/month Remote*  About the role Our client is hiring an Epic Software Engineer to build and integrate apps that run inside Epic and connect with a patient services platform. You will design secure, scalable workflows for enrollment, consent, surveys, provider updates, and analytics using Epic APIs, SMART on FHIR, HL7 v2, and modern web technologies.  4+ years of professional software engineering experience, with at least 2 years in healthcare integrations or EHR app development. Hands-on experience with Epic APIs, SMART on FHIR app development, and FHIR resource modeling. Strong proficiency with web development using JavaScript or TypeScript, React, and HTML/CSS. Proficiency building RESTful APIs and JSON contracts, including request validation, versioning, and backward compatibility strategies. Practical knowledge of HL7 v2 segments, message types, ACK handling, and interface engines. Experience implementing OAuth2.0; familiarity with SAML for enterprise SSO. Solid backend skills in Node.js or PHP; experience with C#/.NET or Java/Kotlin is a plus. SQL expertise with PostgreSQL or MS SQL, including schema design and query tuning. Docker-based development and AWS deployment experience.  Preferred qualifications Experience publishing Epic apps through Epic programs and working with Epic client teams for onboarding. Familiarity with interface engines such as Mirth, Rhapsody, or Cloverleaf. Knowledge of consent frameworks, TCPA considerations for SMS, and de-identification techniques. Experience with SFTP-based data exchange at scale, checksum validation, and idempotent ingestion design. CI/CD with GitHub Actions, GitLab CI, or similar, plus infrastructure as code on AWS. Observability with tools such as CloudWatch, OpenTelemetry, or Datadog. Tableau or analytics pipeline experience for clinical or brand reporting. Mobile development exposure in Swift or Objective-C for iOS and Java or Kotlin for Android for companion apps.  
INTERVIEW ASSURED IN 15 MINS
question

Are these jobs relevant for you?

posted 2 months ago

MLOps engineer

CAPGEMINI TECHNOLOGY SERVICES INDIA LIMITED
CAPGEMINI TECHNOLOGY SERVICES INDIA LIMITED
experience6 to 11 Yrs
location
Noida, Bangalore+4

Bangalore, Chennai, Hyderabad, Pune, Mumbai City

skills
  • machine learning
  • python
  • mlops
Job Description
Fill below mandatory link.  MLops-virtual interview with Capgemini--all locations, Monday 27th Oct 2025 Fill out form   Position: - Machine Learning Ops Harvest Level: Software Engineer  Job Description:    Machine Learning Ops Engineer is the one whos responsible for managing and optimizing the deployment and operation of machine learning models in production environments. He/she works closely with data scientists, software engineers, and DevOps teams to ensure that machine learning models are integrated seamlessly into existing systems and can scale effectively. Role & Responsibilities Responsible for design and implementation of secure and scalable infrastructure in Azure cloud. Build and maintain CI/CD/CT pipeline across azure cloud platform for Data science projects. Own and automate the infrastructure provisioning, demand forecasting and capacity planning. Build tools and automation to improve systems availability, reliability, performance, monitoring and scalability. Setting up the alerts and monitoring for the platform. Monitoring system, application health, security controls and cost Envision, implement and rollout best MLOPs/DevOps tools and automation.  Requirements: Strong understanding of concepts related to Machine Learning, Architecture, and MLOps practices. Proficient in Azure Cloud practices and managing Azure Kubernetes Service infrastructure. Hands on experience in Python or any other scripting languages (Groovy, Shell) Experience with monitoring tools. Hands-on experience in Docker, Kubernetes. Excellent experience with source code management tools (git) Experience with modern cloud development practices such as Microservices Architecture, REST Interfaces etc. Experience in implementing Data/Model drift. Strive for Continuous Improvement and build continuous integration, continuous development, and constant deployment pipeline. Excellent Troubleshooting skills Always ready to learn more and adopt new cutting-edge technology with right value proposition.  
INTERVIEW ASSURED IN 15 MINS
posted 3 weeks ago

Software Or Platform Engineer

HYRE GLOBAL RECRUITMENT AND SERVICES PRIVATE LIMITED
experience5 to 6 Yrs
Salary7 - 10 LPA
WorkContractual
location
Noida, Delhi+7

Delhi, Bangalore, Chennai, Hyderabad, Kolkata, Gurugram, Pune, Mumbai City

skills
  • python
  • kubernetes
  • azure
  • angular
  • terraform
  • typescript
  • bicep
Job Description
Job Title: Software Engineer Platform & Infrastructure Department: Software Engineering Job Category: Individual Contributor Reporting Structure: Reports To: Manager Software Engineering Reportees: None Peers: Software Engineers, Software Architects Stakeholders: Product Managers Location: Europe (Travel Required) Job Summary: The Software Engineer Platform & Infrastructure is responsible for building and maintaining the foundational systems that enable efficient, secure, and reliable software delivery across the organization. This includes cloud infrastructure automation, CI/CD pipeline development, Kubernetes environments, SDKs, data engineering support, and observability tooling. The engineer collaborates closely with development teams to create performant, developer-friendly platforms for both applications and data systems. Key Responsibilities: Design and implement Infrastructure as Code (IaC) using tools like Terraform, Bicep, and manage Azure cloud and Kubernetes environments. Develop and maintain CI/CD pipelines in Azure DevOps. Implement and manage monitoring and alerting systems (e.g., Prometheus, Grafana, Azure Monitor). Maintain and enhance the internal Developer Portal (Backstage). Build reference or sample applications using C#, TypeScript, or Java. Support data workloads and pipelines (Spark, PySpark) under team guidance. Identify and automate repetitive operational tasks to improve reliability and efficiency. Enhance monitoring and observability for better visibility, reliability, and proactive issue detection. Collaborate with development teams to define requirements, architecture, and specifications. Coach and support teams on DevSecOps best practices and platform utilization. Contribute to improving developer onboarding and maintaining comprehensive platform documentation. Qualifications, Skills, and Competencies: Education: Bachelors degree in Computer Science, Engineering, or a related field. Language: Proficiency in English (spoken and written). Experience: Proven experience as a Software Engineer or Platform Engineer. Technical Proficiency: Strong programming experience in C#, Python, TypeScript, or Angular. Expertise in Azure Cloud, Terraform, and Bicep. Hands-on with Kubernetes, networking, and monitoring tools. Deep understanding of CI/CD pipelines and DevSecOps principles. Exposure or interest in data engineering (Python, PySpark). Soft Skills: Excellent problem-solving, communication, teamwork, and attention to detail. Mindset: Proactive, automation-driven, and collaborative approach to platform engineering. Success Criteria / KPIs: Software Development & Reliability High platform reliability and reproducibility (pipeline success rates, infra drift metrics). Increased automation across delivery processes (manual task reduction). Fewer manual interventions or deployment incidents. Compliance with Service Level Objectives (SLOs). Improved developer satisfaction with platform usability. Collaboration & Communication Active participation in defining platform standards and requirements. Effective collaboration with development and operations teams. Constructive contributions to design reviews and incident postmortems. Code Quality & Best Practices Adherence to coding and infrastructure standards (code review success rate, test coverage). Positive impact on reliability and maintainability (fewer failures, reusable IaC modules, faster builds). Problem Solving & Debugging Efficient incident detection and resolution (MTTR, incident count). Proactive prevention of recurring issues and improvement in system stability. Learning & Growth Continuous learning and implementation of emerging technologies. Tangible skill growth through certifications, trainings, and project impact. Timely Delivery On-time delivery of automation and infrastructure improvements. Minimal delays in development or deployment due to platform dependencies.
INTERVIEW ASSURED IN 15 MINS
posted 1 week ago

Data Engineer

Lorven Technologies Private Limited
experience6 to 10 Yrs
WorkRemote
location
Noida, Delhi+6

Delhi, Bangalore, Chennai, Kolkata, Gurugram, Pune, Mumbai City

skills
  • azure
  • data engineering
  • data
  • bricks
  • factory
  • .net
  • python
Job Description
Data Engineer Azure Data Factory, SQL, Python, Databricks, ETL  Data Engineer .Net + Azure Data Factory  We are seeking a skilled Data Engineer to join our team. The ideal candidate will be responsible for designing, implementing, and maintaining data systems and architectures that support our business needs. This role requires expertise in working with large datasets, cloud technologies, and advanced data platforms to ensure the availability, quality, and accessibility of data for business analytics and decision-making. Key Responsibilities: Design, build, and maintain data pipelines to collect, process, and store data efficiently from various sources. Work with cross-functional teams to understand business requirements and deliver data solutions that meet the needs of the organization. Optimize data storage and retrieval methods to enhance system performance and data quality. Integrate data from multiple sources (databases, data lakes, cloud storage, etc.) to build comprehensive data sets. Build and maintain data infrastructure, ensuring it is scalable, reliable, and secure. Implement and manage data governance policies, ensuring data accuracy, consistency, and compliance. Conduct data modeling and provide insights into the data through advanced analytics tools and reports. Perform data transformations and data wrangling tasks to prepare data for analysis. Troubleshoot data issues and collaborate with stakeholders to resolve technical challenges. Ensure the integrity of data pipelines and systems by conducting routine testing and validation. Required Skills and Qualifications: Bachelors degree in Computer Science, Engineering, Information Systems, or related field. Proven experience as a Data Engineer or in a similar role, with a minimum of [X] years of experience. Proficiency in data engineering tools and technologies such as SQL, Python, Java, Scala, and ETL frameworks. Experience working with cloud platforms (AWS, Azure, GCP) and data storage solutions (e.g., Redshift, Snowflake, BigQuery). Solid understanding of data modeling, database design, and cloud data architectures. Hands-on experience with data warehousing concepts and tools (e.g., Apache Hive, Apache Spark). Familiarity with data orchestration tools such as Apache Airflow, Azure Data Factory. Knowledge of real-time data streaming technologies (e.g., Kafka, Kinesis) is a plus. Strong problem-solving skills and ability to troubleshoot complex data issues. Familiarity with data security best practices and data privacy regulations (e.g., GDPR, HIPAA).
INTERVIEW ASSURED IN 15 MINS
posted 1 week ago

Senior Snowflake Data Engineer

Hucon Solutions India Pvt.Ltd.
Hucon Solutions India Pvt.Ltd.
experience8 to 13 Yrs
Salary8 - 18 LPA
location
Noida, Delhi+7

Delhi, Bangalore, Chennai, Hyderabad, Gurugram, Kolkata, Pune, Mumbai City

skills
  • aws
  • sql
  • data modeling
  • snowflake
  • data build tool
  • five tran
Job Description
Senior Snowflake Data Engineer Location: PAN India Experience: 8+ Years Skills: Snowflake, dbt, FiveTran, Snowpipe, AWS (MWAA, S3, Lambda), GitHub CI/CD Job Description We are looking for an experienced Senior Snowflake Data Engineer with strong expertise in modern data warehousing, cloud technologies, and ELT pipeline development. The ideal candidate should have deep hands-on experience in Snowflake, dbt, cloud environments, and CI/CD practices, with the ability to design scalable and efficient data solutions. Key Responsibilities Analyze, integrate, model, and interpret large and complex datasets from multiple sources. Design and implement ELT data pipelines using dbt with Snowflake as the primary cloud data warehouse. Build efficient and scalable data transformation pipelines using dbt at an advanced level. Work with ETL/ELT and data governance tools such as FiveTran and Alation. Utilize advanced Snowflake features such as RBAC, Dynamic Tables, and various optimization techniques. Ensure strong data modelling and warehousing practices across diverse database technologies. Manage orchestrations using Apache Airflow or AWS MWAA, along with CI/CD pipelines. Oversee continuous deployment, monitoring, and operations of data solutions using GitHub Actions, Terraform, and other DevOps tools. Collaborate with technical and non-technical stakeholders through effective communication. Contribute to building future-state Data Warehouse capabilities using cutting-edge technologies. Adapt and work efficiently within Agile methodology. Preferred Qualifications 2+ years of hands-on experience with Snowflake as a Cloud Data Warehouse and Data Lake platform. Strong understanding of cloud environments, especially AWS (S3, Lambda, MWAA). Airline industry domain experience is a plus.
INTERVIEW ASSURED IN 15 MINS
posted 2 months ago

Senior Automation Engineer

LTIMindtree Limited
LTIMindtree Limited
experience7 to 12 Yrs
location
Noida, Bangalore+5

Bangalore, Chennai, Hyderabad, Kolkata, Pune, Coimbatore

skills
  • ansible
  • cloud
  • automation
  • engineer
Job Description
We are looking for an experienced Network Automation Engineer to design, implement, and optimize automation solutions for our Private Cloud datacenter network, which underpins large-scale AI/ML GPU and TPU workloads. This role focuses on automating configuration, provisioning, and monitoring of high-performance networking devices to ensure low latency, high throughput, and reliability in a mission-critical environment. This role involves automating network device management as well as OS-level network configurations on servers. Expertise in Ansible and Python is essential, and experience with GoLang is a strong plus.  Key Responsibilities: Develop and maintain network automation frameworks for large-scale datacenter environments supporting AI/ML workloads. Build Ansible playbooks, roles, and modules to automate device configurations, software upgrades, and compliance checks across multi-vendor environments. Design and implement Python-based automation scripts and tools to integrate with APIs, orchestration platforms, and monitoring systems. Automate OS core networking configurations on servers (Linux / Windows / Hypervisor) including bonding, VLANs, routing tables, kernel network parameters, MTU tuning, and NIC performance optimization.
INTERVIEW ASSURED IN 15 MINS
posted 1 day ago

Senior Software Engineer

LTIMindtree Limited
LTIMindtree Limited
experience3 to 12 Yrs
location
Chennai, Gurugram
skills
  • Adobe Journey Optimizer
Job Description
Key Responsibilities Design and implement customer journeys using Adobe Journey Optimizer Build and manage event driven journeys time based campaigns and real time personalization use cases Collaborate with cross functional teams marketing data and engineering to translate business requirements into AJO workflows Utilize AEP capabilities such as RealTime Customer Profile Segments Schemas XDM and Datasets to create journey conditions and triggers Work with APIs DecisioningAJOD Module and Event Forwarding for real time data ingestion and outbound integrations Design and test email templates push notifications and in app messages within AJO Implement offer decisioning and personalized content blocks using AJO Offer Library Perform unit testing UAT and AB testing within journeys to ensure optimal performance Monitor journey performance trouble shoot issues and optimize for KPIs eg engagement conversion Ensure governance and best practices around data privacy consent and compliance GDPR CCPA etcTechnical Skills Strong experience with Adobe Journey Optimizer AJO Working knowledge of Adobe Experience Platform AEP especially Schemas XDM modeling Segment Builder RealTime Customer Profile RTCP Datasets and data ingestion Familiarity with Customer Journey Analytics CJA and Adobe Offer Decisioning is a plus Proficient in JavaScript JSON and RESTful API integration Good understanding of data sources batch and streaming identity stitching and audience activation Experience with Tag Management Tools Adobe Launch or Tealium is a plus Preferred Qualifications Adobe Certified Expert Adobe Journey Optimizer Adobe Experience Platform Experience with other Adobe tools like Adobe CampaignMarketoEloqua Adobe Target or Adobe Analytics is a plus Experience in industries such as Retail Telecom BFSI or Healthcare is a plus Understanding of privacy regulations GDPR CCPA and consent frameworks
INTERVIEW ASSURED IN 15 MINS
posted 2 months ago
experience2 to 6 Yrs
location
Noida, Uttar Pradesh
skills
  • coding
  • database programming
  • Agile methodology
  • GitHub
  • Jira
  • communication
  • infrastructure tools
  • CICD principles
  • SDLC knowledge
  • Confluence familiarity
  • problemsolving
Job Description
As a Release Engineer, you will be responsible for designing, implementing, and maintaining CI/CD pipelines using industry-standard tools. You will collaborate with development and QA teams to support Agile delivery cycles and manage version control and branching strategies using GitHub. Your role will involve defining and executing release strategies aligned with business goals, troubleshooting deployment issues, and optimizing automation workflows. Additionally, you will work closely with stakeholders to ensure smooth communication and delivery. Key Responsibilities: - Design, implement, and maintain CI/CD pipelines using industry-standard tools. - Collaborate with development and QA teams to support Agile delivery cycles. - Manage version control and branching strategies using GitHub. - Define and execute release strategies aligned with business goals. - Troubleshoot deployment issues and optimize automation workflows. - Work closely with stakeholders to ensure smooth communication and delivery. Qualifications Required: - Proficiency in coding, database programming & infrastructure tools. - Strong understanding of CI/CD principles. - Experience with Agile methodology and SDLC. - Familiarity with GitHub, Jira, and Confluence. - Excellent communication and problem-solving skills. If you're ready to make an impact and grow with a collaborative team, we'd love to hear from you! Apply now or tag someone who'd be a great fit for this role.,
ACTIVELY HIRING
posted 2 days ago
experience3 to 7 Yrs
location
Delhi
skills
  • APIs
  • Performance Optimization
  • Quality Assurance
  • AI Software Engineer
  • LLMs
  • models
  • embeddings
  • RAG
  • LLMPowered Business Solutions
  • RetrievalAugmented Generation RAG Pipelines
  • Prompt Engineering
  • API
  • Microservices Development
  • Cloud Architecture Integration
  • nodejs
  • TensorFlow
  • PyTorch
  • LangChain
  • GPT models
Job Description
Role Overview: You are being sought after to fill the role of Lead AI Software Engineer at Occams Advisory. As part of the AI Initiatives team, your primary responsibility will be to spearhead the development of AI-powered solutions, generative AI, and AI business applications to create impactful outcomes. If you are deeply immersed in AI, LLMs, and APIs, this position is tailored for you. Your expertise in AI evolution, LLMs, generative models, and AI business use cases will be crucial in ensuring the success of our AI-driven solutions. Your proactive, analytical, and detail-oriented approach will be highly valued in this role. Key Responsibilities: - Develop LLM-Powered Business Solutions: Design, build, and deploy robust applications using open source LLM frameworks like Llama and Langchain to address specific business use cases. - Implement Retrieval-Augmented Generation (RAG) Pipelines: Develop and maintain scalable RAG systems by integrating with vector databases and memory modules to enhance response accuracy. - Advanced Prompt Engineering: Utilize advanced prompt engineering techniques to optimize user interactions and generate precise responses aligned with business requirements. - API and Microservices Development: Create secure RESTful APIs and microservices to seamlessly integrate LLM functionalities into existing systems with a focus on data security and compliance. - Cloud Architecture Integration: Collaborate with cloud services to ensure highly scalable, resilient, and cost-effective LLM applications. - Collaboration with Cross-Functional Teams: Partner with product managers, designers, and stakeholders to translate business requirements into technical solutions and refine applications iteratively based on feedback. - Performance Optimization and Quality Assurance: Monitor, troubleshoot, and optimize applications for performance to deliver high-quality, business-aligned outputs. Qualification Required: - Education: Bachelor's or master's degree in computer science, Data Science, Engineering, or related field; relevant certifications are a plus. - Experience: 3+ years of experience with OpenAI, Gemini, Lambda, Llama, Langchain integration into business use cases. - Essential Technical Skills: Proficiency in node.js, TensorFlow, PyTorch, LangChain, open source LLMs, transformers, embeddings, generative AI, AI model optimization, API development, cloud platforms, and working with large datasets. - Preferred Skills: Experience with OpenAI's GPT models, problem-solving skills, and collaboration with cross-functional teams. Additional Company Details: Occams Advisory, founded in 2012, is a prominent business solutions, professional services, and financial advisory provider. The company has achieved significant recognition for its rapid growth and innovative solutions, with a team of over 100 professionals operating worldwide. Occams is committed to delivering transformative growth and advisory solutions to its clients. (Note: Benefits & Perks and Job Details have been omitted as per the instructions),
ACTIVELY HIRING
posted 2 days ago
experience4 to 8 Yrs
location
Noida, Uttar Pradesh
skills
  • Manual Testing
  • API Testing
  • Database Testing
  • SDLC
  • STLC
  • Test Cases
  • Test Scenarios
  • Test Management Tools
  • Quality Center
  • Jira
  • TestRail
  • SQL Queries
  • Mobile Application Testing
  • Regression Testing
  • Integration Testing
  • Exploratory Testing
  • Defect Tracking
  • Documentation
  • Mentoring
  • Automation
  • Agile Scrum Methodology
  • Requirements Traceability Matrix
  • Requirement Tracking Tools
  • Bug Tracking Tools
  • QA Processes
  • Crossfunctional Collaboration
Job Description
As a Senior Manual QA Engineer with 4-6 years of experience, you will be responsible for the following key responsibilities: - Designing, developing, and executing comprehensive test plans and test cases for new features and regression testing. - Participating in requirements and design review meetings to provide QA perspective and identify potential issues early. - Performing functional, regression, integration, and exploratory testing across web and/or mobile platforms. - Tracking and reporting software defects using tools like Jira or similar; working closely with developers to troubleshoot and resolve issues. - Creating and maintaining documentation related to QA processes, test cases, and bug reports. - Leading and mentoring junior QA engineers, providing technical guidance and best practices. - Continuously evaluating and improving QA processes and suggesting automation opportunities. - Collaborating with cross-functional teams including development, product management, and support to ensure timely and high-quality releases. Priority qualifications for this role include: - Experience in Manual, API, and database testing. - Good understanding of Agile scrum Methodology. - Excellent verbal & written communication skills. - Good understanding of Software Development Life Cycle (SDLC) & Software Testing Life Cycle (STLC). - Expertise in writing Test Cases, Test Scenarios & Requirements traceability matrix. - Good knowledge of using any test management tools like Quality Center/Jira/TestRail etc. - Good knowledge of using any requirement and bug tracking tools like Jira/Bugzilla/Asana/Zoho etc. - Good experience in Database Testing (Writing and executing SQL queries). - Practical knowledge of API testing. - Ability to work as an individual contributor. - Experience in testing Mobile applications will be an added advantage. In addition to the above, as part of this role, you can benefit from: - Flexible Work Hours. - Working on impactful, industry-changing projects. - Continuous learning and career growth opportunities. - Collaborative, innovative, and inclusive environment. - Insurance coverage including Group Medical Coverage, Employee Personal Accident, Maternity and Paternity Leaves. - Performance Feedback Sessions. - Competitive Compensation. - Rewards and Recognition Program, Employee Referral Program, Fun Events.,
ACTIVELY HIRING
posted 2 days ago
experience3 to 7 Yrs
location
Noida, Uttar Pradesh
skills
  • MongoDB
  • AWS
  • Docker
  • MERN Stack
  • Electronjs
  • Expressjs
  • React
  • Nodejs
  • OpenAI
Job Description
Role Overview: As an experienced MERN (MongoDB, Electron.js, Express.js, React, and Node.js) Stack Developer with a solid background in web development and expertise in AWS, Docker, and OpenAI, you will be a key member of the team responsible for creating and implementing cutting-edge web applications and integrating AI-driven functionalities into products. Key Responsibilities: - Develop and maintain top-notch web applications using the MERN stack - Collaborate with designers and fellow developers to build user-friendly interfaces - Design efficient database schemas in MongoDB - Write server-side logic with Node.js, Electron.js, and Express.js - Create responsive front-end components using React - Seamlessly integrate third-party APIs and libraries - Ensure the security and privacy of user data - Stay updated with code versioning tools and best practices - Work on cloud infrastructure deployment on AWS - Containerize applications with Docker - Implement AI and natural language processing capabilities with OpenAI technologies - Troubleshoot technical issues - Keep abreast of emerging technologies and industry trends Qualifications Required: - Bachelor's degree in computer science or a related field (or equivalent work experience) - Proven experience as a MERN Stack Developer with a strong portfolio of developed web applications - Expertise in MongoDB, Electron.js, Express.js, React, and Node.js - Proficiency in AWS services like EC2, S3, Lambda, and CloudFormation - Experience with Docker and Kubernetes for containerization and orchestration - Familiarity with AI technologies and RESTful API design - Strong problem-solving skills - Excellent teamwork and communication abilities - Self-motivated approach to work independently while providing constructive feedback in code reviews,
ACTIVELY HIRING
posted 1 day ago
experience4 to 8 Yrs
location
Noida, Uttar Pradesh
skills
  • Docker
  • Kubernetes
  • Python
  • Bash
  • AWS ecosystem
  • CICD tools
  • Terraform
  • FHIR R4 UK
  • SNOMED CT
  • MIG connectors
  • EMR integration
  • Agile tools
  • Material UI
  • AuthenticationAuthorization mechanisms
Job Description
As a DevOps Engineer in the healthcare domain, your role will involve designing, building, and maintaining robust CI/CD pipelines using tools like GitHub Actions, GitLab CI, or Jenkins. You will be responsible for managing and automating containerized applications using Docker and orchestration tools like Kubernetes (EKS) or AWS ECS. Collaborating with development teams for frontend-backend deployment coordination and ensuring smooth rollout of Material UI-based interfaces will be crucial. Additionally, integrating and managing Electronic Medical Record (EMR) workflows, MIG connectors, and related systems will be part of your responsibilities. Key Responsibilities: - Design, build, and maintain robust CI/CD pipelines using tools like GitHub Actions, GitLab CI, or Jenkins. - Manage and automate containerized applications using Docker and orchestration tools like Kubernetes (EKS) or AWS ECS. - Collaborate with development teams for frontend-backend deployment coordination; ensure smooth rollout of Material UI-based interfaces. - Integrate and manage Electronic Medical Record (EMR) workflows, MIG connectors, and related systems. - Implement and maintain monitoring, logging, and alerting systems using CloudWatch and third-party tools. - Leverage AWS services (e.g., EC2, EKS, ECS, IAM, Secrets Manager, CloudWatch) for scalable and secure infrastructure. - Use Infrastructure as Code (IaC) tools like Terraform or AWS CDK for environment provisioning and consistency. - Ensure compliance with healthcare data standards (FHIR R4 UK, SNOMED CT) and manage data staleness indicators, caching strategies, and performance testing. - Develop and maintain automation scripts using Python and Bash for infrastructure and application tooling. - Support authentication and authorization frameworks aligned with healthcare security and regulatory needs. - Collaborate in Agile teams using Rally (or similar project management tools); contribute to sprint planning and retrospectives. - Manage code in mono-repo environments, ensuring version control best practices and release hygiene. Qualifications Required: - 4-6 years of experience as a DevOps Engineer in complex, regulated environments. - Strong background in AWS ecosystem: EC2, ECS/EKS, IAM, Secrets Manager, CloudWatch. - Proficiency in Docker, Kubernetes, and container orchestration. - Hands-on experience with CI/CD tools like GitHub Actions, GitLab CI, Jenkins. - Experience with Terraform or AWS CDK for Infrastructure as Code. - Expertise in scripting languages such as Python and Bash for automation. - Familiarity with healthcare data standards like FHIR R4 UK, SNOMED CT, and NHS data protocols. - Exposure to MIG connectors, EMR integration, and NHS data interoperability workflows. - Understanding of performance testing, caching strategies, and data freshness indicators. - Knowledge of Agile tools (e.g., Rally) and coordination with Material UI for frontend-backend development. - Understanding of authentication/authorization mechanisms and security best practices. In addition to the above requirements, experience in NHS or healthcare compliance environments, observability tools like Prometheus, Grafana, ELK stack, and AWS certifications (e.g., Solutions Architect, DevOps Engineer) would be considered a plus. Please note that this is a contractual/temporary position with a contract length of 12 months, and the work location is in person at various locations including Noida, Bangalore, Hyderabad, Pune, and Chennai.,
ACTIVELY HIRING
posted 2 days ago
experience2 to 6 Yrs
location
Delhi
skills
  • Java
  • Spring Boot
  • Spring Framework
  • Git
  • Maven
  • REST
  • SOAP
  • Agile application development
  • JIRA
  • Jenkins
  • AWS
  • SOLID principles
  • RDBMS
  • Spring Projects
  • React JS
  • Spring Cloud
  • continuous integration environments
  • Cloudbees
  • Cloud Platforms
  • Pivotal Cloud Foundry
  • Fraud handling workflows
  • microservice design patterns
  • structured problem solving
Job Description
As a Software Engineer II at Synchrony, you will be playing a crucial role in designing and developing cutting-edge tech solutions that redefine industry standards. You will have the opportunity to work on diverse projects, from fintech to data analytics, contributing to shaping the future of technology. If you are passionate about engineering and innovation, Synchrony's Engineering Team is the perfect place for you to be. **Key Responsibilities:** - Build Microservices, MFEs & APIs leveraging cloud technologies with high quality and sound design principles - Manage DevOps pipelines for Microservices, MFEs & APIs - Analyze production defects, troubleshoot systems, identify root cause, and implement fixes - Collaborate with third-party vendors to develop software and integrate their solutions into our products - Undertake any other duties and/or special projects as assigned **Qualifications Required:** - Bachelors Degree or equivalent with a minimum of 2+ years of experience in technology. In lieu of degree, a minimum of 4+ years of experience is required. - Strong Technical Expertise in Java, Spring Projects (Spring Boot, Spring Framework), React JS, Spring Cloud, Git, Maven, REST and SOAP - Experience with Agile application development and familiarity with tools like JIRA - Proficiency in continuous integration environments and tools like Jenkins, Cloudbees - Knowledge of Cloud Platforms such as Pivotal Cloud Foundry & AWS - Understanding of Fraud handling workflows **Desired Skills/Knowledge:** - Familiarity with designing applications using SOLID principles, Java, and microservice design patterns with business acumen - Working knowledge in RDBMS - Ability to analyze, use structured problem-solving and available tools to troubleshoot systems, identify root cause, action plans, impact, and resolution options Join Synchrony's Engineering Team to be part of a collaborative environment that encourages creative problem-solving and fosters career growth. Embrace this opportunity to work on innovative projects and contribute towards the future of technology.,
ACTIVELY HIRING
posted 1 day ago
experience5 to 9 Yrs
location
Noida, Uttar Pradesh
skills
  • Release Management
  • Product Operations
  • Jira
  • Confluence
  • Git
  • Stakeholder Management
  • Change Management
  • Data Migration
  • Analytics Reporting
  • ITIL
  • AWS Cloud environments
  • AgileScrum development
  • CICD tools
  • Documentation Skills
  • Healthcare CRM systems
Job Description
As a Release Manager Engagement Platform, your role combines the technical orchestration skills of a Release Manager with the business communication and adoption readiness focus of a Product Operations Specialist. You are required to have a deep understanding of healthcare compliance considerations and manage complex dependencies among various teams while closely collaborating with the Change Board. Key Responsibilities: - Develop and maintain a comprehensive release calendar covering various releases, migrations, environment refreshes, demos, and analytics updates. - Collaborate with multiple teams to plan releases based on priorities and environment availability. - Plan release scope and timelines across multiple scrum teams and environments. - Coordinate environment usage to minimize conflicts during different events. Environment & Configuration Management: - Maintain an inventory of all environments with their purpose, versions, and scheduled changes. - Manage multi-environment configurations ensuring environment integrity. - Ensure environment refreshes and configurations comply with healthcare data security regulations. Release Execution: - Oversee the end-to-end release lifecycle ensuring readiness and validation with stakeholders. - Facilitate go/no-go meetings before deployments and manage hotfix and emergency releases. - Validate readiness with all stakeholders and ensure compliance with data privacy regulations. Documentation, Compliance & Audit Readiness: - Ensure accuracy and consistency of change control docs, known issues lists, and implementation plans. - Prepare user-facing release notes and enablement materials. - Standardize templates for release documentation to support compliance. Risk & Issue Management: - Identify and mitigate release risks related to dependencies and conflicts. - Maintain a lessons-learned repository to improve release processes continuously. Stakeholder Communication: - Be the primary point of contact for release-related queries. - Prepare clear release notes for operational and business audiences. - Provide regular release progress updates to leadership and cross-functional teams. Process & Tooling: - Implement release management best practices and optimize CI/CD pipelines. - Identify and resolve process inefficiencies in release workflows. - Utilize tools like Jira, Azure DevOps, Git, Jenkins, and ServiceNow for release tracking. Qualifications & Skills: Essential: - Bachelor's degree in computer science or related field. - 5+ years of experience in Release Management or related roles in AWS Cloud environments. - Strong understanding of Agile/Scrum development and release processes. - Proficiency in tools like Jira, Confluence, Git, Jenkins, Azure DevOps. - Excellent communication and stakeholder management skills. Preferred: - Experience with healthcare CRM systems. - Familiarity with data migration projects and analytics reporting. - Exposure to change management frameworks and healthcare compliance requirements. - ITIL or Agile/Scrum certification.,
ACTIVELY HIRING
posted 5 days ago
experience9 to 13 Yrs
location
Noida, Uttar Pradesh
skills
  • software development
  • Java
  • Cryptography
  • Web Application Development
  • J2EE
  • NoSQL
  • RDBMS
  • HTML5
  • JavaScript
  • Angular
  • Debugging
  • Analytical Skills
  • Design Patterns
  • Relational databases
  • SQL
  • Redis
  • Kafka
  • AESRSA Encryption decryption
  • React
  • Problemsolving
  • Application Server platforms
  • ELK
Job Description
As a member of the new initiatives team at AeroQube, your day-to-day responsibilities will include: - Evaluating technology and formulating solutions to solve specific technology problems - Helping others build prototypes and pass on implementation - Architecting and designing software components for modeling non-functional requirements like performance and scalability - Working in teams to achieve common deliverables and collaborating to ensure everyone collectively participates in their part - Working as an individual and delivering specific targets with a focus on great performance - Mentoring and guiding the team to scale up and develop both hard and soft skillsets - Managing conflicts arising from gaps in technical implementation and customer expectations - Coding and implementing high-performance code - Conducting design and code reviews - Working on identifying project risks and planning mitigation actions with the project manager at the project level - Participating in organizational processes like recruitments, appraisals, and certification audits Signs of success in this role include: - Ensuring the safety and security of business systems and identifying security threats - Changing legacy systems to new future systems - Venturing into new technologies to solve specific problems - Adapting to situations and problems quickly, as well as juggling multiple tasks as needed - Having a passion for coding Skills we can't do without: - Regular coding experience - 9+ years of experience in software development - Proficiency in encryption and decryption in Java with AES/RSA, including key generation, export, and transfer - Expertise in various aspects of cryptography such as Secure Random generation, Message Digest, Signature, Cipher selection and padding, Cypher modes, KeyPair Generator, Key Generator, BCrypt based Key Derivation Function BKDF, Key Agreement, etc. - Experience with certificates import/export and awareness of various certificate formats, Root Certificates, Self-signed Certificates, CA, etc. - Familiarity with OpenSSL command line tool or Microsoft tools for generating and converting certificates - A master's degree in computer science or a graduate degree in Engineering - Development experience in Web Applications using J2EE with NoSQL or RDBMS, and/or HTML5, JavaScript (Angular and/or React) - Track record of delivering mission-critical systems that directly contribute to organizational revenue - Designing scalable web applications for consumers - Proficiency in programming languages, operating systems, software design, and good coding discipline - Strong analytical and problem-solving skills, attention to detail, and excellent interpersonal and communication skills Skills we'd like to see: - Experience with Java/.Net/NodeJS, Spring, ORACLE (SQL /PL/SQL), and/or HTML5, JavaScript (Angular and/or React) - Experience with ElasticSearch/Solr/Lucene, Redis, Kafka, or MongoDB in a work project - Knowledge of agile software development and Clean Code principles - Interest in learning new technologies quickly and customer-oriented mindset - Ability to deal with change management, risk aversion, and complex technical conflicts - Strong debugging skills, analytical skills, and problem-solving abilities - Willingness to work under delivery pressure, adapt to different technologies, and explore new solutions - Hands-on experience in Java and common open-source ecosystems - Extensive experience with Application Server platforms and relational databases Benefits at AeroQube include: - Medical Insurance - Open Door Policy - Flexible Working Hours - Skills Development - Food & Beverage - Employee clubs, activities & gifts - Work-Life Balance - Financial Benefits,
ACTIVELY HIRING
posted 3 days ago

Reservoir Engineer

SAIKOR SECURITY TRAINING AND SERVICES PRIVATE LIMITED
experience5 to 10 Yrs
Salary1.5 - 12 LPA
WorkContractual
location
Noida, Delhi+8

Delhi, Bangalore, Chennai, Hyderabad, Kolkata, Gurugram, South Goa, Pune, Mumbai City

skills
  • ndt
  • testing
  • drilling engineer
  • petroleum engineer
  • instrument technician
  • mechanical engineer
  • chemical engineer
  • reservoir engineer
  • geoscientist/geologist
  • non-destructive
  • technician
  • electrical engineer
  • pipeline engineer
Job Description
A reservoir engineer analyzes and manages oil and gas reservoirs to maximize extraction efficiency. They use geological data and computer models to predict fluid flow, estimate reserves, and develop strategies for optimal production, often working with drilling and production teams. Key activities include creating field development plans, monitoring reservoir performance, and implementing recovery techniques. Key responsibilitiesData analysis: Analyze geological, seismic, and reservoir data to understand properties like pressure, temperature, and rock and fluid characteristics.Reservoir modeling: Use sophisticated software to build realistic computer models of the reservoir to simulate fluid flow and predict behavior.Production forecasting: Predict future oil and gas production rates and estimate reserves, including their uncertainties.Development planning: Develop and manage field development plans and strategies to optimize resource extraction over the life of the field.Production optimization: Monitor current production, identify opportunities for improvement, and propose and coordinate well interventions, such as workovers or new well placements.Strategy and collaboration: Work with other engineering disciplines (drilling, completions, production) to ensure that field operations are aligned with reservoir management goals.Economic evaluation: Conduct economic analyses to support development decisions and ensure that recovery strategies are profitable.Reporting: Prepare reports on reserves, production, and financial forecasts for management. Required skills and qualificationsEducation: A bachelor's degree in petroleum engineering, reservoir engineering, or a related field like chemical engineering is typically required. A master's degree is often preferred.Technical skills: Strong understanding of reservoir physics, fluid flow, and geology, along with proficiency in reservoir simulation software and data analysis tools.Soft skills: Analytical, problem-solving, and communication skills are essential, as is the ability to work effectively in a multidisciplinary team.
posted 3 weeks ago

Sales engineers

ASSETCARE SERVICES PRIVATE LIMITED
ASSETCARE SERVICES PRIVATE LIMITED
experience1 to 5 Yrs
Salary7 - 16 LPA
WorkContractual
location
Delhi, Bangalore+8

Bangalore, Chennai, Ongole, Tirupati, Nellore, Vijayawada, Hyderabad, Gurugram, Kolkata

skills
  • technical
  • analytical
  • communication
  • thinking.
  • logical
  • ability.
  • mathematical
  • skills.
Job Description
A sales engineer acts as a bridge between technical products and customer needs, combining technical expertise with sales skills to help clients find the right solutions Prepare and deliver presentations explaining technical products to existing and prospective customers. Talk with customers, the sales team, and engineers to determine product needs and system requirements. Provide customers with product sales support Pre-sales technical support: Provide technical expertise during the sales cycle, answer customer questions, and act as a liaison between the sales team and potential clients. Solution development: Work with clients to understand their specific technical and business needs and then design customized solutions using the company's products or services.   Product demonstrations: Deliver technical presentations and product demonstrations that clearly explain complex technologies and highlight their practical business benefits. Proposal creation: Develop and write proposals that outline the technical solutions, project scope, and costs for customers. Customer relationship management: Build and maintain strong relationships with customers, both new and existing, which can involve travel to client sites and industry events. Post-sales support: Assist with the implementation and integration of products, provide training to customers, and troubleshoot issues to ensure customer satisfaction. Product feedback: Act as a liaison between customers and the product development team by gathering and relaying client feedback to help improve products and services.     Technical skills. Analytical skills. Logical thinking. Mathematical ability. Communication skills.  
posted 3 weeks ago

Databricks Lead Engineer

Vy Systems Private Limited
experience8 to 13 Yrs
Salary50,000 - 3.5 LPA
WorkRemote
location
Delhi, Bangalore+5

Bangalore, Chennai, Hyderabad, Gurugram, Kolkata, Mumbai City

skills
  • aws
  • azure
  • databricks
Job Description
Job Title: Databricks LeadLocation: RemoteExperience: 8+ YearsApply: sanjai@vysystems.comAbout the RoleWe are seeking a skilled Databricks Developer with hands-on experience in Apache Spark, PySpark, AWS and data engineering on the Azure Databricks platform. The ideal candidate will design and implement scalable data pipelines, optimize data processing workflows, and contribute to building robust data solutions supporting analytics and business intelligence initiatives.Key ResponsibilitiesDesign, develop, and maintain data pipelines and ETL workflows using Databricks (PySpark, SQL, Scala, or Python).Develop scalable and optimized data processing solutions for large datasets.Work with Azure Data Lake, Delta Lake, and Azure Data Factory (ADF) to build end-to-end data solutions.Implement data transformations, aggregations, and cleansing within Databricks.Collaborate with data architects, analysts, and business stakeholders to translate requirements into technical designs.Optimize Spark jobs for performance and cost efficiency.Monitor, debug, and troubleshoot Databricks jobs and clusters.Ensure best practices for data governance, quality, and security.Contribute to CI/CD pipelines for data workflows and infrastructure automation.Document processes, workflows, and code to ensure maintainability.
logo

@ 2025 Shine.com | All Right Reserved

Connect with us:
  • LinkedIn
  • Instagram
  • Facebook
  • YouTube
  • Twitter