scripting-python-jobs-in-pune, Pune

308 Scripting Python Jobs in Pune

Toggle to save search
posted 2 months ago

MLOps engineer

CAPGEMINI TECHNOLOGY SERVICES INDIA LIMITED
CAPGEMINI TECHNOLOGY SERVICES INDIA LIMITED
experience6 to 11 Yrs
location
Pune, Bangalore+4

Bangalore, Noida, Chennai, Hyderabad, Mumbai City

skills
  • machine learning
  • python
  • mlops
Job Description
Fill below mandatory link.  MLops-virtual interview with Capgemini--all locations, Monday 27th Oct 2025 Fill out form   Position: - Machine Learning Ops Harvest Level: Software Engineer  Job Description:    Machine Learning Ops Engineer is the one whos responsible for managing and optimizing the deployment and operation of machine learning models in production environments. He/she works closely with data scientists, software engineers, and DevOps teams to ensure that machine learning models are integrated seamlessly into existing systems and can scale effectively. Role & Responsibilities Responsible for design and implementation of secure and scalable infrastructure in Azure cloud. Build and maintain CI/CD/CT pipeline across azure cloud platform for Data science projects. Own and automate the infrastructure provisioning, demand forecasting and capacity planning. Build tools and automation to improve systems availability, reliability, performance, monitoring and scalability. Setting up the alerts and monitoring for the platform. Monitoring system, application health, security controls and cost Envision, implement and rollout best MLOPs/DevOps tools and automation.  Requirements: Strong understanding of concepts related to Machine Learning, Architecture, and MLOps practices. Proficient in Azure Cloud practices and managing Azure Kubernetes Service infrastructure. Hands on experience in Python or any other scripting languages (Groovy, Shell) Experience with monitoring tools. Hands-on experience in Docker, Kubernetes. Excellent experience with source code management tools (git) Experience with modern cloud development practices such as Microservices Architecture, REST Interfaces etc. Experience in implementing Data/Model drift. Strive for Continuous Improvement and build continuous integration, continuous development, and constant deployment pipeline. Excellent Troubleshooting skills Always ready to learn more and adopt new cutting-edge technology with right value proposition.  
INTERVIEW ASSURED IN 15 MINS

Top Companies are Hiring in Your City

For Multiple Roles

Jio Platforms Ltd
Jio Platforms Ltdslide-preview-Genpact
posted 3 weeks ago
experience7 to 12 Yrs
location
Pune, Chennai
skills
  • testing
  • webdriver
  • etl testing
  • automation
  • automation testing
  • python
  • database testing
  • selenium testing
  • pytest
Job Description
We are seeking a skilled ETL & Python Automation Tester to join our QA team. The ideal candidate will have strong experience in ETL testing, data validation, and Python-based automation to ensure the integrity, accuracy, and quality of data across data pipelines, warehouses, and reporting systems. Role : Automation Testing - ETL+ Python Automation Mode : HybridLocation : Pune/Chennai Key Responsibilities Design, develop, and execute ETL test cases for data ingestion, transformation, and loading processes. Validate data movement between source, staging, and target systems to ensure end-to-end data integrity. Automate ETL and data validation tests using Python and relevant testing frameworks. Create and maintain automated scripts for regression and functional testing. Work closely with Data Engineers, Developers, and Business Analysts to understand data flows and business rules. Perform data reconciliation, transformation validation, and schema testing. Implement and maintain test frameworks using pytest, unittest, or Robot Framework. Report and track defects in JIRA (or similar tools) and work with teams to resolve them. Contribute to test strategy, planning, and continuous integration (CI/CD) processes. Required Skills & Qualifications Bachelors degree in Computer Science, Information Technology, or related field. 3+ years of experience in software testing, with a focus on ETL and data testing. Strong knowledge of SQL for data verification and validation. Hands-on experience with Python scripting for automation. Experience with ETL tools (e.g., Informatica, Talend, SSIS, DataStage, Glue, etc.). Familiarity with data warehouses (Snowflake, Redshift, BigQuery, etc.). Knowledge of automation frameworks (pytest, unittest, Robot Framework). Experience with version control systems (Git) and CI/CD pipelines (Jenkins, GitLab CI, etc.). Strong analytical and problem-solving skills.
INTERVIEW ASSURED IN 15 MINS
posted 7 days ago
experience5 to 8 Yrs
location
Pune
skills
  • domain
  • python
  • scripting languages
  • test design
  • test automation
  • c++
  • rail
Job Description
Job Description: Senior Software Test Engineer Position Overview We are looking for a skilled Senior Software Test Engineer with strong expertise in the Rail domain, capable of designing high-quality test scenarios and contributing to both manual and automated testing efforts. The ideal candidate should have hands-on experience in C++, Python or scripting languages, and developing robust test automation frameworks for safety-critical rail systems. Key Responsibilities Test Design & Planning Analyze rail system requirements and create detailed test plans, test cases, and test design documents. Apply structured test design methodologies to ensure complete coverage of functional and safety requirements. Review requirement specifications and provide feedback on testability. Test Execution & Automation Develop and maintain test automation scripts using Python or other scripting languages. Execute manual and automated test cases for embedded or control system components built in C++. Implement reusable and scalable automation frameworks tailored to rail industry needs. Defect Management & Reporting Identify, document, and track defects, collaborating with development teams for timely resolution. Perform root-cause analysis on failures and validate fixes through regression testing. Prepare detailed test reports, traceability matrices, and documentation for audits. Domain-Specific Activities Work with rail signaling, control, or onboard systems, ensuring compliance with rail safety standards. Understand system behavior, communication protocols, and operational workflows in the Rail domain. Required Skills & Qualifications Strong experience in Rail domain testing (signaling, interlocking, onboard systems, etc.). Proficiency in Test Design, including creation of test cases, scenarios, and coverage analysis. Hands-on experience with C++ for test understanding, validation, and support. Strong scripting skills in Python or equivalent scripting languages. Experience in building or maintaining test automation frameworks. Good understanding of requirement-based testing and safety-critical system testing. Familiarity with tools such as version control (Git), CI/CD pipelines, and defect tracking systems. Preferred Qualifications Experience with EN 50128/50129 rail safety standards. Exposure to communication protocols used in rail systems. Knowledge of embedded systems, real-time systems, and Linux environments. ISTQB certification or equivalent. Bachelors in Engineering in required.
INTERVIEW ASSURED IN 15 MINS
question

Are these jobs relevant for you?

posted 1 month ago
experience5 to 9 Yrs
location
Pune, Maharashtra
skills
  • Vulnerability Research
  • Code Analysis
  • Reverse Engineering
  • OWASP
  • NIST
  • Authentication Protocols
  • Network Security
  • Cryptography
  • CEH
  • ECSA
  • OSCP
  • LPT
  • Communication Skills
  • Collaboration
  • SIEM
  • IDSIPS
  • Scripting Python
  • Bash
  • Automation Tools
  • Threat Modelling
  • SANS
  • Cloud Computing Technologies
  • Ransomware Attacks
  • Certifications CISSP
  • Frameworks NIST
  • MITRE ATTCK
  • CIS Controls
  • Endpoint Protection
Job Description
As a part of the security team at Cohesity, your mission is to assist organizations worldwide in safeguarding their data and staying ahead of evolving cyber threats. Your responsibilities will include identifying vulnerabilities through code analysis and reverse engineering, documenting findings with detailed reports, and providing actionable recommendations to strengthen the security posture of target systems. Staying updated with cybersecurity trends, emerging threats, and technologies is crucial as you contribute implementation plans to enhance product security. **Role Overview:** In this role, you will be responsible for: - Penetrating applications, networks, and systems to identify vulnerabilities - Understanding product architecture and attack vectors - Finding 0-Day vulnerabilities through code analysis or reverse engineering - Performing threat modeling and categorizing threat levels - Understanding technical domains and human interaction domains related to cybersecurity - Knowing modern authentication protocols, cloud computing technologies, and network security **Key Responsibilities:** You will: - Identify vulnerabilities through code analysis and reverse engineering - Document findings with detailed reports and actionable recommendations - Contribute implementation plans to enhance product security - Perform threat modeling and categorize threat levels - Stay updated with cybersecurity trends, emerging threats, and technologies - Have knowledge of Windows, Linux, Unix internals and network security **Qualifications Required:** We would love to talk to you if you have many of the following qualifications: - Preferred Certifications: CISSP, CEH, ECSA, OSCP, LPT - Minimum 5-7 years of experience in the Security Domain - Excellent communication skills and great collaboration - Familiarity with frameworks like NIST, MITRE ATT&CK, and CIS Controls - Knowledge of scripting (Python, Bash) or automation tools for security operations If you are driven to shape the future of cybersecurity and have a passion for safeguarding data against cyber threats, Cohesity invites you to join our mission. Embrace the opportunity to work in a dynamic environment where your expertise and skills can make a difference. (Data Privacy Notice: For information on personal data processing, please see our Privacy Policy.) (Note: In-Office Expectations - Cohesity employees within a reasonable commute work out of our core offices 2-3 days a week of their choosing.),
ACTIVELY HIRING
posted 2 months ago

Architect / Senior Architect

Pivotchain Solution Technologies
experience2 to 6 Yrs
location
Pune, Maharashtra
skills
  • H264
  • HEVC
  • AAC
  • scripting languages
  • Python programming
  • streaming protocols
  • AC3 AV
  • H265
  • VP8VP9
  • AV1 codecs
  • Frontend Technologies
  • container technologies
  • database languages
  • version control practices
  • video players
  • CICD pipelines
  • FFmpegGstreamer commandline tools
  • FullStack technologies
  • Deep Learning Machine Learning
  • common programming languages
Job Description
As a hands-on technical architect, you will be leading the Video Analytics & Surveillance platform. Your responsibilities will include: - Architecting end-to-end video streaming infrastructure with a focus on optimizing and improving video codec efficiency, encoder rate control, transcode speed, video pre/post-processing, and error resilience. - Enhancing end-to-end video experiences on lossy networks across various user scenarios. You should possess the following skills: - Expert level Python programming language skills are a MUST. - Knowledge of streaming protocols RTSP, RTMP, HLS (streaming at scale), MPEG-DASH. - Working knowledge of industry standard codecs such as H.264, HEVC, AAC, AC-3 A/V, H.265, VP8/VP9, and AV1. - Familiarity with Frontend Technologies including HTML5, CSS3, Javascript, JQuery, NodeJS, and VideoJS. - Hands-on experience with container technologies Docker and Kubernetes. - Knowledge of database languages (e.g. MongoDB, SQL, No-SQL) and version control practices and tools (Git, Gitlab etc.). - Experience in dealing with video players on multiple platforms like iOS, ExoPlayer, Roku, video.js, and hls.js. - Experience in deploying applications at scale on serverless and on-prem platforms. - Experience in integrating IoT/sensors with existing solutions. - Proficiency in CI-CD pipelines, automating training/deployments. - Optimize software to run in real-time on multiple computing platforms. - Define the video optimization roadmap for both low-end and high-end networks and devices. - Proficient in FFmpeg/Gstreamer command-line tools and libraries. - Review code & mentor the team in improving the quality and efficiency of the delivery. - Experience in writing technical architecture documents. - Build tools and distributed software to automatically gather user stats to improve algorithms. Qualification and Experience: - Must have 2+ years of experience with expert-level understanding of Full-Stack technologies including some understanding of Deep Learning/ Machine Learning. - Masters degree in computer science/engineering/mathematics. - Deep knowledge of common programming languages and scripting languages, such as Python & C++. - Experience with streaming quality optimization on mobile devices. - Experience collaborating cross-functionally between groups with different video technologies and pipelines.,
ACTIVELY HIRING
posted 6 days ago
experience3 to 7 Yrs
location
Pune, Maharashtra
skills
  • Snowflake
  • SQL
  • data transformation
  • data analysis
  • data reporting
  • data governance
  • data security
  • data monitoring
  • Airflow
  • data warehousing
  • data modeling
  • clustering
  • performance tuning
  • dbt
  • Python
  • data engineering
  • data governance
  • Google Cloud Platform
  • Cloud Storage
  • GCPnative tools
  • Cloud Composer
  • partitioning
  • ELT tools
  • Fivetran
  • Cloud Data Fusion
  • scripting language
  • privacy
  • BigQuery
  • Dataflow
  • PubSub
Job Description
You will be responsible for: - Designing, developing, and maintaining data pipelines and ETL/ELT workflows using GCP-native tools and services. - Building and optimizing data warehouses using Snowflake. - Writing complex and efficient SQL queries for data transformation, analysis, and reporting. - Collaborating with analysts, data scientists, and business stakeholders to understand data needs and deliver reliable solutions. - Implementing data governance, security, and monitoring best practices across GCP projects. - Tuning queries and optimizing performance of large-scale datasets. - Automating workflows using Cloud Composer (Airflow) or similar orchestration tools. Your Skills & Qualifications should include: - 3+ years of experience in a data engineering or data platform role. - Strong hands-on experience with Snowflake data warehousing. - Expert-level skills in SQL, able to write optimized, scalable, and complex queries. - Experience with data modeling (star/snowflake schema), partitioning, clustering, and performance tuning in a data warehouse. - Familiarity with modern ELT tools such as dbt, Fivetran, or Cloud Data Fusion. - Experience in Python or similar scripting language for data engineering tasks. - Understanding of data governance, privacy, and Google Cloud Platform services, especially BigQuery, Cloud Storage, Dataflow, Pub/Sub, Composer.,
ACTIVELY HIRING
posted 2 months ago

GCP & Python Scripting

Kudos Technolabs
experience10 to 14 Yrs
location
Pune, Maharashtra
skills
  • GCP
  • GitLab
  • Google Cloud Platform
  • AWS
  • Azure
  • Helm
  • Docker
  • Kubernetes
  • Bash scripting
  • YAML
  • Python scripting
  • DevOps principles
  • CICD tools
  • Terraform
  • Datadog
  • Prometheus
  • Grafana
  • Networking concepts
Job Description
Role Overview: As an experienced professional with over 10+ years of experience, you should have a strong expertise in GCP and Python scripting. Your role will involve collaborating with product owners and engineering teams to develop Infrastructure as Code modules and application build and deployment pipelines. Additionally, you will be responsible for setting up monitoring and alerting systems for applications and providing timely solutions to any arising issues. Key Responsibilities: - Work closely with product owners and engineering teams to create Infrastructure as Code modules and application build and deployment pipelines - Assist in establishing monitoring and alerting capabilities for applications, addressing any challenges that may arise - Automate tasks wherever possible - Adapt to various situations while maintaining a positive attitude - Implement cloud-based applications with a hands-on approach - Deliver infrastructure improvements efficiently and cost-effectively - Understand the infrastructure and workflow needs of development teams - Monitor security aspects of infrastructure and address potential vulnerabilities - Collaborate with SREs and developers through pair programming sessions - Independently troubleshoot and persistently seek solutions - Interpret architectural decisions impacting infrastructure shape - Execute necessary infrastructure changes to support applications - Cooperate with cross-functional teams spread across multiple time zones - Implement logging and monitoring frameworks effectively - Lead by example with an outcomes-oriented approach Qualifications Required: - Proficiency in DevOps principles - Strong programming skills, particularly in Python - Familiarity with CI/CD tools like GitLab - Knowledge of cloud technologies such as Google Cloud Platform, AWS, and Azure - Experience with cloud infrastructure automation tools like Terraform and Helm - Understanding of container platforms like Docker and container orchestration platforms like Kubernetes - Proficiency in Bash scripting and YAML - Experience with infrastructure and application monitoring tools like Datadog, Prometheus, and Grafana - Knowledge of networking concepts and cloud-based delivery models including CIDR notations, Subnets, Firewalls, DNS, etc. Nice to have: - Familiarity with code and development practices like Test Driven Development, Behaviour Driven Development, and code reviews in GitHub and GitLab - Knowledge of modern testing techniques and frameworks for programming languages like Node.js, Go, and Python - Understanding of OAuth2 and OIDC semantics (Note: Additional details about the company were not provided in the job description.),
ACTIVELY HIRING
posted 2 weeks ago
experience10 to 14 Yrs
location
Pune, All India
skills
  • Azure
  • Python
  • Power BI
  • Snowflake
  • DevOps
  • Agile methodology
  • Pandas
  • CICD
Job Description
You will be responsible for driving the design, development, and optimization of scalable cloud-based data solutions as a Senior Data Engineer at Keyrus. Your role will involve owning the end-to-end data lifecycle, from pipeline architecture to data validation, exploratory analysis, and insight presentation. You should combine strong data engineering capabilities with the ability to collate, analyze, and interpret data to generate meaningful business value. A foundational understanding of data science concepts is expected. **Responsibilities:** - Architect, build, and optimize cloud-native data pipelines and big data ecosystems using Azure. - Perform data ingestion, transformation, and enrichment to ensure high data quality and reliability for downstream use cases. - Conduct basic exploratory data analysis (EDA) to generate insight-ready datasets for business stakeholders and analytics teams. - Present data findings and performance insights clearly to both technical and non-technical stakeholders. - Ensure that all Non-Functional Requirements (security, performance, scalability, DR, compliance) are integrated into the design and implementation. - Evaluate and prototype modern data technologies, leading POCs to drive innovation and continuous improvement. - Collaborate with Data Governance teams to ensure accurate metadata and lineage tracking. - Advocate for data mesh principles, reusable data products, and best practices across engineering and business teams. - Utilize DevSecOps and automation tools (Terraform, GitHub, CI/CD) for streamlined deployment and operational excellence. **Qualifications Required:** - 10+ years of hands-on experience in data engineering in enterprise/cloud environments. - Proven experience building data pipelines on Azure (Azure Data Factory, Databricks, Synapse, or equivalent). - High proficiency in Python and analytical scripting (Pandas, data inspection, validation scripts). - Experience in data joins, validation, quality checks, and exploratory insights aggregation. - Understanding of data migration processes and working with distributed client/service data across multiple sources. - Strong understanding of data lake/data warehouse architectures and modern data platform patterns. - Ability to build Power BI dashboards (experience with Tableau or another BI tool is also acceptable). - Previous experience in Snowflake development is a plus. - Skills in DevOps, CI/CD, and Agile methodology. - Ability to translate complex technical concepts into clear business-relevant insights. - Experience collaborating with multidisciplinary teams, including Data Scientists, Analysts, Governance, and Product stakeholders. Joining Keyrus as a Senior Data Engineer means becoming a part of a market leader in the Data Intelligence field and an international player in Management Consultancy and Digital Experience. You will have the opportunity to work with a network of thought-leading professionals and showcase your talents in a dynamic atmosphere. Keyrus offers competitive benefits, including a comprehensive Private Medical Plan, Flexible working patterns, and Training & Development opportunities via KLX (Keyrus Learning Experience). You will be responsible for driving the design, development, and optimization of scalable cloud-based data solutions as a Senior Data Engineer at Keyrus. Your role will involve owning the end-to-end data lifecycle, from pipeline architecture to data validation, exploratory analysis, and insight presentation. You should combine strong data engineering capabilities with the ability to collate, analyze, and interpret data to generate meaningful business value. A foundational understanding of data science concepts is expected. **Responsibilities:** - Architect, build, and optimize cloud-native data pipelines and big data ecosystems using Azure. - Perform data ingestion, transformation, and enrichment to ensure high data quality and reliability for downstream use cases. - Conduct basic exploratory data analysis (EDA) to generate insight-ready datasets for business stakeholders and analytics teams. - Present data findings and performance insights clearly to both technical and non-technical stakeholders. - Ensure that all Non-Functional Requirements (security, performance, scalability, DR, compliance) are integrated into the design and implementation. - Evaluate and prototype modern data technologies, leading POCs to drive innovation and continuous improvement. - Collaborate with Data Governance teams to ensure accurate metadata and lineage tracking. - Advocate for data mesh principles, reusable data products, and best practices across engineering and business teams. - Utilize DevSecOps and automation tools (Terraform, GitHub, CI/CD) for streamlined deployment and operational excellence. **Qualifications Required:** - 10+ years of hands-on experience in data engineering in enterprise/cloud environments. - Proven experience building data pipelines on Azure (Azure Data Fac
ACTIVELY HIRING
posted 3 weeks ago
experience12 to 16 Yrs
location
Pune, Maharashtra
skills
  • Oracle
  • Postgres
  • MySQL
  • Database Management
  • Database Administration
  • Scripting
  • Python
  • Perl
  • Bash
  • YAML
  • Linux
  • AWS
  • Project Management
  • Jira
  • GIT
  • Troubleshooting
  • Coordination
  • Time Management
  • RDS
Job Description
As a Data Engineer at NiCE, you will be responsible for managing Oracle, Postgres, MySQLRDS databases. You will collaborate with DBAs, DevOps, Engineering, and SRE teams to develop database strategies, frameworks, and best practices, focusing on infrastructure as code. High availability, reliability, and performance are top priorities, requiring you to rapidly understand the product portfolio and ensure platform reliability. Strong communication and documentation skills are essential as you work in a fast-paced global and remote team environment, providing 24x7x365 support on rotation. **Key Responsibilities:** - Be a knowledge expert on critical database activities like replication, failover, archiving, performance tuning, scaling, and security. - Refactor legacy database processes to enhance design and scalability for data integrity, client migration, and schema changes in collaboration with Engineering teams. - Tighten database alerting procedures to achieve 100% platform reliability. - Provide guidance and support to Application Developers and QA teams on database-related projects. - Plan and recommend capacity and resilience improvements with a focus on reliability. - Collaborate with Performance/Engineering teams on application SQL tuning and optimization. - Create and publicize excellent documentation. - Proactively make recommendations for Database improvements. - Evaluate and analyze complex platform issues, resolving and preventing critical incidents. **Qualifications Required:** - Bachelor's (mandatory) or Master's (preferred) Degree in Computer Science, Engineering, Software Engineering, or a relevant field. - 12+ years of experience with Oracle/Postgres Database Management and Administration. - Proficiency in scripting languages like Python, Perl, Bash, and YAML for automating database maintenance. - Experience with Linux-based infrastructure, Linux/Unix administration, and AWS. - Familiarity with project management and workflow tools like Jira, GIT, etc. - Experience in 24/7 support, backup, and recovery practices. - Strong troubleshooting skills to identify and resolve issues proactively. - Exceptional coordination skills to work effectively across global teams. - Time and project management skills with the ability to prioritize tasks. **About NiCE:** NICELtd. is a global software company (NASDAQ: NICE) serving over 25,000 businesses worldwide, including 85 Fortune 100 corporations. With a focus on customer experience, financial crime prevention, and public safety, NiCE manages over 120 million customer interactions daily and monitors 3+ billion financial transactions. Recognized for innovation in AI, cloud, and digital technologies, NiCE has a team of 8,500+ employees across 30+ countries, offering endless career opportunities for passionate and innovative individuals.,
ACTIVELY HIRING
posted 1 month ago

SAS

Cognizant
experience6 to 10 Yrs
location
Pune, Maharashtra
skills
  • SAS
  • Statistics
  • EDA
  • Machine Learning
  • Alteryx
  • MS Excel
  • Python
  • SQL Scripting
  • MS PowerPoint
  • LS Salesforce
  • Incentive Compensation Management
  • Dataiku
  • ROI Calculation
  • Retail Demand Forecasting
  • R Statistical Package
Job Description
As a Lead Data Analyst at our organization in Pune, you will be instrumental in driving data-driven decision-making processes. Your role will involve leveraging your expertise in data analysis and statistical techniques to provide insights that enhance business performance, with a focus on retail demand forecasting. **Key Responsibilities:** - Lead the analysis of complex data sets to identify trends and patterns that inform business strategies. - Oversee the development and implementation of data models using SAS and R Statistical Package. - Provide insights through exploratory data analysis to support decision-making processes. - Utilize LS Salesforce and Dataiku to streamline data workflows and enhance data accessibility. - Conduct incentive compensation management analysis to optimize sales performance. - Apply machine learning techniques to improve predictive accuracy in demand forecasting. - Calculate ROI to assess the effectiveness of business initiatives and strategies. - Employ Alteryx for data preparation and blending to ensure data quality and consistency. - Develop and maintain SQL scripts for data extraction and manipulation. - Use Python for advanced data analysis and automation of repetitive tasks. - Create comprehensive reports and presentations using MS Excel and MS PowerPoint to communicate findings to stakeholders. - Collaborate with cross-functional teams to align data analysis efforts with business objectives. - Ensure data integrity and security in all analytical processes. **Qualifications:** - Possess a strong technical background with mandatory skills in SAS, LS Salesforce, Statistics, EDA, Incentive Compensation Management, Dataiku, Machine Learning, ROI Calculation, Alteryx, Retail Demand Forecasting, R Statistical Package, MS Excel, Python, SQL Scripting, and MS PowerPoint. - Demonstrate excellent analytical and problem-solving skills with a focus on actionable insights. - Exhibit proficiency in English for effective communication and collaboration. - Have domain experience in Finance & Accounting as a nice-to-have skill. - Show ability to work in a hybrid model with a focus on day shifts. - Display strong organizational skills and attention to detail in managing data projects.,
ACTIVELY HIRING
posted 1 week ago
experience3 to 7 Yrs
location
Pune, Maharashtra
skills
  • Kubernetes
  • Docker
  • Bash scripting
  • Ansible
  • AWS
  • Azure
  • GCP
  • Linux systems administration
  • Python scripting
Job Description
As an experienced Red Hat OpenShift Administrator, your primary role will involve designing, deploying, and maintaining Red Hat OpenShift Container Platform environments. Your responsibilities will include: - Overseeing the administration of Red Hat OpenShift Container Platform. - Managing OpenShift clusters using the command-line interface and web console. - Customizing cluster configurations to align with specific application needs. Additionally, you will be tasked with: - Deploying and configuring Red Hat OpenShift clusters following industry best practices. - Setting up and managing cluster authentication, Role-Based Access Control (RBAC), and security policies. - Configuring networking components, pod scheduling, and cluster scaling to ensure optimal performance. - Connecting Kubernetes workloads to appropriate storage for application data. Your role will also involve: - Conducting regular system checks and updates to maintain peak cluster performance. - Troubleshooting and resolving issues related to cluster operations and application deployments. - Implementing high availability and disaster recovery solutions for OpenShift clusters. - Monitoring application health and performance to ensure a seamless user experience. In terms of security and compliance, you will be expected to: - Apply security patches, manage access controls, and implement network policies to secure OpenShift environments. - Ensure compliance with industry security standards and regulations. - Develop and implement backup and recovery procedures for OpenShift clusters. Qualifications: - Education: Bachelors degree in Computer Science, Information Technology, or a related field (or equivalent work experience). - Certifications: Red Hat Certified Engineer (RHCE) or Red Hat Certified System Administrator (RHCSA) is essential. Red Hat Certified OpenShift Administrator (EX180/280) is a plus. - Experience: Proven experience in managing Red Hat OpenShift Container Platform. Technical Skills Required: - Strong knowledge of Linux systems administration. - Proficiency with Kubernetes and container technologies (e.g., Docker). - Scripting skills (e.g., Bash, Python) and familiarity with automation tools like Ansible. - Knowledge of cloud platforms (AWS, Azure, GCP) is highly desirable.,
ACTIVELY HIRING
posted 2 weeks ago
experience4 to 8 Yrs
location
Pune, Maharashtra
skills
  • Power BI
  • Python
  • Data Analysis
  • Data Visualization
  • SQL
  • Exploratory Data Analysis
  • Communication Skills
  • Statistical Methods
  • ProblemSolving
Job Description
Your Role and Impact As a Python Developer/Data Analyst, you will play a crucial role in utilizing data analytics to drive informed decision-making within the organization. Your contributions will involve data analysis, visualization, management, scripting, automation, collaboration, reporting, and continuous learning. By leveraging your skills in Python, Power BI, SQL, and data analysis techniques, you will extract valuable insights from complex datasets and present them in a clear and concise manner to stakeholders. Key Responsibilities: - Collaborate with cross-functional teams to gather and understand data requirements. - Perform exploratory data analysis to identify trends, patterns, and anomalies. - Design, develop, and maintain interactive Power BI dashboards and reports for effective communication of insights. - Utilize Python for data manipulation, analysis, and visualization tasks. - Support the integration of data from various sources, ensuring data quality and accuracy. - Present findings and recommendations to stakeholders in a clear and concise manner. Qualification Required: - Minimum of 4 years and a maximum of 6 years of experience in data analytics, with a strong focus on Power BI. - Proficiency in Python for data analysis and visualization. - Solid experience in exploratory data analysis techniques. - Strong analytical and problem-solving skills with attention to detail. - Excellent communication skills to convey complex information to non-technical stakeholders. Additional Details: The preferred qualifications include a Bachelor's degree in Data Science, Statistics, Computer Science, or a related field. Previous experience in a similar role within a financial services or insurance environment is considered advantageous. This role offers opportunities for continuous learning and professional development to stay updated on industry trends and enhance your skill set.,
ACTIVELY HIRING
posted 1 month ago

Senior Python Developer

Acclivis Technologies Pvt Ltd
experience5 to 9 Yrs
location
Pune, Maharashtra
skills
  • Python
  • Jenkins
  • Bash scripting
  • Git
  • Docker
  • Yocto build systems
  • CICD
  • GitLab CI
Job Description
You will be joining a cutting-edge automotive software team to work on next-generation embedded automation and CI/CD integration projects. We are seeking a passionate and experienced professional with a strong expertise in Python and Yocto build systems. **Key Responsibilities:** - Develop and maintain Python-based tools, scripts, and automation frameworks. - Manage and optimize Yocto build environments for embedded Linux platforms. - Implement and maintain CI/CD pipelines (Jenkins, GitLab CI, etc.). - Collaborate with cross-functional teams to integrate and deploy software efficiently. - Work within Linux-based environments using Bash scripting and Git version control. **Required Skills:** - Strong expertise in Python programming (tooling, automation, backend). - Hands-on experience with Yocto build system (mandatory). - Good understanding of CI/CD process and DevOps tools. - Proficiency with Linux, Bash scripting, Docker, and Git. **Good To Have:** - Exposure to C/C++ or Embedded system integration. - Experience in Automotive or Functional Safety domains.,
ACTIVELY HIRING
posted 2 months ago
experience5 to 9 Yrs
location
Pune, Maharashtra
skills
  • SAS
  • Machine Learning
  • Alteryx
  • MS Excel
  • Python
  • SQL Scripting
  • MS PowerPoint
  • R Statistical Package
  • LS Salesforce
  • Dataiku
  • Incentive Compensation Management
  • ROI Calculation
  • Retail Demand Forecasting
Job Description
As a Lead Data Analyst, you will play a crucial role in driving data-driven decision-making processes within the organization. Leveraging your expertise in data analysis and statistical techniques, you will provide insights that enhance business performance. This role requires a strong technical background in various data analysis tools and a keen understanding of retail demand forecasting. Responsibilities: - Lead the analysis of complex data sets to identify trends and patterns that inform business strategies. - Oversee the development and implementation of data models using SAS and R Statistical Package. - Provide insights through exploratory data analysis to support decision-making processes. - Utilize LS Salesforce and Dataiku to streamline data workflows and enhance data accessibility. - Conduct incentive compensation management analysis to optimize sales performance. - Apply machine learning techniques to improve predictive accuracy in demand forecasting. - Calculate ROI to assess the effectiveness of business initiatives and strategies. - Employ Alteryx for data preparation and blending to ensure data quality and consistency. - Develop and maintain SQL scripts for data extraction and manipulation. - Use Python for advanced data analysis and automation of repetitive tasks. - Create comprehensive reports and presentations using MS Excel and MS PowerPoint to communicate findings to stakeholders. - Collaborate with cross-functional teams to align data analysis efforts with business objectives. - Ensure data integrity and security in all analytical processes. Qualifications: - Possess a strong technical background with mandatory skills in SAS, LS Salesforce, Statistics, EDA, Incentive Compensation Management, Dataiku, Machine Learning, ROI Calculation, Alteryx, Retail Demand Forecasting, R Statistical Package, MS Excel, Python, SQL Scripting, and MS PowerPoint. - Demonstrate excellent analytical and problem-solving skills with a focus on actionable insights. - Exhibit proficiency in English for effective communication and collaboration. - Have domain experience in Finance & Accounting as a nice-to-have skill. - Show ability to work in a hybrid model with a focus on day shifts. - Display strong organizational skills and attention to detail in managing data projects.,
ACTIVELY HIRING
posted 2 weeks ago
experience5 to 9 Yrs
location
Pune, Maharashtra
skills
  • QA Automation
  • Manual Testing
  • DevOps
  • Jenkins
  • Git
  • Communication Skills
  • Selenium
  • AWS
  • Azure
  • GCP
  • Performance Testing Tools
  • BDD
  • Python Programming
  • Testing Lifecycle
  • Test Automation Development
  • CICD
  • API Test Automation
  • XML Verification
  • UI Test Automation
  • UFT
  • Jenkins Pipeline Scripting
Job Description
As an Infrastructure Engineer at Barclays, your role involves enhancing existing processes, reporting, and controls to ensure flawless execution of BAU. You will drive efficiencies and process improvements, standardizing processes across SBUs where possible. Your responsibilities include: - Strong experience in QA Automation and manual testing - Good understanding of BDD with hands-on experience in test automation development - Proficiency in Python programming - Extensive knowledge of testing lifecycle and core testing processes/techniques - Ability to provide estimations, design test plans and strategies, write test scenarios and cases, execute test cases, report bugs, and verify bugs - Awareness on CI/CD and DevOps tools like Jenkins and Gitlab - Experience with Git source code control system - Excellent communication skills, both verbal and written, essential for global team collaboration - Identifying processes and tools to optimize and enhance testing quality - Familiarity with API test automation, XML verification, and UI test automation tools like Selenium and UFT Additional highly valued skills may include knowledge of AWS, Azure, or GCP, Jenkins pipeline scripting, and performance testing tools like Locust. The purpose of your role is to apply software engineering techniques, automation, and best practices in incident response to ensure reliability, availability, and scalability of systems, platforms, and technology. Your accountabilities include: - Ensuring availability, performance, and scalability of systems and services through proactive monitoring, maintenance, and capacity planning - Resolving system outages and disruptions, implementing preventive measures, and automating operational processes - Monitoring and optimizing system performance, addressing bottlenecks, and implementing best practices for performance tuning - Collaborating with development teams to integrate reliability and scalability best practices - Staying informed of industry technology trends and contributing to the organization's technology communities As an Analyst, you are expected to perform activities in a timely manner with high standards, driving continuous improvement. You should have in-depth technical knowledge and experience in your area of expertise, lead and supervise a team, guide professional development, and coordinate team resources. If the position involves leadership responsibilities, demonstrate clear leadership behaviours. For individual contributors, develop technical expertise, act as an advisor, and impact related teams within the area. You will partner with other functions, take responsibility for end results of operational activities, escalate breaches of policies, and influence decision-making. You will manage risk, strengthen controls, and align work with relevant rules and regulations. Demonstrate understanding of how your sub-function integrates with the function and contribute to achieving organizational objectives. All colleagues at Barclays are expected to demonstrate the Barclays Values of Respect, Integrity, Service, Excellence, and Stewardship, as well as the Barclays Mindset to Empower, Challenge, and Drive.,
ACTIVELY HIRING
posted 5 days ago
experience3 to 7 Yrs
location
Pune, Maharashtra
skills
  • Snowflake
  • SQL
  • data transformation
  • data analysis
  • data reporting
  • data governance
  • data security
  • data monitoring
  • Airflow
  • data warehousing
  • data modeling
  • clustering
  • performance tuning
  • dbt
  • Python
  • data engineering
  • data governance
  • Google Cloud Platform
  • Cloud Storage
  • GCPnative tools
  • Cloud Composer
  • partitioning
  • ELT tools
  • Fivetran
  • Cloud Data Fusion
  • scripting language
  • privacy
  • BigQuery
  • Dataflow
  • PubSub
Job Description
Role Overview: As a Data Engineer, your role will involve designing, developing, and maintaining data pipelines and ETL/ELT workflows using GCP-native tools and services. You will be responsible for building and optimizing data warehouses using Snowflake and writing complex SQL queries for data transformation, analysis, and reporting. Collaboration with analysts, data scientists, and business stakeholders to understand data needs and deliver reliable solutions will be a key part of your responsibilities. Additionally, you will be implementing data governance, security, and monitoring best practices across GCP projects and tuning queries to optimize performance of large-scale datasets. Automation of workflows using Cloud Composer (Airflow) or similar orchestration tools will also be a part of your tasks. Key Responsibilities: - Design, develop, and maintain data pipelines and ETL/ELT workflows using GCP-native tools and services. - Build and optimize data warehouses using Snowflake. - Write complex and efficient SQL queries for data transformation, analysis, and reporting. - Collaborate with analysts, data scientists, and business stakeholders to understand data needs and deliver reliable solutions. - Implement data governance, security, and monitoring best practices across GCP projects. - Tune queries and optimize performance of large-scale datasets. - Automate workflows using Cloud Composer (Airflow) or similar orchestration tools. Qualification Required: - 3+ years of experience in a data engineering or data platform role. - Strong hands-on experience with Snowflake data warehousing. - Expert-level skills in SQL, able to write optimized, scalable, and complex queries. - Experience with data modeling (star/snowflake schema), partitioning, clustering, and performance tuning in a data warehouse. - Familiarity with modern ELT tools such as dbt, Fivetran, or Cloud Data Fusion. - Experience in Python or similar scripting language for data engineering tasks. - Understanding of data governance, privacy, and Google Cloud Platform services, especially BigQuery, Cloud Storage, Dataflow, Pub/Sub, Composer.,
ACTIVELY HIRING
posted 1 week ago

Senior DevOps Engineer

PeerLogic Systems
experience5 to 9 Yrs
location
Pune, Maharashtra
skills
  • MySQL
  • bash scripting
  • Python
  • Perl
  • Ruby
  • Kafka
  • wireshark
  • tcpdump
  • Java DevOps
  • Linux platform
  • HA Proxy
  • Zookeeper
  • NGINX configuration
  • init scripts
Job Description
Job Description: As a member of the build and release team, your role involves building, deploying, monitoring, debugging, and proactively fixing issues related to complex Java applications in a Linux environment. Roles and Responsibilities: - Own the build and installation of applications, collaborating with the US team. - Lead and manage a team of engineers locally. - Troubleshoot complex technical problems escalated from QA. - Track bugs and gather specific technical information to assist development team(s). Required Skills: - 5+ years of experience in Java DevOps and Linux platform. - Expertise in setting up Java-based environments, installing, and debugging Java WARs in various run environments such as JBoss, Apache Tomcat, etc. - Strong experience working with MySQL and HA Proxy. - Proficiency in bash scripting and other scripting languages like python, perl, or ruby. - Experience with Kafka, Zookeeper, and NGINX configuration. - Knowledge of init scripts on the Linux platform. - Good debugging, problem-solving, and analytical skills. - Ability to perform network traffic analysis and debugging using tools like wireshark and tcpdump. - Excellent written and verbal communication skills. Preferred Skills: - Familiarity with tools like Chef, Puppet, Nagios, and Cacti. - Working experience with Perforce and JIRA would be advantageous. - Experience in working with open-source projects. Please note that the company details were not provided in the job description.,
ACTIVELY HIRING
posted 1 week ago
experience4 to 8 Yrs
location
Pune, Maharashtra
skills
  • Python
  • Docker
  • Git
  • NumPy
  • Matplotlib
  • Jenkins
  • SQL
  • MariaDB
  • Azure
  • data structures
  • algorithm optimization
  • Azure Cloud development
  • DevOps practices
  • Pandas
  • GitLab CI
  • RESTful API design
  • InfluxDB
  • Linux environments
  • CICD pipelines
  • DockerKubernetes
  • asynchronous programming
  • unit testing frameworks
Job Description
Role Overview: As a Python Engineer at ENGIE India, you will be responsible for designing, developing, and maintaining scalable, high-performance applications and services using Python. You will collaborate closely with cross-functional teams to understand requirements, build efficient backend solutions, integrate APIs, and ensure system reliability and maintainability. Additionally, you will contribute to code reviews, optimization efforts, and continuous improvement initiatives to enhance overall software quality and team productivity. Your expertise in Python and familiarity with Azure Cloud development and integration will be crucial for the success in this role. Key Responsibilities: - Design and develop scalable and efficient server-side features using Python. - Collaborate with SCADA developers to integrate user-facing elements with server-side scripting. - Implement continuous integration and deployment processes using tools such as Jenkins and GitLab CI. - Conduct code reviews and implement optimizations for code quality, maintainability, and efficiency. - Work closely with cross-functional teams, including data scientists, and product managers, to deliver integrated solutions. - Communicate effectively with team members and stakeholders to gather requirements and provide updates on project progress. - Develop and execute test cases and test scripts (manual and automated) based on product requirements. - Identify, document, and track bugs using standard defect tracking tools (e.g., DevOps). Qualifications Required: - BE/B Tech/MCA qualification. - Minimum of 4+ years of experience. - Strong proficiency in Python and solid understanding of object-oriented programming (OOP) and design patterns. - Experience in RESTful API design, development, and integration. - Good knowledge of database systems (SQL and databases such as InfluxDB, MariaDB). - Familiarity with version control systems (e.g., Git, GitHub, Azure DevOps). - Experience with unit testing frameworks (e.g., pytest, unittest). - Working knowledge of Linux environments, CI/CD pipelines, and containerization (Docker/Kubernetes preferred). - Exposure to cloud platforms (Azure) is a plus. - Knowledge of asynchronous programming, data structures, and algorithm optimization is desirable.,
ACTIVELY HIRING
posted 1 week ago

Engineer, Analyst

Deutsche Bank
experience3 to 7 Yrs
location
Pune, Maharashtra
skills
  • Java
  • Oracle
  • ETL
  • Agile process
  • Test Driven Development
  • BI Analytic Tools
  • Python scripting
  • Database solutions
Job Description
Role Overview: You will be working as an Engineer, Analyst at Deutsche Bank's Pune location. As part of the dbSleuth Team within Regulatory & Cross Product IT, you will be involved in developing and delivering significant components of engineering solutions to meet complex business goals. Your role will require engagement with business stakeholders and collaboration within an innovative team to make a meaningful impact. The team focuses on delivering Trader and Counterparty surveillance solutions across all business sections of Deutsche Bank, emphasizing high-quality architecture, design, and code. Key Responsibilities: - Participate or lead in gathering business requirements and contribute to the development of functional design - Develop source code for software components using Java, Oracle, and ETL - Design and implement automated testing scripts - Integrate software components following the integration strategy and verify through unit and integrated software testing - Support release deployments into controlled environments - Capture Build and Deployment notes and develop automated deployment scripts as necessary Qualifications Required: - 3+ years of technical experience - Database experience in Oracle or SQL server, with preference for both - Basic level proficiency in Oracle PLSQL programming and performance tuning - Experience with ETL Tools and BI & Analytic Tools beneficial - Knowledge of Data Structures, Design Patterns & Frameworks - Recommended experience in Python scripting, Agile processes, Test Driven Development, and testing database solutions Company Details: Deutsche Bank's Technology, Data and Innovation (TDI) strategy focuses on enhancing engineering expertise, adopting an agile delivery model, and modernizing IT infrastructure. The bank invests in visionary tech talent and offers benefits like leave policies, parental leaves, childcare assistance, sponsorship for certifications, and comprehensive insurance coverage. The culture promotes continuous learning, career development, and collaboration among team members to excel together. Visit the company website for more information: [Deutsche Bank Company Website](https://www.db.com/company/company.htm),
ACTIVELY HIRING
posted 2 months ago
experience2 to 6 Yrs
location
Pune, Maharashtra
skills
  • Process Automation
  • Scripting
  • Python
  • SQL
  • Java
  • R
  • MS Excel
  • PowerPoint
  • Power BI
  • Data Analysis
  • Data Extraction
  • Data Cleaning
  • Data Preparation
  • Data Validation
  • Data Reporting
  • Analytical Skills
  • Soft Skills
  • Communication Skills
  • Teamwork
  • Workflow Development
  • Power Automate
  • Power Query
  • Cloud Platforms
  • ProblemSolving Skills
Job Description
Role Overview: As an Automation & Process Efficiency Specialist at DWS Product Division, you will be responsible for designing, implementing, and optimizing process automation solutions. Your role will involve leveraging technical expertise to streamline workflows, enhance operational efficiency, and drive digital transformation initiatives. You will play a key role in building and maintaining automated workflows, optimizing scripts and applications, collaborating with cross-functional teams, ensuring compliance with governance requirements, and providing technical guidance and training to team members. Key Responsibilities: - Design, develop, and maintain automated workflows using Power Automate, Power Query, and other relevant automation tools. - Build and optimize scripts and applications using major programming languages such as Python, SQL, Java, or R to automate data processing and business processes. - Collaborate with IT, data management, and business teams to identify automation opportunities and translate requirements into technical solutions. - Ensure robust documentation, version control, and testing of all automation solutions. - Monitor, troubleshoot, and continuously improve existing automated processes. - Support compliance with governance, risk, and regulatory requirements related to process automation. - Provide technical guidance and training to team members on automation best practices and tools. Qualifications Required: - Bachelor's degree (or equivalent) in Computer Science, Engineering, Information Technology, or a related field. - 2+ years of hands-on experience in process automation, scripting, or workflow development within a corporate environment. - Proficiency in coding with at least one major programming language (e.g., Python, SQL, Java, R). - Advanced knowledge of using MS Excel, PowerPoint, Word, Power BI, Power Automate, and Power Query. - Familiarity with cloud platforms is preferred. - Practical Data Analysis skills including data extracting, cleaning, preparation, validation, and reporting. - Strong analytical and problem-solving skills with the ability to work independently on technical challenges. - Soft skills including the ability to communicate technical concepts to non-technical stakeholders and strong collaboration skills. - Excellent English written and verbal skills (German skills are an added benefit). Note: The job description did not include any additional details about the company.,
ACTIVELY HIRING
logo

@ 2025 Shine.com | All Right Reserved

Connect with us:
  • LinkedIn
  • Instagram
  • Facebook
  • YouTube
  • Twitter