snowflake-jobs-in-gurgaon, Gurgaon

82 Snowflake Jobs in Gurgaon

Toggle to save search
posted 1 week ago

Senior Snowflake Data Engineer

Hucon Solutions India Pvt.Ltd.
Hucon Solutions India Pvt.Ltd.
experience8 to 13 Yrs
Salary8 - 18 LPA
location
Gurugram, Delhi+7

Delhi, Noida, Bangalore, Chennai, Hyderabad, Kolkata, Pune, Mumbai City

skills
  • aws
  • sql
  • data modeling
  • snowflake
  • data build tool
  • five tran
Job Description
Senior Snowflake Data Engineer Location: PAN India Experience: 8+ Years Skills: Snowflake, dbt, FiveTran, Snowpipe, AWS (MWAA, S3, Lambda), GitHub CI/CD Job Description We are looking for an experienced Senior Snowflake Data Engineer with strong expertise in modern data warehousing, cloud technologies, and ELT pipeline development. The ideal candidate should have deep hands-on experience in Snowflake, dbt, cloud environments, and CI/CD practices, with the ability to design scalable and efficient data solutions. Key Responsibilities Analyze, integrate, model, and interpret large and complex datasets from multiple sources. Design and implement ELT data pipelines using dbt with Snowflake as the primary cloud data warehouse. Build efficient and scalable data transformation pipelines using dbt at an advanced level. Work with ETL/ELT and data governance tools such as FiveTran and Alation. Utilize advanced Snowflake features such as RBAC, Dynamic Tables, and various optimization techniques. Ensure strong data modelling and warehousing practices across diverse database technologies. Manage orchestrations using Apache Airflow or AWS MWAA, along with CI/CD pipelines. Oversee continuous deployment, monitoring, and operations of data solutions using GitHub Actions, Terraform, and other DevOps tools. Collaborate with technical and non-technical stakeholders through effective communication. Contribute to building future-state Data Warehouse capabilities using cutting-edge technologies. Adapt and work efficiently within Agile methodology. Preferred Qualifications 2+ years of hands-on experience with Snowflake as a Cloud Data Warehouse and Data Lake platform. Strong understanding of cloud environments, especially AWS (S3, Lambda, MWAA). Airline industry domain experience is a plus.
INTERVIEW ASSURED IN 15 MINS

Top Companies are Hiring in Your City

For Multiple Roles

Jio Platforms Ltd
Jio Platforms Ltdslide-preview-Genpact
posted 2 months ago

Snowflake Developer

Viraaj HR Solutions Private Limited
experience3 to 7 Yrs
location
Delhi
skills
  • cloud computing
  • data modeling
  • etl tools
  • analytical skills
  • data integration
  • azure
  • data governance
  • snowflake
  • metadata management
  • troubleshooting
  • communication
  • team collaboration
  • data warehousing
  • performance tuning
  • sql proficiency
  • cloud platforms aws
  • google cloud
  • problemsolving
Job Description
Role Overview: Viraaj HR Solutions, a dynamic HR consultancy, is seeking a skilled Snowflake Developer to design, develop, and implement data solutions using the Snowflake platform. As part of the team, you will optimize and manage Snowflake databases, develop data integration workflows, and ensure data quality and integrity across all data stores. Your role will involve collaborating with data scientists, conducting system performance tuning, and staying updated on Snowflake features and best practices. Key Responsibilities: - Design, develop, and implement data solutions using the Snowflake platform. - Optimize and manage Snowflake databases and schemas for efficient data access. - Develop data integration workflows using ETL tools with Snowflake. - Implement and maintain data models to support business analytics. - Prepare and optimize SQL queries for performance improvements. - Collaborate with data scientists and analysts to address data-related issues. - Monitor Snowflake usage and provide recommendations for resource management. - Ensure data quality and integrity across all data stores. - Document data architecture and design processes for future reference. - Conduct system performance tuning and troubleshooting of the Snowflake environment. - Integrate Snowflake with cloud-based services and tools as required. - Participate in code reviews and provide constructive feedback. - Stay updated on Snowflake features and industry best practices. - Assist in training team members on Snowflake capabilities and tools. - Work closely with stakeholders to gather requirements and define project scope. Qualifications: - Bachelor's degree in Computer Science, Information Technology, or a related field. - At least 3 years of experience in data engineering or database development. - Strong expertise in Snowflake platform; certification is a plus. - Proficient in SQL with hands-on experience in complex queries. - Experience with ETL tools and data integration techniques. - Understanding of data warehousing concepts and best practices. - Familiarity with cloud platforms like AWS, Azure, or Google Cloud. - Ability to troubleshoot and solve data-related issues. - Excellent analytical and problem-solving skills. - Strong communication and team collaboration skills. - Experience in data modeling and metadata management. - Ability to work independently and manage multiple tasks effectively. - Knowledge of additional programming languages is a plus. - Understanding of data governance practices. - Willingness to learn and adapt to new technologies.,
ACTIVELY HIRING
posted 2 months ago
experience3 to 8 Yrs
location
Noida, Uttar Pradesh
skills
  • SQL
  • Python
  • ETL
  • Data warehousing
  • MongoDB
  • MySQL
  • APIs
  • Hadoop
  • HDFS
  • Hive
  • Spark
  • ETL tools
  • DBT
  • Azure Data Services
  • Databricks
  • PySpark
  • Nodejs
  • Snowflake Datawarehouse
  • Bigdata technologies
  • Matillion
Job Description
You will be responsible for supporting data and analytics solutions across projects as an Azure Data Engineer with 3 to 8 years of experience. Collaborating with various projects, you will create data solutions for a variety of business use cases. Your role will involve driving innovation within Data Engineering to shape the future of data and analytics technology. As a critical team member, you will design and develop highly complex data projects, identifying gaps in the data stack and guiding advancements for the team. You will provide technical expertise in all project phases, including analysis, design, development, and implementation of cutting-edge solutions. Your input will continuously shape and improve the data environment beyond the team boundaries. Key Responsibilities: - Collaborate with various projects to create data solutions for different business use cases - Drive innovation within Data Engineering to influence the future of data and analytics technology - Design and develop highly complex data projects - Identify gaps and weaknesses in the data stack, guiding advancements for the team - Provide technical expertise in all project phases, including analysis, design, development, and implementation - Shape and improve the data environment beyond team boundaries - Understand UI needs, backend design, and data design to create backend code - Run diagnostic tests, repair defects, and provide technical support - Document Node.js processes and prepare reports - Design and implement APIs, integrating with third-party APIs as needed - Develop and maintain databases using technologies such as MongoDB or MySQL - Create reusable code and libraries for future use - Optimize applications for maximum speed and scalability - Troubleshoot and debug issues, providing timely resolutions Qualifications Required: - Excellent knowledge and experience in Azure Data Services (like ADF, etc.) along with Databricks - Good knowledge and expertise in SQL, Python, PySpark - Good understanding of ETL & Data warehousing concepts - Excellent written & spoken communication skills for interactions with customers and stakeholders - Ability to understand requirements and architecture and translate them into implementation following best practices - Self-driven with an excellent learning attitude - Experience in Snowflake Datawarehouse Good to have: - Experience & conceptual knowledge of Bigdata technologies such as Hadoop, HDFS, Hive, Spark - Knowledge and exposure to Databricks - Experience with ETL tools like Matillion, DBT, etc. - Exposure to other cloud platforms,
ACTIVELY HIRING
question

Are these jobs relevant for you?

posted 3 weeks ago

Tech Data Lead

Boyen Haddin
experience5 to 9 Yrs
location
Gurugram, All India
skills
  • ETL tools
  • Snowflake
  • SQL
  • Python
  • Java
  • Data governance
  • Communication skills
  • Project management
  • ETLELT processes
  • Redshift
  • BigQuery
  • Compliance standards
Job Description
You have an exciting opportunity as a Data Engineering Leader with 8+ years of experience in data engineering, including at least 5 years in a leadership role. Your success in this position will rely on your proficiency in ETL/ELT processes, hands-on experience with ETL tools such as Talend or Informatica, and expertise in cloud-based data platforms like Snowflake, Redshift, or BigQuery. Additionally, your strong SQL skills, knowledge of database tuning, data modeling, and schema design will be key to excelling in this role. Your familiarity with programming languages like Python or Java for data processing, as well as your understanding of data governance and compliance standards, will be essential. Moreover, your excellent communication and project management skills will play a vital role in your success, demonstrating your ability to prioritize and manage multiple projects simultaneously. The location for this role is Gurgaon, with the flexibility of working 3-4 days from the office. **Key Responsibilities:** - Lead data engineering projects and teams, leveraging your expertise in ETL/ELT processes - Utilize ETL tools such as Talend or Informatica for effective data processing - Work with cloud-based data platforms like Snowflake, Redshift, or BigQuery to drive data initiatives - Develop and optimize SQL queries, perform database tuning, and design efficient schemas - Use programming languages like Python or Java for data manipulation and analysis - Ensure adherence to data governance and compliance standards within all projects - Manage multiple projects simultaneously, demonstrating strong project management skills **Qualifications:** - Bachelor's or Master's Degree in IT or equivalent - Excellent verbal and written communication skills Please note that the additional details of the company mentioned in the JD have been omitted as they were not provided in the text. You have an exciting opportunity as a Data Engineering Leader with 8+ years of experience in data engineering, including at least 5 years in a leadership role. Your success in this position will rely on your proficiency in ETL/ELT processes, hands-on experience with ETL tools such as Talend or Informatica, and expertise in cloud-based data platforms like Snowflake, Redshift, or BigQuery. Additionally, your strong SQL skills, knowledge of database tuning, data modeling, and schema design will be key to excelling in this role. Your familiarity with programming languages like Python or Java for data processing, as well as your understanding of data governance and compliance standards, will be essential. Moreover, your excellent communication and project management skills will play a vital role in your success, demonstrating your ability to prioritize and manage multiple projects simultaneously. The location for this role is Gurgaon, with the flexibility of working 3-4 days from the office. **Key Responsibilities:** - Lead data engineering projects and teams, leveraging your expertise in ETL/ELT processes - Utilize ETL tools such as Talend or Informatica for effective data processing - Work with cloud-based data platforms like Snowflake, Redshift, or BigQuery to drive data initiatives - Develop and optimize SQL queries, perform database tuning, and design efficient schemas - Use programming languages like Python or Java for data manipulation and analysis - Ensure adherence to data governance and compliance standards within all projects - Manage multiple projects simultaneously, demonstrating strong project management skills **Qualifications:** - Bachelor's or Master's Degree in IT or equivalent - Excellent verbal and written communication skills Please note that the additional details of the company mentioned in the JD have been omitted as they were not provided in the text.
ACTIVELY HIRING
posted 1 week ago
experience7 to 11 Yrs
location
Noida, Uttar Pradesh
skills
  • Snowflake
  • ETL
  • Azure
  • Python
  • ETLELT pipelines
  • Data Modelling
  • Azure Data Factory ADF
  • Apache Airflow
Job Description
Job Description: As a Tech Lead specializing in Snowflake (Data Engineering) at Nuaav, you will play a crucial role in leading the design, development, and optimization of data pipelines and Snowflake-based solutions. Your responsibilities will include architecting scalable Snowflake data models, guiding and mentoring a team of data engineers, collaborating with clients to design efficient data solutions, and ensuring high standards of code quality and performance tuning. Additionally, you will integrate Snowflake pipelines with Spark, Airflow, Kafka, and Azure cloud services to deliver robust and high-performance data architectures for enterprise environments. Key Responsibilities: - Lead the design, development, and optimization of data pipelines and Snowflake-based solutions. - Architect scalable Snowflake data models and Snowpark applications. - Guide and mentor a team of data engineers, ensuring high-quality delivery. - Collaborate with clients to understand business needs and design efficient and maintainable data solutions. - Work with large-scale semi-structured datasets and complex schema environments. - Drive best practices in ETL/ELT design, data modeling, and performance tuning. - Integrate Snowflake pipelines with Spark, Airflow, Kafka, and Azure cloud services. - Ensure high standards of code quality, CI/CD, DevOps practices, and automation. Key Requirements: - 7+ years of experience in Data Engineering, including 3+ years of recent hands-on Snowflake work. - 1+ year of experience in a Tech Lead / Team Lead capacity. - Deep expertise in Snowflake job optimization, Snowpark applications, and scalable data modeling. - Proven experience designing complex ETL/ELT pipelines and transformations in Snowflake environments. - Experience processing and normalizing high-volume semi-structured data using Snowpark. - Strong understanding of data modeling and schema design; experience working with enterprise or telecom domain datasets is preferred. - Working knowledge of Spark for large-scale data transformations. - Proficiency in Python, Apache Airflow, Apache Kafka, Spark, and related pipeline tools. - Strong experience with Azure services: ADF, ADLS, Azure SQL, and CI/CD/DevOps workflows. - Excellent communication skills with the ability to mentor teams and engage with clients effectively. About Uplers: At Uplers, our aim is to streamline the hiring process, making it reliable, simple, and fast. We are committed to assisting talents in finding and applying for relevant contractual onsite opportunities to advance their careers. We are here to provide support for any challenges or grievances you may encounter during the engagement. If you are seeking a new challenge, a supportive work environment, and an opportunity to elevate your career, seize the moment and apply now. We are excited to welcome you aboard!,
ACTIVELY HIRING
posted 3 weeks ago

Big Data Developer

Obrimo Technologies
experience6 to 10 Yrs
location
Gurugram, All India
skills
  • Power BI
  • SQL
  • DAX
  • Snowflake
  • Python
  • ETL
  • Data Modeling
  • Power Query
  • Azure Data Factory
  • Databricks
Job Description
As a Lead Data Analyst in the Data & Analytics team, you will play a crucial role in analyzing large and complex datasets to extract actionable insights. Your responsibilities will include designing and developing end-to-end BI solutions using Power BI, SQL, DAX, and Power Query. Additionally, you will be building scalable ETL pipelines and data workflows utilizing Azure Data Factory, Databricks, and Delta Lake. Your expertise will be essential in leading data modeling efforts to optimize reporting performance with Star and Snowflake schemas. Your role will also involve migrating and managing cloud data platforms, especially Snowflake and Azure-based environments. Automation of repetitive data processing tasks using Python and ensuring high data quality will be part of your routine tasks. Key Responsibilities: - Analyze large and complex datasets to extract actionable insights and support strategic decision-making. - Design and develop end-to-end BI solutions using Power BI, SQL, DAX, and Power Query. - Build scalable ETL pipelines and data workflows using Azure Data Factory, Databricks, and Delta Lake. - Lead data modeling efforts using Star and Snowflake schemas to optimize reporting performance. - Migrate and manage cloud data platforms, especially Snowflake and Azure-based environments. - Automate repetitive data processing tasks using Python and streamline reporting operations. - Collaborate with cross-functional teams to gather business requirements, resolve data issues, and ensure high data quality. - Mentor junior team members and ensure documentation and adherence to best practices in data engineering. - Work with stakeholders across US and Indian teams, maintaining strong communication and delivery alignment. Qualifications Required: - 6-8 years of experience in delivering robust business intelligence solutions, advanced data modeling, and cloud-based data engineering. - Proficiency in BI & Visualization tools such as Power BI, DAX, Power Query, and Streamlit. - Strong knowledge of Data Platforms including SQL, Snowflake, Azure Synapse, and Delta Lake. - Experience with ETL & Pipelines using Azure Data Factory, PySpark, and Databricks. - Skilled in Programming & Automation with Python, Pandas, VBA, and Git. - Familiarity with other tools like SSAS, Confluence, HTML/CSS, and basics of Machine Learning. Please note: The JD does not provide any additional details of the company. As a Lead Data Analyst in the Data & Analytics team, you will play a crucial role in analyzing large and complex datasets to extract actionable insights. Your responsibilities will include designing and developing end-to-end BI solutions using Power BI, SQL, DAX, and Power Query. Additionally, you will be building scalable ETL pipelines and data workflows utilizing Azure Data Factory, Databricks, and Delta Lake. Your expertise will be essential in leading data modeling efforts to optimize reporting performance with Star and Snowflake schemas. Your role will also involve migrating and managing cloud data platforms, especially Snowflake and Azure-based environments. Automation of repetitive data processing tasks using Python and ensuring high data quality will be part of your routine tasks. Key Responsibilities: - Analyze large and complex datasets to extract actionable insights and support strategic decision-making. - Design and develop end-to-end BI solutions using Power BI, SQL, DAX, and Power Query. - Build scalable ETL pipelines and data workflows using Azure Data Factory, Databricks, and Delta Lake. - Lead data modeling efforts using Star and Snowflake schemas to optimize reporting performance. - Migrate and manage cloud data platforms, especially Snowflake and Azure-based environments. - Automate repetitive data processing tasks using Python and streamline reporting operations. - Collaborate with cross-functional teams to gather business requirements, resolve data issues, and ensure high data quality. - Mentor junior team members and ensure documentation and adherence to best practices in data engineering. - Work with stakeholders across US and Indian teams, maintaining strong communication and delivery alignment. Qualifications Required: - 6-8 years of experience in delivering robust business intelligence solutions, advanced data modeling, and cloud-based data engineering. - Proficiency in BI & Visualization tools such as Power BI, DAX, Power Query, and Streamlit. - Strong knowledge of Data Platforms including SQL, Snowflake, Azure Synapse, and Delta Lake. - Experience with ETL & Pipelines using Azure Data Factory, PySpark, and Databricks. - Skilled in Programming & Automation with Python, Pandas, VBA, and Git. - Familiarity with other tools like SSAS, Confluence, HTML/CSS, and basics of Machine Learning. Please note: The JD does not provide any additional details of the company.
ACTIVELY HIRING
posted 2 weeks ago

Tech Lead - Data Engineering

Ameriprise Financial Services, LLC
experience4 to 8 Yrs
location
Noida, All India
skills
  • SQL
  • Snowflake
  • Python
  • data governance
  • security
  • performance optimization
  • ETL tool
  • Informatica
  • SSIS
  • technology infrastructure
  • data profiling
  • SDLC
  • AWS infrastructure
  • ITSM process
  • CICD deployments
  • AWS Athena
  • Parquet
  • AthenaSnowFlake
  • Lambda
Job Description
As a highly skilled Data Delivery Developer with expertise in SQL, Snowflake, and AWS infrastructure, your role will involve working on complex data migration projects, particularly migrating from SSIS/SQL Server to Snowflake. Your responsibilities will include: - Designing, developing, and maintaining scalable data pipelines using Snowflake and AWS services. - Writing complex and optimized SQL queries for data extraction, transformation, and reporting. - Automating data workflows and delivery processes using Python. - Collaborating with data architects, analysts, and business stakeholders to understand data requirements. - Monitoring and troubleshooting data delivery issues to ensure data quality and integrity. - Implementing best practices for data governance, security, and performance optimization. - Driving the optimization of technology operations through automation and solution delivery, system/services performance monitoring. - Developing, enhancing, and troubleshooting database objects such as stored procedures, functions, tables, and views. - Creating and maintaining data dictionaries, data model, and change management documentation. - Conducting data profiling to understand data relationships and the quality of data through AWS Athena. - Participating in all phases of the software development life cycle (SDLC), including requirements, technical design, development, testing, and implementation. - Working with column-oriented data storage formats such as Parquet, interactive query services such as Athena/Snowflake, and event-driven computing cloud service Lambda. Preferred qualifications for this role include: - 4+ years of strong experience in data delivery projects with exposure to writing complex SQL queries, performing SQL optimization, and working on complex data migration assignments. - Experience in Snowflake, AWS data suite of services like S3, Athena, Glue, etc. - Experience on AWS infra using CI/CD deployments. - Experience with any one of ETL tools like Informatica, SSIS, etc. - Thorough understanding of ITSM process and aptitude to understand the technology infrastructure and cost. - Demonstrated ability to manage multiple and varied assignments. - Ability to quickly perform critical analysis and use creative approaches for solving complex problems. - Strong academic background. - Excellent written and verbal communication skills. About Our Company: Ameriprise India LLP has been providing client-based financial solutions for 125 years, helping clients plan and achieve their financial objectives. As a U.S.-based financial planning company headquartered in Minneapolis with a global presence, the firm's focus areas include Asset Management and Advice, Retirement Planning, and Insurance Protection. Join our inclusive and collaborative culture that rewards your contributions and work alongside other talented individuals who share your passion for doing great work. You'll have opportunities to make your mark at the office and a difference in your community. If you're talented, driven, and want to work for a strong ethical company that cares, take the next step and create a career at Ameriprise India LLP. This is a full-time position with timings from 2:00 pm to 10:30 pm in the India Business Unit of AWMPO AWMP&S President's Office under the Job Family Group of Technology. As a highly skilled Data Delivery Developer with expertise in SQL, Snowflake, and AWS infrastructure, your role will involve working on complex data migration projects, particularly migrating from SSIS/SQL Server to Snowflake. Your responsibilities will include: - Designing, developing, and maintaining scalable data pipelines using Snowflake and AWS services. - Writing complex and optimized SQL queries for data extraction, transformation, and reporting. - Automating data workflows and delivery processes using Python. - Collaborating with data architects, analysts, and business stakeholders to understand data requirements. - Monitoring and troubleshooting data delivery issues to ensure data quality and integrity. - Implementing best practices for data governance, security, and performance optimization. - Driving the optimization of technology operations through automation and solution delivery, system/services performance monitoring. - Developing, enhancing, and troubleshooting database objects such as stored procedures, functions, tables, and views. - Creating and maintaining data dictionaries, data model, and change management documentation. - Conducting data profiling to understand data relationships and the quality of data through AWS Athena. - Participating in all phases of the software development life cycle (SDLC), including requirements, technical design, development, testing, and implementation. - Working with column-oriented data storage formats such as Parquet, interactive query services such as Athena/Snowflake, and event-driven computing cloud service Lambda. Preferred qualifications for this role include:
ACTIVELY HIRING
posted 1 day ago

Data & AI Manager

InTalent Asia
experience12 to 16 Yrs
location
Delhi
skills
  • Data Management
  • Analytics
  • Stakeholder Management
  • Strategic Thinking
  • AWS
  • GCP
  • Snowflake
  • Power BI
  • Data Architecture
  • Forecasting
  • Risk Management
  • Predictive Analytics
  • Procurement
  • Finance
  • Supply Chain
  • NLP
  • AIML
  • Cloud Platforms Azure
  • Data Stacks Azure
  • Databricks
  • Data LakeWarehouse Design
  • ERP Integration
  • BI Platforms Integration
  • AIML Models Development
  • Operational Optimization
  • Digital Twins
  • AIpowered Bots
  • Emerging Technologies Generative AI
  • RPA
  • Experimentation
  • Agility
  • Continuous Learning
Job Description
As a Data & AI Manager in the fashion/textile and apparel manufacturing & export industry in India, you will play a critical role in leading enterprise-level data and AI transformation initiatives for a prominent apparel manufacturing company. Your passion for building a data-driven organization and implementing AI solutions at scale will be instrumental in driving innovation and delivering measurable business impact through data and AI. **Key Responsibilities:** - Define and implement enterprise data strategy, governance frameworks, and data quality standards. - Oversee data architecture, data lake/warehouse design, and integration with ERP and BI platforms (e.g., SAP S/4HANA, SAC, other data sources). - Lead the development and deployment of AI/ML models for forecasting, risk management, and operational optimization. - Drive initiatives such as digital twins, AI-powered bots, and predictive analytics for procurement, finance, and supply chain. - Manage cross-functional data and AI programs, ensuring alignment with strategic goals and timely delivery. - Collaborate with IT, business improvement, and factory teams across Hong Kong, Bangladesh, and India. - Build and mentor a high-performing team of data scientists, engineers, and analysts. - Act as a trusted advisor to senior leadership, translating business needs into data-driven solutions. - Evaluate emerging technologies (e.g., generative AI, NLP, RPA) and drive their adoption. - Promote a culture of experimentation, agility, and continuous learning. **Qualifications:** - Bachelors or Masters degree in Data Science, Computer Science, Engineering, or related field. - Minimum 12 years of experience in data management, analytics, and AI/ML. - At least 5 years in leadership roles. - Proven track record in delivering enterprise-scale data and AI solutions. - Strong knowledge of modern data stacks (e.g., Azure, Snowflake, Databricks, Power BI). - Excellent communication, stakeholder management, and strategic thinking skills. - Certifications in cloud platforms (Azure, AWS, GCP) and AI/ML frameworks are a plus. If you are ready to be part of a transformation journey where data and AI drive the future, this role offers you the opportunity to make a significant impact in the industry.,
ACTIVELY HIRING
posted 2 months ago
experience3 to 8 Yrs
location
Noida, Uttar Pradesh
skills
  • Snowflake
  • SQL
  • Data Warehousing
  • Git
  • Azure
  • AWS
  • Matillion
  • CICD
Job Description
As an experienced Matillion Developer with strong knowledge of Snowflake and exposure to the banking or financial services domain, your role will involve designing and maintaining scalable ETL pipelines to ensure data accuracy, performance, and compliance. Key Responsibilities: - Develop and optimize ETL workflows using Matillion and Snowflake - Work with banking data such as transactions, compliance, and risk - Collaborate with analysts and business teams to deliver effective data solutions - Ensure data integrity, security, and performance tuning - Provide support for cloud-based data integration, primarily on Azure and AWS Qualifications: - 6+ years of experience in Data Engineering/ETL development - 4+ years of experience with Matillion and Snowflake - Strong SQL and data warehousing experience - Understanding of banking/financial data structures and compliance - Familiarity with Git, CI/CD, and cloud platforms Any additional details of the company: You will be joining 3i Infotech, a leading IT company with a strong focus on providing innovative solutions in the banking and financial services sector. Your role will involve working with global customers and collaborating with global teams, showcasing a strong customer service orientation. For further details or application, please contact: chritina.ragasia@3i-infotech.com,
ACTIVELY HIRING
posted 2 weeks ago

Devops cloud architect

Technology Next
experience3 to 7 Yrs
location
Noida, Uttar Pradesh
skills
  • AWS
  • Azure
  • GCP
  • Apache Spark
  • Snowflake
  • Automation
  • DevOps
  • DataStage
  • Cloud Architecture
  • CICD
  • Infrastructure as Code
  • ETL Migration
Job Description
As a Cloud Architect with DevOps expertise, your role will be crucial in driving the cloud migration journey and ensuring cloud security at scale. You will be responsible for implementing automation and working with ETL DataStage. **Key Responsibilities:** - Drive cloud migration journey from ETL DataStage - Implement automation for efficient operations - Ensure cloud security at scale - Work with Cloud Architecture on AWS, Azure, and GCP - Utilize Apache Spark and Snowflake for data processing - Implement CI/CD and Infrastructure as Code practices - Follow DevOps best practices for efficient development - Utilize your experience in ETL Migration, with a preference for DataStage **Qualifications Required:** - Minimum 7 years of experience in Cloud Architecture - Proficiency in Apache Spark and Snowflake - Strong experience in DevOps for at least 4 years - Familiarity with CI/CD, Automation, and Infrastructure as Code - Prior experience in ETL Migration, with expertise in DataStage This job is a full-time contractual position for 12 months, based in Noida, Uttar Pradesh. The work schedule is from Monday to Friday, with a day shift timing from 11:00 AM to 8:30 PM IST. If you are passionate about cloud architecture, DevOps, and have hands-on experience in cloud migration and security, this opportunity is for you.,
ACTIVELY HIRING
posted 2 weeks ago
experience3 to 7 Yrs
location
Delhi
skills
  • Snowflake
  • SQL Server
  • SSIS
  • TSQL
  • Python
  • Agile methodologies
  • data validation
  • unit testing
  • QA
  • data modeling
  • database design
  • AWS
  • Azure Data Factory
  • Selfhosted Integration Runtime
  • Logic Apps
  • Azure Data Lake Storage Gen2
  • Blob Storage
  • CICD pipelines
  • GitHub Actions
  • Databricks
  • Google Cloud
  • Azure certifications
  • Snowflake certifications
Job Description
As a Cloud Data Engineer with QA, you will be an integral part of a Cloud Data Modernization initiative aimed at migrating an on-premises Enterprise Data Warehouse (SQL Server) to a modern cloud-based data platform using Azure and Snowflake. Your role will require a blend of data engineering, cloud platform expertise, and data quality assurance. **Key Responsibilities:** - Design, develop, and implement scalable data platform solutions using Azure and Snowflake - Migrate and modernize on-premises ETL pipelines (SSIS, T-SQL) to cloud-native solutions - Build and optimize data pipelines using Azure Data Factory (ADF), Self-hosted Integration Runtime (SHIR), Logic Apps, and Azure Data Lake Storage Gen2 (ADLS Gen2) - Implement DevOps practices and CI/CD workflows using GitHub Actions - Collaborate with cross-functional teams to ensure seamless data integration and delivery - Conduct thorough data validation and QA testing to ensure consistency and accuracy - Troubleshoot and optimize data workflows for performance and reliability - Ensure compliance with data security and industry standards **Qualifications Required:** - 3+ years of experience as a Cloud Data Engineer - Strong hands-on experience with Azure Cloud data tools such as ADF, SHIR, Logic Apps, ADLS Gen2, and Blob Storage - Expertise in Snowflake for cloud data warehousing - Proficient in ETL development using SQL Server, SSIS, and T-SQL - Solid scripting skills in Python for data processing and validation - Experience implementing CI/CD pipelines using GitHub Actions - Familiarity with Agile methodologies - Proven experience in data validation, unit testing, and QA for data migration projects - Strong analytical mindset and attention to detail - Excellent communication and collaboration skills - Ability to work independently and adapt to new technologies quickly If there are any additional details about the company in the job description, please provide them separately for inclusion.,
ACTIVELY HIRING
posted 2 months ago
experience5 to 9 Yrs
location
Noida, Uttar Pradesh
skills
  • Data Integration
  • Optimization
  • Performance Tuning
  • Reverse Engineering
  • Requirement Gathering
  • Technical Documentation
  • Data Analysis
  • Postgres
  • Python
  • Development
  • Data Transformation
  • SSIS
  • Database Designing
  • Data Modelling
  • Core Component Implementation
  • Relational Data Modelling
  • Automating Backup
  • Purging Processes
  • Basic SQL
  • Migration of Database Objects
  • Application Data
  • Oracle DB
  • Cloud Architecture
  • Snowflakes
  • Modification of Major Database Components
Job Description
As a valued member of Iris Software, you will have the opportunity to be a part of one of India's Top 25 Best Workplaces in the IT industry. Join us to make a difference and do the best work of your life in a fast-growing IT services company with an award-winning work culture that values your talent and career aspirations. **Role Overview:** At Iris Software, you will be responsible for database designing, data modeling, and core component implementation. Your role will involve data integration, relational data modeling, optimization, and performance tuning. You will also automate backup and purging processes, reverse engineer existing applications, and handle basic SQL tasks. Additionally, you will be involved in requirement gathering, technical documentation, and data analysis for migration of database objects and application data, specifically for Postgres or Oracle DB. **Key Responsibilities:** - Database Designing, Data Modelling, and Core Component Implementation - Data Integration and Relational Data Modelling - Optimization and Performance Tuning - Automating Backup and Purging Processes - Reverse Engineering of Existing Applications - Basic SQL - Requirement Gathering and Technical Documentation - Data Analysis for Migration of Database Objects and Application Data - Data Analysis for Postgres or Oracle DB **Good to Have:** - Basic Knowledge of Cloud Architecture, Snowflakes, Python - Development/Modification of Major Database Components - Data Transformation Using SSIS **Qualifications Required:** - Mandatory Competencies in Database - Oracle - Data Modelling, Database - Database Programming - SQL, Beh - Communication, Database - Sql Server - DBA, Database - Other Databases - PostgreSQL Joining Iris Software comes with a range of perks and benefits designed to support your financial, health, and well-being needs. From comprehensive health insurance and competitive salaries to flexible work arrangements and ongoing learning opportunities, we are committed to providing a supportive and rewarding work environment. Come be a part of Iris Software and experience a workplace that values your success and happiness.,
ACTIVELY HIRING
posted 2 months ago
experience8 to 12 Yrs
location
Noida, Uttar Pradesh
skills
  • AWS
  • Azure
  • GCP
  • Snowflake
  • Spark
  • Jenkins
  • Git
  • Elasticsearch
  • Kibana
  • Apache Spark
  • Hadoop
  • Python
  • Java
  • Bash
  • SQL
  • HIPAA
  • Databricks
  • Data Lake
  • Data Warehouse
  • Terraform
  • CloudFormation
  • CICD
  • Prometheus
  • Grafana
  • SOC2
Job Description
Role Overview: You will be responsible for designing and implementing scalable and secure cloud architecture on cloud platforms like AWS, Azure, and GCP. Your expertise in building enterprise-grade data platforms using technologies such as Databricks, Snowflake, Spark, and Data Lake/Warehouse solutions will be crucial in supporting advanced analytics and AI/ML integrations. Additionally, you will conduct research and development for custom architecture PoCs, including AI/ML model deployment, security enhancements, and scalable data workflows to meet evolving business needs. Key Responsibilities: - Design and deploy data pipelines with technologies like AWS (S3, Lambda, Glue, Data Pipeline), Databricks, and Snowflake for real-time streaming, batch processing, and analytics workloads. - Provision cloud infrastructure using Terraform, CloudFormation, and Control Tower to ensure multi-AZ, high availability, and multi-tenant architectures. - Implement hybrid cloud architecture for cloud-to-cloud integrations, cross-cloud data migration, and multi-cloud orchestration involving AWS/GCP or Azure, and on-prem systems. - Design and implement CI/CD pipelines using CloudFormation, Jenkins, Git, Terraform to support automated deployment of applications and infrastructure. - Develop SaaS solutions with secure management provisioning, license management, and custom deployment architectures. - Utilize monitoring and logging tools like Prometheus, Grafana, Elasticsearch, Kibana to ensure performance optimization, fault detection, and quick bug resolution. - Design, develop, and optimize data pipelines and implement data warehousing solutions using platforms like AWS, Snowflake, Databricks. - Create scalable, distributed data processing systems using Apache Spark, Hadoop, or similar frameworks. - Develop and maintain logical and physical data models to support analytics, reporting, and ML use cases. - Collaborate with Data Scientists, Analysts, and Business Teams to support data-driven decision-making and provide guidance on data best practices, performance tuning, and data-driven solutions. - Mentor and train junior data engineers and contribute to team growth. Qualifications Required: - 8 years of experience as a Cloud Architect with proficiency in cloud platforms AWS, GCP, and Azure. - Hands-on experience with Data Platforms & Tools such as Databricks, Snowflake, Hadoop, Spark, Data Lake, and Data Warehouse. - Expertise in DevOps & Automation tools like Terraform, CloudFormation, Kubernetes, Docker, and CI/CD pipelines. - Proficiency in Monitoring & Logging tools including Prometheus, Grafana, ELK Stack. - Strong programming and scripting skills in Python, Java, Bash, and SQL. - Knowledge of Security & Compliance standards such as SOC2, HIPAA, audit logs, and security controls.,
ACTIVELY HIRING
posted 2 weeks ago

Aladdin BA- Accounting- Manager

Imbibe Consultancy Services Pvt Ltd
experience8 to 12 Yrs
location
Gurugram, All India
skills
  • Team Management
  • Client Reporting
  • Data Engineering
  • AWS
  • Azure
  • Snowflake
  • Data Engineering
  • Aladdin
  • Cash
  • Position Reconciliation
  • Accrual
  • PNL Validation
  • Asset Class Expertise
  • Fund
  • Client Accounting
  • Onboarding
  • Fund Setup
  • Regulatory Evaluation
  • Aladdin Accounting Experience
  • Aladdin Data Cloud
  • Aladdin Data BA
Job Description
As an Aladdin professional at AceNet Consulting, you will play a crucial role in overseeing fund and client accounting operations with a sharp focus on accuracy, compliance, and timely delivery. Your expertise in cash and position reconciliation, daily accrual and profit-and-loss (PNL) validation, and managing various asset classes will be key to leading the accounting team and ensuring seamless fund and client onboarding. Your responsibilities will include: - Guide the team in performing daily cash and position reconciliations to maintain accuracy and completeness of fund records. - Oversee validation of daily accruals and PNL to ensure accurate reporting of income, expenses, and valuations. - Apply in-depth knowledge of multiple asset classes such as equities, fixed income, derivatives, and alternatives in all accounting processes. - Manage end-to-end accounting tasks including NAV calculation, fee accruals, and general ledger activities for assigned funds and clients. - Lead, mentor, and develop a team of accountants to promote accuracy, accountability, and continuous improvement. - Ensure successful onboarding of new funds and clients by overseeing all systems, controls, and documentation. - Prepare and review periodic and ad hoc client reports to maintain consistency, clarity, and compliance with client requirements. - Evaluate and implement regulatory requirements to ensure adherence to current regulations and best practices. - Utilize Aladdin or similar investment accounting platforms for data management, reconciliation, and reporting functions. Qualifications required for this role include: - Bachelor's or Master's degree in Accounting, Finance, or a related field. - Minimum 8-10 years of relevant fund/client accounting experience. - Experience in Aladdin Data Cloud (ADC) and Snowflake, AWS/Azure, Data Engineering, and Aladdin Data BA. - Strong understanding of fund and client accounting principles and practices. - Proficiency with Aladdin or equivalent investment accounting systems. - Excellent analytical and problem-solving abilities. - Demonstrated experience in leading teams within the financial services sector. - Hands-on operational knowledge of Aladdin accounting platform is highly desirable. Joining AceNet Consulting will provide you with opportunities to work on transformative projects, cutting-edge technology, and innovative solutions with leading global firms. You will benefit from continuous investment in your growth and professional development, competitive compensation, benefits, ESOPs, international assignments, a supportive environment with a focus on work-life balance and employee well-being, and an open culture that values diverse perspectives and encourages transparent communication. If you are passionate about technology and ready to thrive in a fast-paced environment, we encourage you to apply and submit your resume to be considered for this exciting opportunity at AceNet Consulting. As an Aladdin professional at AceNet Consulting, you will play a crucial role in overseeing fund and client accounting operations with a sharp focus on accuracy, compliance, and timely delivery. Your expertise in cash and position reconciliation, daily accrual and profit-and-loss (PNL) validation, and managing various asset classes will be key to leading the accounting team and ensuring seamless fund and client onboarding. Your responsibilities will include: - Guide the team in performing daily cash and position reconciliations to maintain accuracy and completeness of fund records. - Oversee validation of daily accruals and PNL to ensure accurate reporting of income, expenses, and valuations. - Apply in-depth knowledge of multiple asset classes such as equities, fixed income, derivatives, and alternatives in all accounting processes. - Manage end-to-end accounting tasks including NAV calculation, fee accruals, and general ledger activities for assigned funds and clients. - Lead, mentor, and develop a team of accountants to promote accuracy, accountability, and continuous improvement. - Ensure successful onboarding of new funds and clients by overseeing all systems, controls, and documentation. - Prepare and review periodic and ad hoc client reports to maintain consistency, clarity, and compliance with client requirements. - Evaluate and implement regulatory requirements to ensure adherence to current regulations and best practices. - Utilize Aladdin or similar investment accounting platforms for data management, reconciliation, and reporting functions. Qualifications required for this role include: - Bachelor's or Master's degree in Accounting, Finance, or a related field. - Minimum 8-10 years of relevant fund/client accounting experience. - Experience in Aladdin Data Cloud (ADC) and Snowflake, AWS/Azure, Data Engineering, and Aladdin Data BA. - Strong understanding of fund and client accounting principles and practices. - Proficiency with Aladdin or equivalent investment accounting systems.
ACTIVELY HIRING
posted 2 months ago
experience4 to 8 Yrs
location
Noida, Uttar Pradesh
skills
  • Snowflake
  • Data Warehousing
  • ETL
  • RDBMS
  • SQL
  • PLSQL
  • Python
  • Scala
  • Kafka
  • Dimensional Modelling
Job Description
As a Developer / Senior Developer, your role will involve designing and implementing effective Analytics solutions and models with Snowflake. You should have a minimum of 5 years of IT experience, with a significant amount of experience in Snowflake. Your responsibilities will include: - Demonstrating in-depth understanding of Data Warehousing, ETL concepts, and modeling structure principles - Developing big data/ETL data warehouse experience by building cloud-native data pipelines - Utilizing Dimensional Modelling techniques effectively - Working with Snowflake Functions, utilities, and features such as time travel, failsafe, and SnowSQL - Implementing different types of Schema and having knowledge about Snowflake Architecture - Having a good grasp of RDBMS topics and the ability to write complex SQL and PL/SQL queries - Handling engineering platform components like Data Pipelines, Data Orchestration, Data Quality, Data Governance & Analytics - Implementing large-scale data intelligence solutions around Snowflake DW - Using scripting languages such as Python or Scala - Working with streaming services like Kafka - Dealing with Semi-Structured data effectively Qualifications Required: - 4-8 years of overall experience in IT - At least 5 years of experience in Snowflake - Experience in designing and implementing Analytics solutions - Knowledge of Data Warehousing, ETL concepts, and modeling principles - Proficiency in SQL and PL/SQL - Familiarity with Python or Scala scripting - Experience with streaming services like Kafka,
ACTIVELY HIRING
posted 2 weeks ago

Data Engineer

iO Associates
experience3 to 7 Yrs
location
Delhi
skills
  • Snowflake
  • Python
  • Data warehousing
  • ETL development
  • Data modelling
Job Description
Role Overview: You will be responsible for developing and maintaining scalable data pipelines using Snowflake and Python. Your role will involve writing efficient Python scripts for data extraction, transformation, and loading. You will also ensure data quality, consistency, and reliability across all systems. Optimizing ETL processes, resolving performance issues, and collaborating with cross-functional teams will be key aspects of your responsibilities. Additionally, maintaining comprehensive documentation for data processes and pipelines will be crucial. Key Responsibilities: - Develop and maintain scalable data pipelines using Snowflake and Python. - Write efficient Python scripts for data extraction, transformation, and loading. - Ensure data quality, consistency, and reliability across all systems. - Optimize ETL processes and resolve performance issues. - Collaborate with cross-functional teams including data analysts and business stakeholders. - Maintain comprehensive documentation for data processes and pipelines. Qualifications Required: - Strong professional experience with Snowflake and data warehousing. - Advanced Python scripting skills for ETL development. - Solid understanding of data modeling and ETL best practices. - Experience working with structured and unstructured datasets. - Excellent problem-solving and debugging skills. (Note: No additional company details were mentioned in the provided job description),
ACTIVELY HIRING
posted 7 days ago

Senior Reltio Technical Architect

People Prime Worldwide
experience5 to 15 Yrs
location
Noida, Uttar Pradesh
skills
  • SAP
  • Salesforce
  • Snowflake
  • JSON
  • XML
  • Data Governance
  • Data Stewardship
  • Reltio MDM
  • REST APIs
  • AIML
  • Healthcare Domain
  • Insurance Domain
Job Description
As a Senior Reltio Technical Architect, you will be responsible for architecting and implementing scalable Reltio MDM solutions across enterprise environments. Your key responsibilities will include: - Leading integration architecture with SAP, Salesforce, Snowflake, and other systems - Designing and configuring entity models, survivorship rules, matchmerge logic, and workflow orchestration - Collaborating with data science teams to embed AI/ML-driven intelligent matching & anomaly detection - Defining and enforcing data governance and data stewardship frameworks - Mentoring MDM developers/analysts and supporting CoE enablement - Translating complex business data requirements into robust technical MDM solutions - Driving strategic MDM initiatives, especially within Healthcare and Insurance domains - Managing project delivery, planning, and stakeholder engagement Qualifications required for this role include: - Strong hands-on experience with Reltio MDM (modeling, configuration, API integration) - Expertise in core MDM concepts such as matchmerge, survivorship, and hierarchy management - Integration experience with SAP, Salesforce, Snowflake - Strong knowledge of REST APIs, JSON, XML, and data transformation tools - Exposure to AI/ML use cases in data management - Strong analytical, communication, and technical leadership skills - Domain knowledge in Healthcare and Insurance for the Senior Architect role Additionally, having experience with other MDM platforms like Informatica MDM, Stibo Systems MDM, cloud platforms (AWS/Azure/GCP), familiarity with data governance tools such as Collibra, Informatica, agile delivery experience, and Reltio MDM certification would be an added advantage. Please note that you should not apply if you do not meet the following criteria: - Lack strong hands-on experience with Reltio MDM - Lack integration experience with SAP/Salesforce/Snowflake - Have no exposure to core MDM concepts (matchmerge, survivorship, hierarchies) - Are unable to work from Hyderabad/Noida/Gurugram - Have a notice period of more than 30 days - Have no real-time experience in AI/ML-driven data quality (preferred for senior roles) It is essential to apply only if you meet all mandatory skills and conditions mentioned above to be considered for this role. Random or irrelevant applications will not be taken into consideration.,
ACTIVELY HIRING
posted 2 weeks ago

Data Analyst

42SIGNS INFOSYSTEMS LLP
experience6 to 11 Yrs
Salary5 - 10 LPA
WorkContractual
location
Gurugram, Delhi+7

Delhi, Noida, Bangalore, Chennai, Hyderabad, Kolkata, Pune, Mumbai City

skills
  • sql
  • tableau
  • snowflake
Job Description
Company name: Tramway Inc(http://www.tranwayinc.com/), Working for Client-Infosys( work location) Role: Hybrid (3 days in a week working from office) Office timings: Normal timings (UK hours)  Role: Contractual, Duration-6 months   Job Description: Should have strong experience in SQL, Snowflake and Tableau. Excellent problem-solving and communication skills. Ability to work independently and in a team-oriented environment. Mandatory skills: Primary Skills - SQL, Snowflake and Tableau. Secondary Skills - SAS Know how If you are interested, Please share your updated resume to prerna@42signs.com/8197 840 980.
posted 1 month ago
experience5 to 9 Yrs
location
Delhi
skills
  • SQL
  • data modeling
  • Snowflake
  • Spark
  • dbt
  • Talend
  • Informatica
  • Airflow
  • cloud services
  • Python
  • Scala
  • ETLELT pipeline development
  • cloudbased data platforms
  • Databricks
  • ETL workflows
  • data orchestration frameworks
  • Dagster
  • Luigi
  • data warehousing principles
  • data security best practices
  • analytical thinking
  • problemsolving
  • realtime data pipelines
  • data mesh
  • data fabric architecture
Job Description
Role Overview: As a candidate for this position, you will be expected to have in-depth knowledge and experience in SQL, data modeling, ETL/ELT pipeline development, and working with cloud-based data platforms such as Databricks or Snowflake. Your primary responsibility will involve designing scalable data models, managing data workflows efficiently, and ensuring the reliability and performance of vital financial datasets. This role is highly collaborative and will require you to work closely with various teams including engineering, analytics, product, and compliance. Key Responsibilities: - Design, implement, and maintain logical and physical data models to support transactional, analytical, and reporting systems. - Develop and manage scalable ETL/ELT pipelines for processing large volumes of financial transaction data. - Tune and optimize SQL queries, stored procedures, and data transformations to achieve maximum performance. - Build and manage data orchestration workflows using tools like Airflow, Dagster, or Luigi. - Architect data lakes and warehouses using platforms such as Databricks, Snowflake, BigQuery, or Redshift. - Enforce data governance, security, and compliance standards like PCI-DSS and GDPR. - Collaborate closely with data engineers, analysts, and business stakeholders to understand data requirements and provide solutions. - Conduct data profiling, validation, and quality assurance to maintain clean and consistent data. - Maintain comprehensive documentation for data models, pipelines, and architecture. Qualifications Required: - Advanced SQL expertise, including query tuning, indexing, and performance optimization. - Proficiency in developing ETL/ELT workflows using tools like Spark, dbt, Talend, or Informatica. - Experience with data orchestration frameworks such as Airflow, Dagster, Luigi, etc. - Strong hands-on experience with cloud-based data platforms like Databricks, Snowflake, or equivalent technologies. - Deep understanding of data warehousing principles including star/snowflake schema and slowly changing dimensions. - Working knowledge of cloud services like AWS, GCP, or Azure and data security best practices. - Strong analytical thinking and problem-solving skills in high-scale environments. Company Details: The company is focused on leveraging advanced data technologies to drive business success. They place a strong emphasis on collaboration and maintaining high standards of data integrity and governance. Working at this company will provide you with the opportunity to work with cutting-edge tools and technologies in the data space.,
ACTIVELY HIRING
posted 2 months ago

Senior Data Engineer

Sonata Software
experience5 to 9 Yrs
location
Noida, Uttar Pradesh
skills
  • Snowflake
  • Airflow
  • SQL
  • Python
  • AWS
  • Azure
  • GCP
  • Git
  • CICD
Job Description
Role Overview: As a Senior NodeJS Data Engineer at Sonata Software, you will be responsible for utilizing your 5+ years of hands-on experience in data engineering to build and maintain transformation pipelines using Python, Airflow, and Snowflake Stored procedures. You will play a crucial role in ensuring data quality, observability, and pipeline reliability through monitoring and alerting. Additionally, you will work with cloud data warehouses like Snowflake, orchestration tools such as Airflow, and have a strong command over SQL and Python for optimized data manipulation and ETL processes. Your expertise in AWS, Azure, or GCP services will be key in contributing to the data architecture, security, and performance tuning of the projects. Key Responsibilities: - Build and maintain transformation pipelines using Python, Airflow, and Snowflake Stored procedures. - Write optimized SQL and Python scripts for data manipulation and ETL processes. - Ensure data quality, observability, and pipeline reliability through monitoring and alerting. - Work with cloud data warehouses like Snowflake, orchestration tools such as Airflow, and have a strong command over SQL and Python. - Contribute to data architecture, security, and performance tuning using AWS, Azure, or GCP services. - Familiarity with version control (e.g., Git), CI/CD tools, and agile workflows. Qualifications Required: - 5+ years of hands-on experience in data engineering. - Strong experience with Snowflake or similar cloud data warehouses, Airflow or other orchestration tools, SQL, and Python. - Good understanding of data architecture, security, and performance tuning. - Hands-on experience with AWS, Azure, or GCP services. - Familiarity with version control (e.g., Git), CI/CD tools, and agile workflows. (Note: The additional details of the company have been omitted as they were not provided in the job description),
ACTIVELY HIRING
logo

@ 2025 Shine.com | All Right Reserved

Connect with us:
  • LinkedIn
  • Instagram
  • Facebook
  • YouTube
  • Twitter