dwh-jobs-in-mumbai, Mumbai

2 Dwh Jobs nearby Mumbai

Toggle to save search
posted 5 days ago
experience4 to 22 Yrs
location
Navi Mumbai, Maharashtra
skills
  • Data mapping
  • Project planning
  • Team leadership
  • Stakeholder management
  • Analytical skills
  • Communication skills
  • Data Warehouse Analyst
  • Corporate Banking Operations
Job Description
Role Overview: You are a seasoned Data Warehouse Analyst (Team Lead) with strong experience in Corporate Banking Operations domain, responsible for leading a critical Data Warehouse renewal/transformation project. Your key responsibilities include acting as Team Lead/Project Lead, leading and delivering data mapping activities, managing overall planning and execution, and coordinating with business stakeholders, IT teams, and third-party vendors. Your role also involves supervising, mentoring, and coaching team members, defining requirements, performing gap analysis, and ensuring adherence to banking governance standards. Key Responsibilities: - Act as Team Lead/Project Lead for the Data Warehouse renewal/migration project - Lead and deliver as-is and to-be data mapping activities - Manage overall planning, execution, timelines, and quality of deliverables - Coordinate with business stakeholders, IT teams, and third-party vendors - Supervise, mentor, and coach team members - Define requirements, perform gap analysis, and support data migration activities - Ensure high-quality documentation and adherence to banking governance standards Qualifications Required: - Total Experience: 10+ years - Relevant DWH Experience in Corporate Banking/Banking Operations domain: 5+ years - Hands-on experience of 3+ years in: - As-is/To-be data mapping - Data Warehouse/Data Mart projects in banking - DWH system renewal or large-scale data migration/transformation - Proven track record in project planning, execution, and team leadership - Strong stakeholder management with business, IT, and vendors - Excellent analytical skills, attention to detail, and ownership - Strong leadership skills with ability to motivate and develop team members - Excellent communication (verbal & written) and collaboration skills Additional Details: This is a high-visibility leadership role with excellent growth and long-term career prospects in a leading banking environment. Interested candidates are encouraged to share their updated CV with the specified subject line for consideration.,
ACTIVELY HIRING

Top Companies are Hiring in Your City

For Multiple Roles

Jio Platforms Ltd
Jio Platforms Ltdslide-preview-Genpact
posted 2 weeks ago
experience3 to 8 Yrs
location
Navi Mumbai, Maharashtra
skills
  • Data mapping
  • Documentation
  • DWH
  • Data mart
  • Analytical skills
  • Communication skills
  • Leadership skills
  • Data alignment
  • BI tool
  • Collaboration skills
Job Description
As a Data Warehouse Analyst at Mizuho Global Services Pvt Ltd (MGS), you will play a key role in executing mapping, supporting documentation, and practical project activities under the guidance of the AVP. Your participation in a data warehouse renewal project will be crucial for the success of the team. **Key Responsibilities:** - **Stakeholder Coordination** - Coordinate requirements with business users and IT stakeholders - Support requirement clarification, impact analysis, and documentation - **Business Requirement and Data Mapping Coordination** - Gather information from business users and IT specialists - Organize and document details, prepare drafts, and support gap analysis - Ensure accuracy and completeness of mapping documentation - **Documentation and Quality Control** - Prepare, review, and update data mapping documentation - Maintain consistency in project deliverables and compile project documentation - **User Support and Knowledge Transfer** - Promote standardization of deliverables and knowledge sharing within the team - Assist in end-user adoption activities **Qualifications Required:** - Experience in banking roles involving data alignment, mapping, or documentation for information systems (DWH/data mart) - Hands-on experience with DWH/data mart/EUC or BI tools from either the business or IT perspective - Experience collaborating across business and IT teams with strong documentation skills **Preferred Skills and Personal Attributes:** - Experience in DWH system renewal, requirements definition, or data migration/projects - Strong analytical skills, attention to detail, and ownership of tasks - Excellent communication and collaboration skills with a proactive learning attitude - Ability to lead, motivate, and develop team members with a strong sense of responsibility In this role, you can expect immense exposure and learning opportunities, excellent career growth, and the chance to work with highly passionate leaders and mentors. You will have the ability to build things from scratch and contribute to the data warehouse renewal project. To learn more about Mizuho Global Services Pvt Ltd, visit [here](https://www.mizuhogroup.com/asia-pacific/mizuho-global-services). Address: Mizuho Global Services India Pvt. Ltd. 11th Floor, Q2 Building Aurum Q Park, Gen 4/1, Ttc, Thane Belapur Road, MIDC Industrial Area, Ghansoli, Navi Mumbai- 400710. Gain valuable experience and make a significant impact in the world of data warehousing by joining Mizuho Global Services Pvt Ltd as a Data Warehouse Analyst.,
ACTIVELY HIRING
posted 2 months ago
experience3 to 7 Yrs
location
Pune, Maharashtra
skills
  • Java
  • Spring framework
  • RESTful APIs
Job Description
As a software developer at Amdocs, you will play a crucial role in designing, developing, modifying, debugging, and maintaining software code according to functional, non-functional, and technical design specifications. You will adhere to Amdocs software engineering standards, relevant software development methodology, and release processes to ensure code is maintainable, scalable, and supportable. Additionally, you will demonstrate the software products to stakeholders. When issues arise, you will investigate them by reviewing and debugging the code, providing fixes and workarounds, and ensuring changes are operable to maintain existing software solutions. Collaboration is key, as you will work within a team, participate in peer code reviews, offer comments and suggestions, and cooperate with cross-functional teams to achieve common goals. You will also take technical accountability for your specific work products within an application and offer technical support during the solution design for new requirements. Amdocs encourages you to proactively seek innovation, continuous improvement, and efficiency in all assigned tasks. Key Responsibilities: - Design, develop, modify, debug, and maintain software code according to specifications - Adhere to software engineering standards, development methodology, and release processes - Demonstrate software products to stakeholders - Investigate and resolve issues by reviewing and debugging code - Collaborate within a team, participate in code reviews, and work with cross-functional teams - Take technical accountability for specific work products and offer support during solution design Qualifications required for this position include: - Bachelor's degree in Science/IT/Computer Science or equivalent - 3+ years of Java experience (server-side) on Linux/Unix/Windows - Demonstrable experience with Spring framework components, such as Spring Boot, MVC, Integration, Security, etc. - Strong understanding of RESTful APIs and open systems In this role, you will have the opportunity to design and develop new software applications, providing you with challenges that foster personal and professional growth. Amdocs values diversity and inclusivity, welcoming applicants from all backgrounds. Join us in building the future and making it amazing at Amdocs.,
ACTIVELY HIRING
question

Are these jobs relevant for you?

posted 5 days ago

Senior IT Consultant

CIRCOR International, Inc.
experience5 to 10 Yrs
location
Maharashtra
skills
  • SAP FICO
  • SAP BW
  • Power BI
  • SOX compliance
  • finance
  • controlling
  • reporting
  • analytical skills
  • communication skills
  • BI solutions
  • IT audit processes
  • conceptual skills
Job Description
As a Senior IT Consultant specializing in Record to Report (R2R) & Business Intelligence (BI), your role will involve owning the end-to-end IT processes for R2R and BI&A. This includes designing processes, integrating systems, and continuously improving them. Your responsibilities will also include consulting with stakeholders, analyzing requirements, and implementing them in SAP ECC and DWH/BI systems. Additionally, you will be responsible for implementing and configuring SAP FI/CO and BI solutions such as SAP BW and Power BI. Key Responsibilities: - Own end-to-end IT processes for Record to Report (R2R) and Business Intelligence & Analytics (BI&A) - Consult and collaborate with stakeholders for process design and future requirements - Implement and configure SAP FI/CO and BI solutions - Coordinate projects for digitalization and automation of financial processes - Ensure data quality and governance in systems - Manage IT audits and compliance, especially in SOX and IT controls - Train business departments for optimal system usage - Collaborate with internal and external stakeholders Qualifications Required: - Degree in (Business) Informatics, Business Administration, or equivalent - 5-10 years of experience supporting and developing SAP FI/CO and BI systems In addition to strong knowledge in finance, controlling, and reporting, experience with SOX compliance and IT audit processes will be beneficial. Your role will require excellent analytical and conceptual skills, strong communication skills, and a high service orientation. Fluency in both German and English, both written and spoken, is essential. CIRCOR is an EEO Employer of Females / Minorities / Veterans / Individuals with Disabilities.,
ACTIVELY HIRING
posted 2 months ago
experience10 to 14 Yrs
location
Pune, Maharashtra
skills
  • BI reporting
  • DWH
  • Power BI
  • Tableau
  • Oracle
  • SQL Server
  • Agile
  • SDLC
  • ETL solutions
  • analytics solutions
  • SQL programming
  • cloud technology stack
Job Description
As a Manager Software Engineer in the Business Insights & DWH engineering team at Mastercard, you will lead scrum teams of software developers and testers to deliver quality software solutions. Your role will involve providing technical leadership in BI reporting, ETL solutions, DWH, and other analytics solutions. You will work closely with product and architecture teams to drive innovation and ensure efficient project delivery. Additionally, you will be responsible for automating software delivery processes, introducing new technologies, and fostering effective partnerships across organizations. - Manage scrum teams to develop quality software solutions efficiently. - Lead major cross-department initiatives with broad scope and long-term implications. - Provide technical leadership in BI reporting, ETL solutions, DWH, and analytics solutions. - Collaborate with product and architecture teams on product definition and technical design. - Ensure project staffing, training, and management align with approved guidelines. - Drive prioritization decisions and trade-offs with product partners. - Automate and simplify software delivery processes. - Introduce new technologies following enterprise guidelines. - Build effective partnerships across organizations. - Lead by example with hands-on engineering excellence. - Provide technology leadership in BI, DWH, and cloud data platforms. - Stay updated on analytics technology trends and industry best practices. - Participate in architectural discussions and agile ceremonies. - Support System and User Acceptance Testing. - Adhere to Agile processes and actively participate in SDLC practices. - Minimum 10+ years of experience in Requirement management, Business Insights, Data Warehousing, and OnPrem/Cloud data platforms. - Strong understanding of software systems, cloud data platforms, BI reporting, and ETL solutions. - Proficiency in DWH and SQL programming. - Hands-on experience with analytics tools like Power BI, Tableau, etc. - Exposure to cloud technology stack such as Databricks, Snowflake, AWS, or Azure is a plus. - Familiarity with Oracle, SQL Server, or other relational databases. - Experience in designing and implementing solutions focusing on non-functional concerns. - Ability to collaborate with cross-functional teams and stakeholders effectively. - Strong communication skills at all levels of the organization. - Experience in Agile and SDLC practices. - Ability to address challenges and opportunities with urgency, fairness, and decency.,
ACTIVELY HIRING
posted 2 months ago
experience3 to 7 Yrs
location
Pune, Maharashtra
skills
  • relational databases
  • Azure
  • SQL
  • Python
  • JIRA
  • data governance
  • ETL Developer
  • Data Warehouse modeling
  • Microsoft Data stack
  • Synapse Analytics
  • Data Factory
  • PowerBI
  • CICD tools
  • GIT repositories
  • Agile projects
  • problemsolving
Job Description
As an ETL Developer at the EU subsidiary of a Global Financial Bank, your role is crucial in designing, building interfaces, and integrating data from various sources into the Enterprise Data Warehouse environment. You will primarily focus on developing ETL solutions using Microsoft and Azure technologies while following industry standards, architecture, and best practices. - Act as a technical expert throughout the software development lifecycle, including designing, coding, unit testing, supporting, and debugging data warehouse software components. - Expertise in cloud and ETL engineering is key to solving problems and designing effective approaches. - Troubleshoot and debug ETL pipelines, optimize query performance, and create unit tests. - Collaborate with the Development Lead, DWH Architect, QA Engineers, and business analysts to plan, implement, and deliver efficient ETL strategies aligned with end-user requirements. - Create technical documentation, reports, and dashboards in the BI portal while supporting internal audit processes. Key Mandatory Skills: - Proven work experience as an ETL Developer. - Advanced knowledge of relational databases and dimensional Data Warehouse modeling. - Expertise in Microsoft Data stack with experience in Azure and Synapse Analytics. - Designing and implementing data transformation and ETL layers using tools like Data Factory and Notebooks. - Experience with PowerBI for report and dashboard creation. - Strong SQL knowledge for developing complex queries and working with stored procedures, views, indexes, etc. - Familiarity with CI/CD tools and principles, preferably Azure DevOps or Bamboo. - Proficiency in at least one scripting language, with Python as an advantage. - Experience with GIT repositories and version control tools like GitHub, Azure DevOps, or Bitbucket. - Working in Agile projects, preferably using JIRA. - Excellent problem-solving skills, communication abilities, and understanding of data governance concepts. Nice-to-Have Skills: - Microsoft Fabric. - Snowflake. - Background in SSIS / SSAS / SSRS. - Azure DevTest Labs, ARM templates. - Azure PurView. - Banking or finance industry experience. If you can work independently, collaborate effectively in a team environment, and communicate complex information clearly, your success in this role is ensured. If you have a passion for data engineering, a keen eye for detail, and a proactive approach to problem-solving, we encourage you to apply.,
ACTIVELY HIRING
posted 2 months ago

Data Engineer

Regnology
experience5 to 9 Yrs
location
Pune, Maharashtra
skills
  • Java
  • Scala
  • Python
  • Spark
  • GCP
  • Kubernetes
  • Docker
  • TrinoPresto
  • Apache Iceberg
  • Oracle SQLPLSQL
  • Parquet formats
  • CDC tools
  • JDBC connectors
  • AWSAzure cloud ecosystems
  • Helm charts
  • CICD pipelines
  • observability tools
Job Description
As an experienced Data Engineer at Regnology, you will play a crucial role in designing ingestion pipelines, optimizing query performance, and ensuring data quality, governance, and cost efficiency at scale. Your responsibilities will include: - **Migration Strategy & Execution**: - Design and implement data ingestion pipelines to extract data from Oracle into GCS/Iceberg. - Migrate and modernize existing Oracle schemas, partitions, and materialized views into Iceberg tables. - Define CDC (Change Data Capture) strategies using custom ETL. - **Data Lakehouse Architecture**: - Configure and optimize Trino clusters (coordinator/worker, Helm charts, autoscaling). - Design partitioning, compaction, and clustering strategies for Iceberg tables. - Implement schema evolution, time-travel, and versioning capabilities. - **Performance & Cost Optimization**: - Benchmark Trino query performance vs Oracle workloads. - Tune Trino/Iceberg for large-scale analytical queries, minimizing query latency and storage costs. - **Data Quality, Metadata & Governance**: - Integrate Iceberg datasets with metadata/catalog services (Postgre/Hive Metastore, or Glue). - Ensure compliance with governance, observability, and lineage requirements. - Define and enforce standards for unit testing, regression testing, and data validation. - **Collaboration & Delivery**: - Support existing reporting workloads (regulatory reporting, DWH) during and after migration. - Document architecture, migration steps, and provide knowledge transfer. **Why you should decide on us**: - Lets grow together, join a market-leading SaaS company - our agile character and culture of innovation enable you to design our future. - We provide you with the opportunity to take on responsibility and participate in international projects. - In addition to our buddy-program, we offer numerous individual and wide-ranging training opportunities during which you can explore technical and functional areas. - Our internal mobility initiative encourages colleagues to transfer cross-functionally to gain experience and promotes knowledge sharing. - We are proud of our positive working atmosphere characterized by a supportive team across various locations and countries and transparent communication across all levels. - Together we're better - meet your colleagues at our numerous team events. **Qualifications Required**: - 5+ years of experience. - Prior experience migrating financial/regulatory datasets. - Experience with Regulatory Reporting or similar enterprise workloads. - Familiarity with large-scale performance benchmarking and cost modeling. **Required Skills & Experience**: - **Core Expertise**: - Strong hands-on experience with Trino/Presto, Apache Iceberg, and Oracle SQL/PLSQL. - Proven experience with data lakehouse migrations at scale (50 TB+). - Proficiency in Parquet formats. - **Programming & Tools**: - Solid coding skills in Java, Scala, or Python for ETL/ELT pipeline development. - Experience with orchestration (Spark). - Familiarity with CDC tools, JDBC connectors, or custom ingestion frameworks. - **Cloud & DevOps**: - Strong background in GCP (preferred) or AWS/Azure cloud ecosystems. - Experience with Kubernetes, Docker, Helm charts for deploying Trino workers. - Knowledge of CI/CD pipelines and observability tools. - **Soft Skills**: - Strong problem-solving mindset with the ability to manage dependencies and shifting scopes. - Clear documentation and stakeholder communication skills. - Ability to work in tight delivery timelines with global teams. Regnology is a leading international provider of innovative regulatory, risk, and supervisory technology solutions, serving over 7,000 financial services firms with reporting solutions globally. The company offers a positive working atmosphere, numerous training opportunities, and promotes knowledge sharing among colleagues. If this challenging opportunity excites you, apply now at [Regnology Careers](https://www.regnology.net).,
ACTIVELY HIRING
posted 3 weeks ago

Data Engineer

Regnology Group GmbH
experience5 to 9 Yrs
location
Pune, Maharashtra
skills
  • Java
  • Scala
  • Airflow
  • Kubernetes
  • Docker
  • TrinoPresto
  • Apache Iceberg
  • Oracle SQLPLSQL
  • Parquet
  • PySpark
  • CICD tools
  • JDBC connectors
  • Helm charts
Job Description
As an experienced Data Engineer, your role will involve designing ingestion pipelines, optimizing query performance, and ensuring data quality, governance, and cost efficiency at scale. You will be working with large workloads of 50-300 TB. **Key Responsibilities:** - Migration Strategy & Execution - Design and implement data ingestion pipelines to extract data from Oracle into GCS/Iceberg. - Migrate and modernize existing Oracle schemas, partitions, and materialized views into Iceberg tables. - Define CDC (Change Data Capture) strategies using custom ETL. - Data Lakehouse Architecture - Configure and optimize Trino clusters (coordinator/worker, Helm charts, autoscaling). - Design partitioning, compaction, and clustering strategies for Iceberg tables. - Implement schema evolution, time-travel, and versioning capabilities. - Performance & Cost Optimization - Benchmark Trino query performance vs Oracle workloads. - Tune Trino/Iceberg for large-scale analytical queries, minimizing query latency and storage costs. - Data Quality, Metadata & Governance - Integrate Iceberg datasets with metadata/catalog services (Postgre/Hive Metastore, or Glue). - Ensure compliance with governance, observability, and lineage requirements. - Define and enforce standards for unit testing, regression testing, and data validation. - Collaboration & Delivery - Support existing reporting workloads (regulatory reporting, DWH) during and after migration. - Document architecture, migration steps, and provide knowledge transfer. **Required Skills & Experience:** - Core Expertise: - Strong hands-on experience with Trino/Presto, Apache Iceberg, and Oracle SQL/PLSQL. - Proven experience with data lakehouse migrations at scale (50 TB+). - Proficiency in Parquet formats. - Programming & Tools: - Solid coding skills in Java is a MUST for ETL/ELT pipeline development. Knowledge/Experience with PySpark(Python), Scala is a plus. - Experience with orchestration (Airflow). - Familiarity with CI/CD tools, JDBC connectors, or custom ingestion frameworks. - Cloud & DevOps: - Strong background in GCP (preferred) or AWS/Azure cloud ecosystems. - Experience with Kubernetes, Docker, Helm charts for deploying Trino workers. - Knowledge of CI/CD pipelines and observability tools. - Soft Skills: - Strong problem-solving mindset with the ability to manage dependencies and shifting scopes. - Clear documentation and stakeholder communication skills. - Ability to work in tight delivery timelines with global teams. In addition to the above responsibilities and requirements, the company Regnology is an international leader providing innovative solutions in Regulatory, Risk, and Supervisory Technology (RegTech/RiskTech/SupTech), AEOI, and tax reporting. With over 770 employees across 17 locations in 12 countries, Regnology offers a positive working atmosphere, encourages internal mobility, and values transparent communication. If you are passionate about data engineering and have the required skills and experience, this role at Regnology could be the next step in your career growth. Apply now to be part of a market-leading SaaS company and contribute to shaping the future within a supportive and innovative culture. For more information about Regnology, visit [Regnology Website](https://www.regnology.net). If you have any questions, feel free to reach out to us at recruiting@regnology.net.,
ACTIVELY HIRING
posted 2 months ago
experience8 to 12 Yrs
location
Pune, Maharashtra
skills
  • SQL
  • test automation
  • analytical skills
  • project management
  • Microsoft applications
  • Tableau
  • PySpark
  • Big data technologies
  • Collibra
  • Database structures
Job Description
As a Data Quality Sr. Analyst, your role involves ensuring that data sourced and provisioned by different areas meets all required Data Quality standards. You will be evaluating moderately complex and variable issues with substantial potential impact, weighing various alternatives and balancing potentially conflicting situations using multiple sources of information. Your responsibilities include assessing, evaluating, and analyzing data challenges, providing recommendations on their resolution, and collaborating with various teams and groups to develop subject matter expertise and knowledge of industry practices and standards. Key Responsibilities: - Partner with Enterprise Chief Data Office to contribute to the development of procedure guidelines and standards addressing Data Certification and Data Quality management - Develop methods to identify data patterns and trends in available data sources, perform data quality analytics, and identify data challenges - Contribute to Data Quality resolution strategies and practices - Assist in data improvement initiatives in compliance with Citi Data Governance Policy - Assist in the remediation of Audit and Regulatory feedback on issues and concerns - Report Data Quality issues through Citi's corresponding Data/Issue management process and coordinate efforts for Data consumption demand and Data requirement from strategic data repositories - Liaise with other areas of the firm to understand data challenges and solutions - Perform financial reconciliation and measure Data Quality to drive data quality resolution - Define and formulate business requirements documents that contribute to resolve Data Quality issues - Lead project management activities - Support senior management strategic vision Qualifications: - 8-12 years of experience with defining and implementing Data Quality programs; Banking or Finance industry preferred - Strong test automation skills - PySpark, SQL, testing of large DWH applications built on Big data technologies - Strong analytical skills and attention to detail - Strong project management skills - Strong SQL and test automation on BIG data technologies - Optimizes work processes by knowing the most effective and efficient processes to get things done, with a focus on continuous improvement - Proficient in using data management and analysis tools including basic Microsoft applications (Word, Excel, PowerPoint) - Working knowledge of Collibra and other Data Governance industry standard tools - Working knowledge of Tableau (report consumer, developer, etc.) - Demonstrated understanding of Database structures - Proven ability to provide in-depth and sophisticated analysis with interpretive thinking to define problems and develop innovative solutions - Anticipates and balances the needs of multiple stakeholders, while monitoring tight deadlines or unexpected requirement changes - Communicates effectively in non-technical terms at all levels of the organization - Collaborates effectively by building partnerships and working well with others to meet shared objectives As a Data Quality Sr. Analyst, you will play a crucial role in ensuring data quality within the organization, collaborating with various teams, and contributing to the strategic vision of senior management. The role requires a strong background in defining and implementing Data Quality programs, along with expertise in test automation, analytical skills, and project management.,
ACTIVELY HIRING
posted 1 week ago
experience4 to 8 Yrs
location
Pune, Maharashtra
skills
  • Ingestion
  • IoT Event Streaming
  • Data Transformation ETL
  • Cloud DWH Environment
Job Description
**Job Description:** As a Technical Project Delivery Specialist, your role will involve the following key responsibilities: - Accountable for delivering Product architectural strategy and vision within the data team and delivering IT Management solutions to realize business benefits. - Understand customer needs and work closely with the Business IT functions to validate value proposition and fit to requirements of products and services being formulated. - Develop hands-on, in-depth knowledge of competitive products and maintain technical analysis of competitive strengths and weaknesses. Maintain current understanding of relevant technologies and how these may impact in-flight innovation initiatives. - Work closely with Domain Data Product Managers to remain aligned with broader innovation initiatives, and with Enterprise Architects for target state alignment. Engage with multiple stakeholders to build the specification of new technologies driving innovative work practices with our business. Ensure detailed use cases and business requirements documented, working with the Demand and Operations Lead. - Build on successful proof-of-concepts to develop Minimum Viable Product (MVP) and manage live pilots which will see the integration of these technologies in our practice. Scale successful pilots into the mainstream operations of our business, following an agile methodology of refining or pivoting on initial ideas. - Serve as primary product expert for products under your own remit. - Lead product design post pilot, ensuring consideration of scalability, global deployment, and relevant architecture principles applied, and that platform / product can be developed in an evolutionary way. - Support the product lead in formulating the roadmap for products / platforms and funding continuum required. - Control product costs and vendor relationships. - Support the development of products and future releases with design and architecture guidance. - Quantify technical development and test effort for development tasks and liaise with Senior Product Manager and Demand and Operations lead to assign the right skills to product execution pod. - Develop and maintain reference architecture and design pattern documentation for products and platforms. - Manage stakeholders" expectations in relation to deliverables and identify risks on innovation initiatives as they are developed. - Assess the requirements and present architectural recommendations that are in line with the strategic vision for the enterprise and wider best practice. - Evaluate and provide feedback on available and emerging technologies to determine the potential impacts and business value. - Provide leadership to the solution engineer community and develop relationships with suppliers to allow effective and productive evaluation of vendor offerings against business requirements. - Perform solution options analysis and product selection, with a particular focus on Total Cost of Ownership. - Maintain a good working knowledge of business processes and associated functionality. - Management and promotion of Jaguar Land Rover IT central governance processes (Govern IT). - Undertake any other work as directed by your line manager in connection with your job as may be requested. **Qualifications Required:** **Essential:** - Hands-on experience of key technologies for the relevant product area (ingestion, IoT event streaming, Data Transformation ETL, Cloud DWH environments). - In-depth technology and coding background. - Ability to self-learn and cross-train on new technologies, innovation experience. - In-depth IT solutions architecture understanding and hands-on experience. - Product Management experience including delivery of entire technology-based products from start to finish. - Project delivery experience covering analysis, design, test, deployment, and operation of new products. - Ability to successfully interact with senior stakeholders. - Experience in the use of agile development methodologies. - Sound understanding of disruptive technologies and innovation. - Excellent oral and written communication skills. - Experience of working in a lead role in diverse teams, mentoring and coaching junior staff to maximize performance. **Desirable:** - International experience ideally working for Blue-chip organizations.,
ACTIVELY HIRING
logo

@ 2025 Shine.com | All Right Reserved

Connect with us:
  • LinkedIn
  • Instagram
  • Facebook
  • YouTube
  • Twitter