metadata-management-jobs-in-madurai, Madurai

1 Metadata Management Jobs nearby Madurai

Toggle to save search
posted 1 week ago

Social Media Manager

Bell Brand Studios
experience5 to 9 Yrs
location
Madurai, Tamil Nadu
skills
  • Social Media Management
  • Content Strategy
  • Analytics
  • SEO
  • Graphic Design
  • Video Editing
  • Communication Skills
  • Audience Engagement
Job Description
Role Overview: As a Social Media Manager at Bell Brand Studios in Madurai, you will be responsible for managing and growing the brand's presence across various social media platforms. You will be expected to develop and execute content calendars, stay updated on algorithm changes, create engaging content tailored to each platform, monitor performance metrics, engage with the audience, and collaborate with the marketing team for campaign execution. Your role will involve implementing SEO best practices and managing paid promotions to maximize organic reach and community loyalty. Key Responsibilities: - Manage and grow brand presence across Instagram, Facebook, YouTube, and other emerging platforms. - Develop and execute content calendars for brand consistency and visual storytelling. - Stay updated on algorithm changes, platform tools, and content trends to optimize strategy. - Create, schedule, and publish engaging content tailored to each platform. - Use tools like Meta Business Suite, Creator Studio, and YouTube Studio to manage accounts seamlessly. - Monitor performance metrics, provide reports with insights, and recommendations. - Engage with audiences, respond to comments/messages, and build community loyalty. - Collaborate with designers, videographers, and the marketing team for campaign execution. - Implement SEO best practices in content, hashtags, and video metadata. - Manage paid promotions and boost campaigns in coordination with the performance marketing team. Qualification Required: - 5+ years of experience in managing social media accounts for brands. - Strong command over Instagram, Facebook, YouTube, and platform-specific strategies. - Proven track record of growing pages and driving engagement organically and via paid campaigns. - Deep understanding of social media algorithms, formats, and engagement triggers. - Basic graphic design or video editing knowledge is a plus. - Strong writing and communication skills. - SEO knowledge and experience optimizing content for better discoverability. - Ability to multitask, meet deadlines, and manage multiple accounts/brands simultaneously. Why Work With Us: Join a vibrant, design-forward team at Bell Brand Studios where your creative strategies shape how people engage with food, lifestyle, and digital culture. We value original thinkers, trend-hunters, and digital storytellers who want to make bold content that leaves a scroll-stopping impact.,
ACTIVELY HIRING

Top Companies are Hiring in Your City

For Multiple Roles

Jio Platforms Ltd
Jio Platforms Ltdslide-preview-Genpact
posted 2 weeks ago

Abinitio Developer

CAPGEMINI TECHNOLOGY SERVICES INDIA LIMITED
CAPGEMINI TECHNOLOGY SERVICES INDIA LIMITED
experience5 to 10 Yrs
location
Chennai, Bangalore+7

Bangalore, Noida, Hyderabad, Gurugram, Kolkata, Pune, Mumbai City, Delhi

skills
  • ab initio
  • unix shell scripting
  • sql
Job Description
Key Responsibilities Design, develop, and implement ETL processes using Ab Initio GDE (Graphical Development Environment). Build and maintain Ab Initio graphs, plans, and sandboxes for data extraction, transformation, and loading. Work with business teams to understand data integration requirements and deliver efficient solutions. Use Ab Initio EME for version control, dependency management, and metadata governance. Perform data profiling, data validation, and quality checks using Ab Initio components and tools. Optimize ETL workflows for performance, scalability, and maintainability. Implement robust error handling, restartability, and logging mechanisms. Collaborate with DBAs, data modelers, and analysts to ensure data accuracy and consistency. Schedule and monitor jobs using Ab Initio Control Center (AICC) or enterprise schedulers. Support production systems, troubleshoot issues, and perform root cause analysis. Required Technical Skills Strong hands-on experience in Ab Initio GDE, EME, Co>Operating System, and Control Center. Proficiency with Ab Initio components such as Input/Output, Transform, Partition, Sort, Join, Lookup, Rollup, Reformat, Scan, and Dedup Sort, along with error handling using Rejects, Error Tables, and Error Ports for robust ETL design. Expertise in ETL design, development, and deployment for large-scale data environments. Proficiency in SQL and relational databases such as Oracle, Teradata, DB2, or SQL Server. Experience with UNIX/Linux shell scripting for automation and workflow integration. Understanding of data warehousing concepts (star schema, snowflake schema, slowly changing dimensions). Strong performance tuning and debugging skills in Ab Initio. Familiarity with data quality, metadata management, and data lineage.  
INTERVIEW ASSURED IN 15 MINS
posted 3 weeks ago
experience3 to 7 Yrs
location
Chennai, All India
skills
  • SQL
  • Alteryx
  • Python
  • Data Analysis
  • Statistical Modeling
  • Metadata Management
  • Data Modeling
  • DataIKU
Job Description
As a Dataiku Developer, you will be responsible for hands-on experience in DataIKU and SQL, translating business needs into data models, and developing efficient Dataiku workflows by applying best practices. You must have experience in converting SQL/Alteryx scripts in Dataiku and executing flowzone development plans for existing Alteryx flows/SQL scripts. Additionally, you will work with large datasets, manage huge volumes of data, and evaluate multiple data sources. Your role will involve designing and developing KPIs for business decision making, data gathering, exploration, transformation, analysis, mining, and ensuring data quality. Key Responsibilities: - Hands-on experience with DataIKU and SQL/Python functions & programs - Optimization of existing flowzones to save time & improve process efficiency - Efficient use of global variables and procedures to simplify the coding process - Writing complex queries on multiple tables to create a unified view of the data - Converting existing Alteryx workflows/SQL scripts in Dataiku - Preparing technical specifications and documentation for supporting BI reports Qualifications Required: - 3+ years of experience in SQL - Hands-on experience with DataIKU and must have a certification as a Core Designer - Knowledge of databases, statistical modeling, and the ability to develop complex models - Good data analysis skills in identifying different trends, patterns, and other data anomalies - Ability to write macros, advanced queries, models, functions, and formulae - Collaborating with client leads to coordinate the data extraction process and discussing results with engagement teams - Experience in the Snowflake environment would be advantageous Please note that the job description does not include any additional details about the company. As a Dataiku Developer, you will be responsible for hands-on experience in DataIKU and SQL, translating business needs into data models, and developing efficient Dataiku workflows by applying best practices. You must have experience in converting SQL/Alteryx scripts in Dataiku and executing flowzone development plans for existing Alteryx flows/SQL scripts. Additionally, you will work with large datasets, manage huge volumes of data, and evaluate multiple data sources. Your role will involve designing and developing KPIs for business decision making, data gathering, exploration, transformation, analysis, mining, and ensuring data quality. Key Responsibilities: - Hands-on experience with DataIKU and SQL/Python functions & programs - Optimization of existing flowzones to save time & improve process efficiency - Efficient use of global variables and procedures to simplify the coding process - Writing complex queries on multiple tables to create a unified view of the data - Converting existing Alteryx workflows/SQL scripts in Dataiku - Preparing technical specifications and documentation for supporting BI reports Qualifications Required: - 3+ years of experience in SQL - Hands-on experience with DataIKU and must have a certification as a Core Designer - Knowledge of databases, statistical modeling, and the ability to develop complex models - Good data analysis skills in identifying different trends, patterns, and other data anomalies - Ability to write macros, advanced queries, models, functions, and formulae - Collaborating with client leads to coordinate the data extraction process and discussing results with engagement teams - Experience in the Snowflake environment would be advantageous Please note that the job description does not include any additional details about the company.
ACTIVELY HIRING
question

Are these jobs relevant for you?

posted 2 weeks ago
experience13 to 17 Yrs
location
Chennai, Tamil Nadu
skills
  • coordination
  • communication
  • YouTube CMS management
  • OTT delivery
  • digital media operations
  • content workflows
Job Description
As a Content Manager/Digital Operations Executive at SNS Movie Production LLP, your role will involve managing digital content workflows across YouTube CMS, streaming platforms, dashboards, and OTT partner deliveries. Your responsibilities will include: - **CMS & Streaming Platform Management**: - Maintain and manage the YouTube CMS Suite and other streaming platform dashboards. - Oversee uploads, metadata accuracy, channel hygiene, content rights, and claims. - Monitor platform analytics and ensure compliance with platform policies. - Troubleshoot technical issues related to CMS or streaming platforms. - **Dashboard Maintenance**: - Maintain internal dashboards and ensure real-time data accuracy. - Generate, update, and share performance analytics and reports with stakeholders. - Oversee access permissions and ensure dashboards are optimized for team usage. - **Content Delivery to OTT Platforms**: - Manage end-to-end movie/content delivery to all OTT partners. - Coordinate closely with internal production, technical, and QC teams. - Track delivery timelines and ensure on-time publishing. - **Agreements & Documentation**: - Handle agreements related to movie associations and partner collaborations. - Maintain proper documentation and contract records. - Coordinate with legal and management teams for approval workflows. Qualifications Required: - Bachelor's degree in Media, Communication, Film Studies, Marketing, or related field. - 3 years of experience in YouTube CMS management, OTT delivery, digital media operations, or similar roles. - Strong understanding of YouTube CMS, digital rights management, and content workflows. - Excellent coordination and communication skills. - Ability to manage multiple projects with tight deadlines. - Basic familiarity with agreements, documentation, and data dashboards. At SNS Movie Production LLP, you will work at the center of digital media and film content distribution. You will have the opportunity to collaborate with creative teams, streaming partners, and industry professionals, and grow in digital operations, content strategy, and OTT ecosystems.,
ACTIVELY HIRING
posted 2 months ago
experience5 to 9 Yrs
location
Chennai, Tamil Nadu
skills
  • information management
  • data management
  • cataloguing
  • metadata
  • business process analysis
  • stakeholder engagement
  • information governance
  • SQL
  • Python
  • search
  • taxonomy development
  • unstructured content management
  • workshop facilitation
  • catalog
  • classification platforms
  • documentation skills
  • retrieval patterns
  • AI
  • metadata strategies
Job Description
As an Information Management and Governance Business Analyst, your role will be pivotal in the data transformation journey of the company. You will focus on unstructured content and ensure its readiness for AI by implementing robust classification, metadata quality, and governed processes. Your collaboration with business, technical, and data governance teams will be crucial in identifying, cataloguing, and governing critical information to enhance its quality, accessibility, and innovation readiness. **Key Responsibilities:** - Plan and conduct workshops and interviews with business, technical, and data governance stakeholders for requirements gathering. - Elicit, document, and prioritize requirements using frameworks like MoSCoW or RICE for identifying, curating, and enriching critical information. - Develop epics, user stories, and acceptance criteria to support the information management and governance processes. **Qualifications Required:** - Demonstrated experience in information or data management, cataloguing, metadata, and taxonomy development, and unstructured content management. - Proven capability in business process analysis, workshop facilitation, and stakeholder engagement across business, IT, and governance teams. - Strong track record of supporting digital transformation and information management initiatives within large complex corporate environments. - Hands-on experience with information governance and catalog or classification platforms such as Microsoft Purview, Collibra, Alation, OpenText, SharePoint, Teams, OneDrive, Box, or Google Workspace. - Strong documentation skills including business and functional requirements documents, user requirement specifications, agile user stories in tools like Jira or Azure DevOps, and support for testing and user acceptance. - Strong communication, influencing, and change management skills with the ability to translate between business and technical audiences. - Basic knowledge of SQL or Python for profiling and quality sampling, along with experience in search or retrieval patterns for AI. - Experience in industries such as life sciences, financial services, or the public sector, and familiarity with GxP regulations for pharmaceutical contexts. - Relevant certifications such as CDMP, IGP, CIP, Agile BA, CBAP, Prosci, Change Management, PRINCE2, or Scrum are a plus.,
ACTIVELY HIRING
posted 2 weeks ago
experience3 to 7 Yrs
location
Chennai, Tamil Nadu
skills
  • Scala
  • Spark
  • ETL
  • Git
  • SonarQube
  • Data Engineering
  • Performance Tuning
  • Memory Management
  • Databricks
  • Polars
  • DuckDB
  • CICD
  • Workflow Optimization
  • Azure Cloud
Job Description
Role Overview: You will be responsible for designing, building, and optimizing large-scale ETL and data-processing pipelines within the Databricks ecosystem. Your role will involve migrating selected workloads to high-performance engines like Polars and DuckDB while maintaining strong engineering rigor. You will need to apply analytical thinking to address data reliability, performance, and scalability issues. Familiarity with AI concepts will be advantageous in this role. Key Responsibilities: - Develop and maintain distributed data pipelines using Scala, Spark, Delta, and Databricks. - Engineer robust ETL workflows for high-volume ingestion, transformation, and publishing. - Profile pipelines, identify and eliminate bottlenecks, and optimize compute, storage, and job orchestration. - Lead the migration of suitable workloads to high-performance engines such as Polars, DuckDB, or equivalent. - Implement CI/CD workflows with automated builds, tests, deployments, and environment gating. - Enforce coding standards through code coverage targets, unit/integration tests, and SonarQube rules. - Ensure pipeline observability by implementing logging, data quality checks, lineage, and failure diagnostics. - Apply analytical reasoning to troubleshoot complex data issues and provide root-cause clarity. - Collaborate with product, analytics, and platform teams to operationalize data solutions. Qualification Required: - 3+ years of experience in data engineering with a strong command of Scala and Spark. - Proven background in ETL design, distributed processing, and high-volume data systems. - Hands-on experience with Databricks, including jobs, clusters, notebooks, and Delta Lake. - Proficiency in workflow optimization, performance tuning, and memory management. - Experience with high-performance engines like Polars, DuckDB, or similar columnar/accelerated engines. - Knowledge of CI/CD discipline using Git-based pipelines, strong testing practices, and code-quality standards. - Familiarity with SonarQube, coverage metrics, and static analysis. - Strong analytical and debugging capabilities across data, pipelines, and infrastructure. - Exposure to AI concepts such as embeddings, vector stores, retrieval-augmented generation, fine-tuning, and agentic architectures. Additional Details: Experience with Azure cloud environments and metadata-driven or config-driven pipeline frameworks is preferred.,
ACTIVELY HIRING
posted 1 week ago

Data Test Manager

RCG Global Services
experience5 to 15 Yrs
location
Tamil Nadu
skills
  • SQL
  • Data transformation
  • Spark
  • Hive
  • Data Governance
  • Metadata Management
  • AWS
  • ETLELT pipelines
  • Reconciliation processes
  • Data testing tools
  • Big Data Hadoop
  • Cloud Platforms Snowflake
  • Redshift
  • BigQuery
  • Azure Synapse
  • Databricks
  • BI validation
  • CICD integration
  • Program
  • people management
  • Stakeholder communication
  • Test strategy alignment
  • Analytical
  • problemsolving mindset
  • Influence
  • drive quality culture
  • Cloud technologies Azure
  • Google Cloud
  • API Mobile Testing
  • Insurance industry experience
  • Project Management Certification
  • Training
  • Certification in Data Quality Management
  • DataOps Cloud Data Platforms
  • Experience in implementing AIMLdr
Job Description
As a Data Test Manager at Myridius, your role involves strategic ownership of defining and implementing the overall Data Testing and Quality Assurance strategy for enterprise data platforms, data lakes, and data warehouses. Your responsibilities include: - Establishing standardized data testing frameworks, methodologies, and quality governance processes across programs - Managing and overseeing multiple data testing projects to ensure quality, schedule adherence, and efficient resource utilization - Collaborating with Data Architects, Engineering, Product, and Business stakeholders to align data testing with enterprise data initiatives - Driving end-to-end test planning, execution, and sign-off across ETL, data migration, integration, and BI validation layers - Defining automation-first strategy using tools like DataGaps, Tricentis Data Integrity, QuerySurge, dbt tests, or custom frameworks - Working with DevOps teams to embed automated data testing within CI/CD pipelines for continuous data quality checks - Tracking KPIs, DQ metrics, and ROI for testing initiatives and leading and mentoring Data Test Leads and Engineers - Identifying opportunities for process optimization, cost reduction, and test efficiency gains through automation and standardization - Ensuring testing alignment with data governance, security, and compliance standards and managing relationships with testing tool vendors Qualifications Required: - Minimum 5+ years of experience in QA with at least 5+ years in Data Testing leadership or Data Quality Management roles - Proven experience in managing large-scale data testing programs across data lakes, warehouses, and integration ecosystems - Strong hands-on experience in SQL, ETL/ELT pipelines, data transformation, and reconciliation processes - Hands-on experience or strong knowledge of data testing tools such as DataGaps, Tricentis DI, QuerySurge, dbt tests, Apache Griffin, or Great Expectations - Experience with Big Data (Hadoop, Spark, Hive) and Cloud Platforms (Snowflake, Redshift, BigQuery, Azure Synapse, Databricks) - Exposure to BI validation across tools like Tableau, Power BI, Looker - Knowledge of CI/CD integration for automated data testing pipelines and Data Governance and Metadata Management principles About RCG Global Services: Formerly known as RCG Global Services, Myridius offers tailored solutions in AI, data analytics, digital engineering, and cloud innovation to drive significant growth and maintain a competitive edge in the global market. With a commitment to exceeding expectations and fostering sustainable innovation, Myridius collaborates with clients to co-create transformative outcomes in a world of continuous change. Visit www.myridius.com to start leading the change.,
ACTIVELY HIRING
posted 7 days ago

Data Management Associate

The Depository Trust & Clearing Corporation (DTCC)
experience4 to 8 Yrs
location
Chennai, Tamil Nadu
skills
  • Data Management
  • Data Analysis
  • Business Requirements
  • Data Modeling
  • Metadata Management
  • Data Quality
  • Project Management
  • Presentation Skills
  • Agile Methodology
  • SQL
Job Description
As a member of the Data Management Center of Excellence at DTCC, you will play a crucial role in supporting the leadership team to establish, mature, and scale the CoE, as well as drive the implementation of industry-leading data management practices. Your ability to prioritize, own and drive projects, quickly learn new concepts and frameworks, work in a fast-paced environment, and adapt to shifting priorities will be key to your success in this role. You will be responsible for providing critical support for the management of information domains across the enterprise and supporting enterprise-wide data management initiatives. **Key Responsibilities:** - Participate in preliminary data analysis for new projects - Collaborate with business and operational partners to analyze data and produce business data element names - Engage with domain experts to understand and document business requirements for data - Assist data modelers in producing conceptual and logical data models - Input and handle required metadata for critical data elements in the enterprise metadata repository - Support data quality through data profiling, defining data quality rules, and supporting data cleansing efforts - Assist the Data Steward with project planning, identification of resource requirements, and effort estimation - Independently manage assigned projects across the life cycle - Develop and deliver high-quality presentations to partners and leadership - Participate as a core member of agile squads, including attending daily stand-up and grooming sessions, documenting user stories, and tracking project metrics In addition to the responsibilities mentioned above, you will play a key role in bridging the gap between IT and the business, accessing, profiling, and analyzing data, understanding and documenting business requirements, and communicating effectively with technical and business stakeholders. **Qualifications:** - Minimum 4 years of related experience - Bachelor's degree preferred or equivalent experience **Talents Needed for Success:** - Solid grasp of data management processes and standard methodologies - Excellent written and verbal communication skills - Excellent analytical and quantitative skills - Proficiency with Microsoft Office tools required; experience with data management tools (e.g. Collibra) and business process tools (e.g. Mega HOPEX) preferred - Basic technical skills required, including the ability to profile data using SQL queries; basic understanding of Logical Data Modeling and Database design is a plus,
ACTIVELY HIRING
posted 1 month ago
experience0 to 3 Yrs
location
Chennai, Tamil Nadu
skills
  • Document Control
  • iDocs
  • Technical Skills
  • Metadata Management
  • MS Office
  • Project Information Management
  • Electronic Database Management System EDMS
  • Adobe
Job Description
As an Associate Document Controller, you will play a crucial role in supporting the Document Control group by following defined procedures and developing new skills in Document Control. Your responsibilities will include: - Assisting the Lead Document Management lead in day-to-day activities - Gaining experience in iDocs and developing basic technical skills - Tracking, logging, and distributing releases of internal and external documentation accurately - Archiving project records - Indexing and distributing project documents in a timely manner - Familiarizing yourself with contract and Project Information Management requirements - Reading and implementing project-specific information management procedures and work instructions You will report to the Project Lead Document Management and the Department Manager. In your role, you will liaise with the Lead Document Management, Project Engineer, Project Manager, and Department Manager. To qualify for this role, you should have: - A Bachelor's Degree in any stream or equivalent diploma (preferably in Computer Science) - Entry level to 3 years of Document Management experience - Good knowledge of Electronic Database Management System (EDMS) software, MS Office, and Adobe - Proficiency in English (both verbal and written) - Experience with EDMS and MS Office applications This position offers the opportunity to develop your Document Control skills and contribute to the efficient management of project documents.,
ACTIVELY HIRING
posted 2 months ago

Data Governance Lead

OptiSol Business Solutions
experience10 to 14 Yrs
location
Chennai, Tamil Nadu
skills
  • Data Governance
  • Data Architecture
  • Data Modeling
  • Data Migration
  • Data Integration
  • Data Visualization
  • Data Analytics
  • Data Security
  • Data Privacy
  • Metadata Management
  • Stakeholder Engagement
  • Communication Skills
  • Stakeholder Management
  • Analytical Skills
  • Agile Methodology
  • Snowflake
  • Problemsolving Skills
  • DAMADMBOK Principles
  • Data Modeling Tools
  • Data Governance Platforms
  • AWS Cloud Platform
  • Data Privacy Regulations
  • Security Frameworks
Job Description
As a Data Governance Lead, your role will involve driving enterprise-wide data governance, architecture, and management initiatives. You will play a crucial part in establishing data standards, ensuring data quality, and enabling secure, compliant, and value-driven data usage across the organization. **Key Responsibilities:** - Lead the implementation of data governance frameworks, policies, and standards. - Define and manage data ownership, stewardship roles, and accountability structures. - Ensure compliance with regulatory and privacy requirements (e.g., GDPR, SOX etc). - Design and maintain conceptual, logical, and physical data models. - Collaborate with business and technical teams to ensure models align with business needs. - Promote standardization and reuse of data assets across platforms. - Lead data migration strategies for legacy systems and cloud platforms. - Ensure data integrity, lineage, and quality during migrations. - Work with ETL and integration teams to streamline data flows. - Partner with BI teams to ensure data is structured for effective visualization. - Support dashboard and reporting initiatives with clean, well-governed data. - Promote self-service analytics through governed data catalogs. - Collaborate with InfoSec to implement data protection controls. - Ensure sensitive data is classified, encrypted, and accessed securely. - Monitor and audit data access and usage patterns. - Implement metadata management tools and practices. - Maintain data lineage documentation to support traceability and impact analysis. - Act as a liaison between business units, IT, and data teams. - Educate stakeholders on data governance principles and best practices. - Drive cultural change toward data-driven decision-making. **Required Qualifications:** - 10+ years of experience in data management, with at least 3 years in a data governance or architecture leadership role. - Excellent communication and stakeholder management skills. - Strong analytical and problem-solving abilities. - Experience in Agile project environments. - Experience in projects using Snowflake for data migration, data warehouse, and governance. - Strong understanding of DAMA-DMBOK principles. - Hands-on experience with data modeling tools (e.g., Erwin, PowerDesigner). - Experience with data governance platforms (e.g., Collibra, Informatica, Alation). - Familiarity with AWS cloud platform. - Knowledge of data privacy regulations and security frameworks. In this role, certifications such as CDMP (Certified Data Management Professional) are considered a plus.,
ACTIVELY HIRING
posted 2 weeks ago
experience3 to 7 Yrs
location
Chennai, Tamil Nadu
skills
  • Issue Resolution
  • Content Provider Management
  • SLA Adherence
  • Workflow Optimization
  • Data Fluency
  • Stakeholder Communication
  • Digital Content Licensing
  • AIpowered Discovery
  • STM Content
  • Product Development Methodologies
Job Description
As the Operational Onboarding Manager for third-party content partners at Elsevier, your primary role is to manage the end-to-end operational onboarding and life-cycle management of all third-party content partners contributing to Elsevier's AI products. You will act as the central orchestrator between partner organizations and internal teams such as Product, Engineering, Legal, and Finance to ensure smooth content flows, contract progression, and stakeholder alignment with ingestion targets. Key Responsibilities: - Manage day-to-day relationships and SLA adherence for approximately 100 live providers across journals, books, and OA aggregators - Ensure compliance with internal standards and timelines - Drive workflow optimization to reduce onboarding time and proactively identify operational inefficiencies - Track and manage content ingestion and delivery pipelines - Lead the rollout of automated contract flows and maintain playbooks for new partner types - Publish monthly Tableau dashboards on ingestion volume, partner health, and revenue share - Escalate risks early and maintain a transparent knowledge base for internal stakeholders - Support the creation and maintenance of a centralized knowledge base for stakeholder transparency and accountability Required Qualifications: - 3+ years of experience in content operations or partner enablement within publishing or SaaS - Proven program-management skills across functions - Working knowledge of Salesforce or comparable CRM workflows - Data fluency with the ability to interpret dashboards and spot anomalies - Strong stakeholder communication and issue-resolution abilities - Familiarity with digital content licensing concepts (rights, metadata, revenue share) Nice-to-Haves: - Experience with AI-powered discovery, STM content, or open-access aggregation - Exposure to product development methodologies and software like JIRA/Confluence Working Environment: - Location: Remote or hybrid within North America or Europe time zones - Travel: Less than 10% for occasional partner or team off-sites - Reports to Director of Product Management and partners daily with Business Development, Engineering, Legal, and Finance teams Elsevier offers you the opportunity to sit at the core of the strategy to transform trusted research content into AI-ready assets, enabling breakthrough discovery for millions of scientists worldwide. Your operational excellence will directly contribute to accelerating this mission. If you are ready to streamline how the world's research enters the AI era, apply with your resume and a brief note highlighting a process improvement you spearheaded in a prior role. Please be aware of potential scams during the hiring process and refer to the Candidate Privacy Policy for more information.,
ACTIVELY HIRING
posted 2 weeks ago
experience3 to 7 Yrs
location
Chennai, All India
skills
  • Java development
  • SQL
  • Data analysis
  • Production support
  • Informatica MDM
  • Informatica P360
  • ActiveVOS workflow management
Job Description
As an Informatica MDM-PIM Consultant at Infoya, your role will involve designing & configuring Informatica P360 solutions, Informatica MDM, Java development, SQL, and data analysis. Your expertise in production support and ActiveVOS workflow management will be crucial for delivering successful outcomes. **Roles & Responsibilities:** - Implement and manage Informatica Product 360 (iP360) solutions across various business units - Configure and deploy P360 on AWS/Azure provisioned environment - Integrate Informatica MDM with P360 for streamlined master data management - Create and manage catalogues - Profile source data and determine source data and metadata characteristics - Design and execute data quality mappings for data cleansing and de-duplication - Develop workflows using ActiveVO designer - Utilize IDQ designer to develop mapplets and data quality rules - Extract, transform, and load data using Informatica Data Integration platform - Develop data integration mapping designs - Implement ingestion and integration processes on Cloud technologies like AWS or GCP - Build Java-based custom components for integration and automation - Perform data transformation on XML or JSON documents - Unit test process flows for accuracy - Provide production support including monitoring, incident resolution, and system optimization **Qualifications:** **Technical Experience:** - 6+ years of experience as an Informatica Developer with MDM-PIM configuration experience - Proficient in Informatica P360 and Informatica Developer Components - Knowledgeable in Informatica Data Quality standards and best practices - Experience deploying P360 solutions on AWS and/or Azure - Familiarity with Java/J2EE technologies for SDK work - Strong database technology experience with SQL databases (Oracle, SQL Server, PostgreSQL, etc.) - Experience in defining and deploying data quality programs on enterprise projects - Data analyst experience including data profiling, mapping, validation, manipulation, and analysis - Workflows development experience using Informatica ActiveVOS - Ability to translate business problems into data quality initiatives - Familiar with Data Quality technologies trends As an Informatica MDM-PIM Consultant at Infoya, you will have the opportunity to work in a dynamic and collaborative environment, with competitive salary and benefits, professional development opportunities, and the chance to collaborate with a diverse and talented team. As an Informatica MDM-PIM Consultant at Infoya, your role will involve designing & configuring Informatica P360 solutions, Informatica MDM, Java development, SQL, and data analysis. Your expertise in production support and ActiveVOS workflow management will be crucial for delivering successful outcomes. **Roles & Responsibilities:** - Implement and manage Informatica Product 360 (iP360) solutions across various business units - Configure and deploy P360 on AWS/Azure provisioned environment - Integrate Informatica MDM with P360 for streamlined master data management - Create and manage catalogues - Profile source data and determine source data and metadata characteristics - Design and execute data quality mappings for data cleansing and de-duplication - Develop workflows using ActiveVO designer - Utilize IDQ designer to develop mapplets and data quality rules - Extract, transform, and load data using Informatica Data Integration platform - Develop data integration mapping designs - Implement ingestion and integration processes on Cloud technologies like AWS or GCP - Build Java-based custom components for integration and automation - Perform data transformation on XML or JSON documents - Unit test process flows for accuracy - Provide production support including monitoring, incident resolution, and system optimization **Qualifications:** **Technical Experience:** - 6+ years of experience as an Informatica Developer with MDM-PIM configuration experience - Proficient in Informatica P360 and Informatica Developer Components - Knowledgeable in Informatica Data Quality standards and best practices - Experience deploying P360 solutions on AWS and/or Azure - Familiarity with Java/J2EE technologies for SDK work - Strong database technology experience with SQL databases (Oracle, SQL Server, PostgreSQL, etc.) - Experience in defining and deploying data quality programs on enterprise projects - Data analyst experience including data profiling, mapping, validation, manipulation, and analysis - Workflows development experience using Informatica ActiveVOS - Ability to translate business problems into data quality initiatives - Familiar with Data Quality technologies trends As an Informatica MDM-PIM Consultant at Infoya, you will have the opportunity to work in a dynamic and collaborative environment, with competitive salary and benefits, professional development opportunities, and the chance to collaborate with a diverse and talented team.
ACTIVELY HIRING
posted 3 weeks ago
experience1 to 5 Yrs
location
Coimbatore, All India
skills
  • Jira
  • SAS 9x
  • Linux basics
  • SAS Metadata Management
Job Description
As a software Product Developer at TechUnity, Inc., you will be involved in engineering next generation systems through an agile process of ideation and recreation. You will also work on modernizing aging applications to current platforms, upkeeping operational systems to zero downtime, and developing disruptive technologies. Key Responsibilities: - Serve as the first point of contact for SAS platform issues, troubleshooting performance and configuration problems. - Manage user access, metadata updates, backups, and security settings. - Support server maintenance activities like patches, updates, and log reviews. - Document issues and resolutions in ticketing systems and contribute to a knowledge base. - Assist users with platform navigation and queries. Qualifications Required: - Education: Bachelor's in Computer Science, IT, or equivalent experience. - Skills: Familiarity with SAS 9.x, Linux basics, SAS Metadata Management, and ticketing systems (e.g., Jira). Basic scripting knowledge is a plus. - Certifications (Preferred): SAS Certified Specialist. TechUnity, Inc. is a high-tech Business Consultancy and software Product Developer headquartered in Washington, DC. The company implements growth strategies for technology companies and delivers systems using a wide range of technologies. TechUnity serves clients in various industries such as Retail and Commerce, Transportation and Logistics, Supply Chain, and more. They provide two service offerings - product development and business consulting, focusing exclusively on these areas. For enquiries, call (202) 747-3443 or email publicrelations@TechUnity.com. What's on Offer - Competitive salary, benefits, and relocation assistance. - Professional growth opportunities in emerging technologies like AI and analytics. As a software Product Developer at TechUnity, Inc., you will be involved in engineering next generation systems through an agile process of ideation and recreation. You will also work on modernizing aging applications to current platforms, upkeeping operational systems to zero downtime, and developing disruptive technologies. Key Responsibilities: - Serve as the first point of contact for SAS platform issues, troubleshooting performance and configuration problems. - Manage user access, metadata updates, backups, and security settings. - Support server maintenance activities like patches, updates, and log reviews. - Document issues and resolutions in ticketing systems and contribute to a knowledge base. - Assist users with platform navigation and queries. Qualifications Required: - Education: Bachelor's in Computer Science, IT, or equivalent experience. - Skills: Familiarity with SAS 9.x, Linux basics, SAS Metadata Management, and ticketing systems (e.g., Jira). Basic scripting knowledge is a plus. - Certifications (Preferred): SAS Certified Specialist. TechUnity, Inc. is a high-tech Business Consultancy and software Product Developer headquartered in Washington, DC. The company implements growth strategies for technology companies and delivers systems using a wide range of technologies. TechUnity serves clients in various industries such as Retail and Commerce, Transportation and Logistics, Supply Chain, and more. They provide two service offerings - product development and business consulting, focusing exclusively on these areas. For enquiries, call (202) 747-3443 or email publicrelations@TechUnity.com. What's on Offer - Competitive salary, benefits, and relocation assistance. - Professional growth opportunities in emerging technologies like AI and analytics.
ACTIVELY HIRING
posted 2 months ago
experience7 to 11 Yrs
location
Chennai, Tamil Nadu
skills
  • Data Architecture
  • Master Data
  • Integration
  • Transaction Processing
  • Data Governance
  • Data Migration
  • Metadata Management
  • Master Data Management
  • Data Governance
  • Data Processing
  • Data Models
  • Data Security
  • Data Integration
  • Technical Leadership
  • Solution Design
  • Data Warehouse
  • Data Modelling
  • Database Technologies
  • Data Strategy
  • Data Service Provider Platforms
  • Data Management Processes
  • Data Mart Structures
  • Relational Data Modelling
  • AI Methods
Job Description
Role Overview: As a Data Architect at Wipro Limited, your role is to define and develop Enterprise Data Structure, Data Warehouse, Master Data, Integration, and transaction processing while maintaining and strengthening modeling standards and business information. Key Responsibilities: - Partner with business leadership to provide strategic, information-based recommendations to maximize the value of data and information assets, while embracing innovation - Assess benefits and risks of data using tools such as business capability models to create a data-centric view - Create data strategy and roadmaps for Reference Data Architecture - Engage stakeholders to implement data governance models and ensure compliance with Data Modeling standards - Oversee and monitor frameworks to manage data across the organization - Provide insights for database storage and platform for ease of use - Collaborate with vendors to ensure system integrity and configuration - Develop high-level data migration plans and knowledge of Data service provider platforms - Collect, aggregate, match, consolidate, quality-assure, persist, and distribute data throughout the organization - Provide solutions for RFPs from clients and ensure overall implementation assurance - Develop direction to manage the portfolio of all databases and analyze technology environment and client requirements for collaboration solutions Qualifications Required: - Strong understanding of activities within primary disciplines such as Master Data Management, Metadata Management, and Data Governance - Experience in Conceptual, Logical, and physical database architectures, design patterns, best practices, and programming techniques - Ability to build and maintain relationships with delivery and practice leadership teams - Strong experience in developing and establishing technical and business process metrics - Ability to identify technical, process, and structural risks and prepare risk mitigation plans - Strong technical capability in presenting solution designs to clients and acting as a trusted advisor Additional Details: Wipro Limited is a leading technology services and consulting company focused on building innovative solutions for digital transformation. With over 230,000 employees and business partners across 65 countries, Wipro helps clients realize their boldest ambitions and build future-ready, sustainable businesses. Join Wipro to be a part of a business powered by purpose and a place that empowers you to design your own reinvention.,
ACTIVELY HIRING
posted 1 week ago
experience2 to 8 Yrs
location
Chennai, Tamil Nadu
skills
  • Analytical Skills
  • Data analysis
  • Data visualization
  • MS Excel
  • MS PowerPoint
  • MS Word
  • SAS
  • Python
  • SQL
  • Data Governance
  • Process improvement
  • Issue resolution
  • Communication skills
  • Interpersonal skills
  • Data Lineage Tools
  • Metadata Management Tools
  • Understanding of Enterprise Data Warehouse
  • Data definitions
  • Data discovery
  • Data quality framework
  • KPIs design
  • Control gaps identification
  • Finance Regulations knowledge
  • Retail Business understanding
  • Project management skills
Job Description
Role Overview: As a Data/Information Management Analyst 2 at Citigroup, you will be a developing professional who applies specialty area knowledge in monitoring, assessing, analyzing, and evaluating processes and data. You will be responsible for identifying policy gaps, formulating policies, interpreting data, making recommendations, and researching factual information. Your role will involve identifying inconsistencies in data, defining business issues, and formulating recommendations on policies, procedures, or practices. Additionally, you will integrate established disciplinary knowledge within your specialty area with a basic understanding of related industry practices. Key Responsibilities: - Responsible for various data activities including database administration, data analysis, maintenance, data quality, and database management or architecture/design engineering - Perform routine operational or administrative work with a focus on administering defined procedures, analyses, and report preparation - Develop working knowledge of industry practices and standards - Appropriately assess risk when making business decisions and demonstrate consideration for the firm's reputation and compliance with applicable laws and regulations - Engage resources outside of your direct control to achieve objectives Qualifications: - 0-2 years of relevant experience - Proficiency in analytic tools such as SAS E-miner, Knowledge Seeker, SPSS, etc. - Experience with big data and machine learning tools like R, Python, etc. - Ability to engage resources outside of direct control to achieve objectives - Bachelor's/University degree or equivalent experience About the Company: Citigroup's Analytics Information Management (AIM) is a global community dedicated to driving data-driven transformation across the organization. The AIM team works with Citi businesses and functions worldwide to create actionable intelligence for business leaders. Additional Details: The Data Management- Metadata and Data Lineage Team within AIM is responsible for managing the implementation of data quality measurement programs in the retail consumer bank. The team supports regulatory compliance, metrics design, data governance, issue management, and audit support. As a Data Steward Analyst, you will be responsible for defining, documenting, maintaining, and communicating metadata, performing data lineage, and supporting the use of metadata for data consumers. You will work with business partners to identify critical data elements, capture metadata, review business glossary, and lead efforts to implement data governance policies and procedures. Overall, this role requires expertise in analytical skills, data analysis, visualization, tools and platforms like data lineage and metadata management tools, MS Excel, PowerPoint, and Word, as well as a good understanding of data definitions, data quality framework, and data governance. Soft skills such as problem-solving, communication, project management, and the ability to work well across multiple functional areas are essential for success in this role. Education and Experience: - MBA/Master's degree in Economics/Statistics/Mathematics/Information Technology/Computer Applications/Engineering or equivalent - Post Graduate in Computer Science, Mathematics, Operations Research, Econometrics, Management Science, or related fields (preferred) - 2 to 8 years of hands-on experience in delivering data quality solutions, with at least 2 years of experience in Banking Industry.,
ACTIVELY HIRING
posted 2 months ago
experience3 to 7 Yrs
location
Chennai, Tamil Nadu
skills
  • Metadata Management
  • SQL
  • PLSQL
  • Python
  • Oracle EDMCS
  • Smart Push
  • Smart View
  • EPM Automate
  • AWS Cloud
  • Lambda
  • Step Functions
  • EventBridge
Job Description
Role Overview: As a member of Capgemini, you will have the opportunity to develop the Oracle EDMCS module, from translating requirements to design, build, testing, and training. Your responsibilities will include maintaining Metadata, Smart Push, Smart view reports, and objects. Additionally, you will be involved in designing and creating data forms, business rules, reports, and automation using EPM Automate. Collaboration with various technical experts within the organization will be crucial for implementing within the Enterprise Performance Management (EPM) system. Writing and analyzing SQL, PL/SQL scripts will also be part of your role. Key Responsibilities: - Develop Oracle EDMCS module including translating requirements, design, build, testing, and training - Maintain Metadata, Smart Push, Smart view reports, and objects - Design and create data forms, business rules, reports, and automation using EPM Automate - Collaborate with technical experts for the implementation within the Enterprise Performance Management (EPM) system - Write and analyze SQL, PL/SQL scripts Qualifications Required: - Experience in Implementation in EDMCS Modules - Proven ability to collaborate with internal clients in an agile manner, leveraging design thinking approaches - Experience with Python, AWS Cloud (Lambda, Step functions, EventBridge etc.) is preferred About Capgemini: Capgemini is a global business and technology transformation partner, helping organizations accelerate their dual transition to a digital and sustainable world. With a diverse team of over 340,000 members in more than 50 countries, Capgemini is trusted by clients to unlock the value of technology to address the entire breadth of their business needs. The company delivers end-to-end services and solutions leveraging strengths from strategy and design to engineering, all fueled by market-leading capabilities in AI, generative AI, cloud, and data, combined with deep industry expertise and partner ecosystem.,
ACTIVELY HIRING
posted 2 weeks ago
experience5 to 9 Yrs
location
Chennai, All India
skills
  • SAP BODS
  • Data Engineering
  • SAP
  • SQL
  • PLSQL
  • Data Migration
  • Data Cleansing
  • Data Transformation
  • AWS
  • Azure
  • GCP
  • Data Warehousing
  • Data Visualization
  • Agile Methodologies
  • ETL Development
  • Data Quality Management
  • Cloud Platforms
  • ETL Design
  • SQL Performance Tuning
  • AI OPS
  • ML Models
  • API Solutions
Job Description
As a Senior Engineer in SAP BODS, your role will involve the following key responsibilities: - Designing and implementing data integration solutions across SAP and non-SAP systems - Hands-on experience in SAP BODS, including data flows, workflows, transforms, and job scheduling - Proficiency in ETL design and optimization for handling large volumes of structured and semi-structured data - Experience with data migration, cleansing, and transformation across Enterprise Data Warehouse systems - Solid understanding of SQL, PL/SQL, and database performance tuning - Knowledge of data quality management, error handling, and exception reporting - Exposure to cloud data platforms such as AWS Redshift, Azure Synapse, and GCP BigQuery - Familiarity with data visualization tools like Power BI, Tableau, or SAP BO - Understanding of data governance, metadata management, and compliance standards - Working experience in Agile/Scrum methodologies - Exposure to leveraging AI OPS for predictive analytics and operational efficiency - Experienced in deploying and maintaining ML models in production environments - Familiarity with design and implementation of scalable API solutions Qualifications required for this role include: - Bachelors or Masters degree in Computer Science, Information Systems or a related field - SAP Certified Application Associate Data Services is preferred - Cloud Certifications (AWS, Azure, or GCP) at Associate level or higher - 5+ years of experience in data engineering or ETL development, with at least 3 years of hands-on experience in SAP BODS - Exposure to large-scale enterprise data environments and agile delivery models If you are looking for an exciting opportunity in Chennai and have the relevant experience and qualifications, we are seeking immediate joiners who can contribute effectively to our team. As a Senior Engineer in SAP BODS, your role will involve the following key responsibilities: - Designing and implementing data integration solutions across SAP and non-SAP systems - Hands-on experience in SAP BODS, including data flows, workflows, transforms, and job scheduling - Proficiency in ETL design and optimization for handling large volumes of structured and semi-structured data - Experience with data migration, cleansing, and transformation across Enterprise Data Warehouse systems - Solid understanding of SQL, PL/SQL, and database performance tuning - Knowledge of data quality management, error handling, and exception reporting - Exposure to cloud data platforms such as AWS Redshift, Azure Synapse, and GCP BigQuery - Familiarity with data visualization tools like Power BI, Tableau, or SAP BO - Understanding of data governance, metadata management, and compliance standards - Working experience in Agile/Scrum methodologies - Exposure to leveraging AI OPS for predictive analytics and operational efficiency - Experienced in deploying and maintaining ML models in production environments - Familiarity with design and implementation of scalable API solutions Qualifications required for this role include: - Bachelors or Masters degree in Computer Science, Information Systems or a related field - SAP Certified Application Associate Data Services is preferred - Cloud Certifications (AWS, Azure, or GCP) at Associate level or higher - 5+ years of experience in data engineering or ETL development, with at least 3 years of hands-on experience in SAP BODS - Exposure to large-scale enterprise data environments and agile delivery models If you are looking for an exciting opportunity in Chennai and have the relevant experience and qualifications, we are seeking immediate joiners who can contribute effectively to our team.
ACTIVELY HIRING
posted 1 week ago
experience10 to 14 Yrs
location
Chennai, Tamil Nadu
skills
  • Data Analytics
  • Data Science
  • Machine Learning
  • NLP
  • Anomaly Detection
  • Customer Segmentation
  • Data Governance
  • BI Tools
  • Data Quality
  • Metadata Management
  • Stakeholder Management
  • AI Enablement
  • Generative AI
  • Recommendation Systems
  • Predictive Models
  • Structured Data Enablement
  • AIML Concepts
  • Data Platforms
  • SaaS Metrics
  • PLG
  • UsageBased Pricing Strategies
  • Product Instrumentation
  • Event Tracking Platforms
Job Description
Role Overview: You are a visionary and hands-on Director of Data Analytics, Science & AI Enablement at Chargebee. Your primary responsibility is to lead the creation and growth of a data function that powers enterprise-wide AI initiatives. You will be instrumental in designing, building, and leading a cross-functional team responsible for enterprise data analytics, data science, data governance, and structured data enablement to support advanced AI/ML use cases. As a strategic partner to business and technology executives, you will enable insight-driven decision-making and scalable AI applications through modern data architecture and practices. Key Responsibilities: - Lead the development and deployment of machine learning, generative AI, recommendation systems, and predictive models to improve product intelligence and automation. - Build and scale AI capabilities across the platform, including personalization, NLP, anomaly detection, and customer segmentation. - Drive insights into user behavior, product performance, churn prediction, and lifecycle value using customer and usage data. - Develop dashboards, KPIs, and self-service analytics tools for marketing, product, sales, and support teams. - Build and lead a high-performance team of data scientists, AI/ML engineers, analysts, and data product managers. - Ensure scalable data architecture and high-quality data pipelines in collaboration with Data Engineering. - Oversee data quality, governance, and compliance across all analytical and operational systems. - Drive data availability, quality, and governance across the organization to support AI and advanced analytics initiatives. - Lead the development of dashboards, metrics, and decision-support tools that empower business leaders. Qualification Required: - Bachelors or Masters degree in Computer Science, Statistics, Data Science, Engineering, or related discipline. - Proven experience working in a SaaS or tech environment with subscription-based metrics. - 10+ years of experience in data analytics, data science, or related fields, with at least 3-5 years in a leadership capacity. - Strong knowledge of AI/ML concepts, data platforms (e.g., Snowflake, Databricks), and BI tools (e.g., Tableau, Power BI). - Deep understanding of data governance, data quality, and metadata management. - Demonstrated ability to lead change in a global, fast-paced, and matrixed environment. - Excellent communication and stakeholder management skills. - Prior experience with product instrumentation and event tracking platforms. - Experience scaling data teams in a high-growth or startup environment.,
ACTIVELY HIRING
posted 2 months ago

Document Management Assistant

Namlabs technologies pvt ltd
experience1 to 5 Yrs
location
Chennai, Tamil Nadu
skills
  • System Management
  • Document Control
  • Security
  • Document Organization
  • Policy
  • Procedure Development
  • Data Entry
  • Processing
  • Auditing
  • Compliance
  • Training
  • Support
  • Metadata
  • Classification
Job Description
Role Overview: As a Document Management Specialist, your main responsibility will be developing and implementing systems for organizing, storing, and retrieving documents in a digital format. You will also be involved in creating and maintaining document management policies, procedures, and workflows. Additionally, you will administer and maintain document management systems (DMS), develop procedures for data entry and processing, conduct regular audits for compliance, train employees on document management procedures, and provide support to end-users. You will also be responsible for managing document versions, metadata, classification systems, and ensuring the security of digital documents and information. Key Responsibilities: - Develop and implement systems for organizing, storing, and retrieving documents in a digital format - Create and maintain document management policies, procedures, and workflows - Administer and maintain document management systems (DMS) - Develop procedures for data entry and processing - Conduct regular audits to ensure compliance with policies and procedures - Train employees on document management procedures and provide support to end-users - Manage document versions, ensure the correct version is in use, and track changes - Develop and maintain document metadata and classification systems - Ensure the security of digital documents and information, including access control and data encryption Qualifications Required: - Previous experience in document management or related field - Strong understanding of document management systems and procedures - Excellent organizational and time management skills - Attention to detail and accuracy - Knowledge of metadata and classification systems - Familiarity with data security practices (Note: The document does not provide any additional details about the company.),
ACTIVELY HIRING
posted 3 weeks ago
experience10 to 14 Yrs
location
Chennai, Tamil Nadu
skills
  • Data Analytics
  • Data Science
  • Machine Learning
  • NLP
  • Anomaly Detection
  • Customer Segmentation
  • Data Governance
  • Data Engineering
  • Data Quality
  • Compliance
  • BI Tools
  • Metadata Management
  • Stakeholder Management
  • AI Enablement
  • Generative AI
  • Recommendation Systems
  • Predictive Models
  • Structured Data Enablement
  • Data Privacy Regulations
  • AIML Solutions
  • Data Platforms
  • SaaS Metrics
  • PLG
  • UsageBased Pricing Strategies
  • Product Instrumentation
  • Event Tracking Platforms
Job Description
As the Director of Data Analytics, Science & AI Enablement at Chargebee, your role will be crucial in leading the development and growth of a data function that drives enterprise-wide AI initiatives. You will play a key part in designing, building, and managing a team responsible for data analytics, data science, data governance, and structured data enablement to support advanced AI/ML use cases. **Roles and Responsibilities:** - **Data Analytics, Science and AI Enablement Leadership** - Lead the development and deployment of machine learning, generative AI, recommendation systems, and predictive models to enhance product intelligence and automation. - Scale AI capabilities across the platform, including personalization, NLP, anomaly detection, and customer segmentation. - Ensure that models are interpretable, ethical, and in alignment with business and customer trust standards. - **Business Analytics & Insights** - Drive insights into user behavior, product performance, churn prediction, and lifecycle value utilizing customer and usage data. - Develop dashboards, KPIs, and self-service analytics tools for marketing, product, sales, and support teams. - Own the customer analytics roadmap to enhance onboarding, conversion, retention, and upsell opportunities. - **Team Building & Cross-Functional Leadership** - Establish and lead a high-performance team of data scientists, AI/ML engineers, analysts, and data product managers. - Collaborate with various departments to align data strategy with business objectives and legal requirements. - Communicate findings to senior leadership and influence roadmap decisions through data-backed recommendations. - **Data Infrastructure & Governance** - Work with Data Engineering to ensure scalable data architecture and high-quality data pipelines. - Oversee data quality, governance, and compliance across all analytical and operational systems. - Implement scalable data architecture and governance frameworks, ensuring compliance with data privacy regulations. - **Team Leadership & Vision** - Build and lead a high-performing global team of data analysts, data scientists, and data engineers. - Define and execute a comprehensive data and AI enablement roadmap aligned with company goals. - **Data & AI Strategy Enablement** - Drive data availability, quality, and governance across the organization to support AI and advanced analytics initiatives. - Partner with engineering, product, and business stakeholders to identify opportunities for AI/ML solutions. - **Analytics & Business Intelligence** - Lead the development of dashboards, metrics, and decision-support tools. - Foster a culture of data literacy and insight-driven decision-making throughout the organization. - Provide executive-level insights through advanced data analysis and reporting. **Required Qualifications:** - Bachelor's or Master's degree in Computer Science, Statistics, Data Science, Engineering, or related discipline. - Experience in a SaaS or tech environment with subscription-based metrics. - 10+ years of experience in data analytics, data science, or related fields, with 3-5 years in a leadership role. - Strong knowledge of AI/ML concepts, data platforms, and BI tools. - Deep understanding of data governance, data quality, and metadata management. - Demonstrated ability to lead change in a global, fast-paced environment. - Excellent communication and stakeholder management skills. - Knowledge of SaaS metrics, PLG, and usage-based pricing strategies. - Experience with product instrumentation and event tracking platforms. - Experience scaling data teams in a high-growth or startup environment.,
ACTIVELY HIRING
logo

@ 2025 Shine.com | All Right Reserved

Connect with us:
  • LinkedIn
  • Instagram
  • Facebook
  • YouTube
  • Twitter