Skip to content

Conversation

@e06084
Copy link
Collaborator

@e06084 e06084 commented Dec 18, 2025

No description provided.

@gemini-code-assist
Copy link

Summary of Changes

Hello @e06084, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request significantly enhances the project's documentation by revamping the README.md and its localized versions. The updates provide a more comprehensive and structured overview of Dingo, emphasizing its role as an AI data, model, and application quality evaluation tool. Key improvements include detailed explanations of its production-ready features, an expanded list of metrics, and clearer guidance on usage and extensibility, ensuring that users can quickly grasp Dingo's capabilities and how to leverage them.

Highlights

  • Comprehensive Documentation Overhaul: The main README.md file and its Japanese and Simplified Chinese translations have been extensively updated to provide a more detailed and structured overview of Dingo.
  • Enhanced Project Introduction: The introduction now clearly positions Dingo as a "Comprehensive AI Data, Model and Application Quality Evaluation Tool" and includes a new "Why Dingo?" section outlining its core benefits and capabilities.
  • Detailed Production-Readiness Concepts: A new section, "Key Concepts for Practitioners," explains Dingo's suitability for production environments, covering multi-field evaluation, stream processing, field isolation, hybrid rule-LLM strategies, and extensibility.
  • Expanded Data Quality Metrics: The metrics section has been revamped to list over 70 evaluation metrics, categorized for clarity across various dimensions like Pretrain Text Quality, RAG Evaluation, and Security.
  • New Feature Highlights Sections: Dedicated sections now elaborate on Multi-Source Data Integration, RAG System Evaluation, Hybrid Evaluation System, Flexible Execution & Integration, and Rich Reporting & Visualization.
  • Refined User Guide and Extensibility: The user guide is updated, and the extensibility section provides clearer examples for custom rule and LLM/prompt registration.
  • Updated Roadmap: The "Future Plans" section has been revised into a "Roadmap & Contributions" section, outlining new features such as Agent-as-a-Judge and a SaaS Platform.
  • New LLMTextQualityV5 Metric: A new LLMTextQualityV5 metric has been introduced and documented, alongside minor terminology adjustments in its source file.
Ignored Files
  • Ignored by pattern: .github/workflows/** (1)
    • .github/workflows/metrics-validation.yml
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

Copy link

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request significantly overhauls the README files, making them more structured, informative, and user-friendly. The changes introduce key concepts, detailed feature highlights, and clearer usage examples, which will greatly benefit the project's documentation. I've identified a couple of high-severity issues in the new code examples within the English README that would cause them to fail. Additionally, there are some minor issues with documentation links that could be improved for non-Chinese speakers. Overall, these are excellent improvements.

Comment on lines +262 to +268
@Model.rule_register('QUALITY_BAD_CUSTOM', ['default'])
class MyCustomRule(BaseRule):
@classmethod
def eval(cls, input_data: Data) -> EvalDetail:
# Your logic here
return EvalDetail(status=False, label=['QUALITY_GOOD'])
```

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

The code example for a custom rule is not functional and can be misleading.

  1. It will raise a pydantic.ValidationError because the EvalDetail constructor is missing the required metric argument.
  2. The example always returns a "good" result (status=False), which doesn't effectively demonstrate how a rule identifies issues.

A more illustrative and correct example would be helpful for users. Here is a suggestion that fixes the crash and provides a more meaningful implementation:

Suggested change
@Model.rule_register('QUALITY_BAD_CUSTOM', ['default'])
class MyCustomRule(BaseRule):
@classmethod
def eval(cls, input_data: Data) -> EvalDetail:
# Your logic here
return EvalDetail(status=False, label=['QUALITY_GOOD'])
```
@Model.rule_register('QUALITY_BAD_CUSTOM', ['default'])
class MyCustomRule(BaseRule):
@classmethod
def eval(cls, input_data: Data) -> EvalDetail:
# Example: check if content is empty
if not input_data.content:
return EvalDetail(
metric=cls.__name__,
status=True, # Found an issue
label=[f'{cls.metric_type}.{cls.__name__}'],
reason=["Content is empty."]
)
return EvalDetail(metric=cls.__name__, status=False, label=['QUALITY_GOOD'])

Comment on lines +453 to +457
return EvalDetail(
status=not is_valid, # False = good, True = bad
label=['QUALITY_GOOD' if is_valid else 'QUALITY_BAD_CUSTOM'],
reason=["Validation details..."]
)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

The code example for DomainSpecificRule is not correct as it will raise a pydantic.ValidationError. The EvalDetail constructor is missing the required metric argument.

Suggested change
return EvalDetail(
status=not is_valid, # False = good, True = bad
label=['QUALITY_GOOD' if is_valid else 'QUALITY_BAD_CUSTOM'],
reason=["Validation details..."]
)
return EvalDetail(
metric=cls.__name__,
status=not is_valid, # False = good, True = bad
label=['QUALITY_GOOD' if is_valid else 'QUALITY_BAD_CUSTOM'],
reason=["Validation details..."]
)

| **Security** | PII detection, Perspective API toxicity | Privacy and safety |

📊 **[View Complete Metrics Documentation →](docs/metrics.md)**
📖 **[RAG Evaluation Guide →](docs/rag_evaluation_metrics_zh.md)**

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The "RAG Evaluation Guide" link points to a Chinese document (docs/rag_evaluation_metrics_zh.md). For better accessibility for English-speaking users, it would be ideal to link to an English version of this guide. If an English version doesn't exist yet, consider creating one. This also applies to the link on line 359.

| **セキュリティ** | PII 検出、Perspective API 毒性 | プライバシーと安全性 |

📊 **[完全なメトリクス文書を表示 →](docs/metrics.md)**
📖 **[RAG 評価ガイド →](docs/rag_evaluation_metrics_zh.md)**

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

「RAG 評価ガイド」のリンクが中国語のドキュメント(docs/rag_evaluation_metrics_zh.md)を指しています。日本のユーザー向けには、日本語版または英語版のガイドにリンクするのが望ましいです。

@e06084 e06084 merged commit 9e5f3a5 into MigoXLab:dev Dec 18, 2025
2 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant