Loghub-2.0 is an improved collection of large-scale annotated datasets for log parsing based on Loghub.
Based on Loghub-2.0, we propose a more comprehensive benchmark of log parsers. The detailed evaluation results could be found at RQ_experiments 🔗.
Software systems | # Annotated Logs (Loghub-2.0) | # Templates (Loghub-2.0) | # Templates (Loghub-2k) |
---|---|---|---|
Distributed systems | |||
Hadoop | 179,993 | 236 | 114 |
HDFS | 11,167,740 | 46 | 14 |
OpenStack | 207,632 | 48 | 43 |
Spark | 16,075,117 | 236 | 36 |
Zookeeper | 74,273 | 89 | 50 |
Supercomputer systems | |||
BGL | 4,631,261 | 320 | 120 |
HPC | 429,987 | 74 | 46 |
Thunderbird | 16,601,745 | 1,241 | 149 |
Operating systems | |||
Linux | 23,921 | 338 | 118 |
Mac | 100,314 | 626 | 341 |
Server application | |||
Apache | 51,977 | 29 | 6 |
OpenSSH | 638,946 | 38 | 27 |
Standalone software | |||
HealthApp | 212,394 | 156 | 75 |
Proxifier | 21,320 | 11 | 8 |
Average | 3,601,187 | 249.1 | 81.9 |
Please first download the full datasets of Loghub-2.0 via Zenodo.
Then, you need to put these datasets into full_dataset/
following the format of 2k_dataset
.
├── 2k_dataset/ # the original Loghub-2k datasets
├── full_dataset/ # unzip the Loghub-2.0 into this directory
│ └── post_process.py # we provide the heuristic roles used in our annotation of templates
├── benchmark/
│ ├── evaluation/
│ ├── logparser/
│ ├── old_benchmark/
│ ├── LogPPT/ # contains the modified source code of LogPPT
│ ├── UniParser/ # contains the source code of implemented UniParser
│ ├── run_statistic_2k.sh # the script to run all statistic-based log parsers on Loghub-2k datasets
│ └── run_statistic_full.sh # the script to run all statistic-based log parsers on Loghub-2.0 datasets
├── result/
│ ├── ...... #
│ └── ...... # contains the output evaluation metric files and all parsed results
├── RQ_experiments/ # contains the experimental results of RQs
│ ├── RQ1/
│ ├── RQ2/
│ └── RQ3/
├── requirements.txt
└── README.MD
Owing to the large scale of the benchmark in the experiments, the requirements of the benchmark of all log parsers are:
- At least 16GB memory.
- At least 100GB storage.
- GPU (for LogPPT and UniParser).
Installation
- Install
python >= 3.8
pip install -r requirements.txt
Running the entire benchmark using Loghub-2.0 datasets requires more than 48 hours to complete.
Note that if you would like to evaluate your parser, one can easily put their parsed results following the format as the files shown in result/
, and run our evluation code.
If you woud like to re-run all parsers using Loghub-2.0, please follow our large-scale benchmarking steps.
We give a demo script to run Drain on both Loghub-2k and Loghub-2.0, this will takes about 2-3 hours.
cd benchmark/
./demo.sh
One can follow the steps to evaluate all parsers using Loghub-2k or the proposed Loghub-2.0 datasets. The overall time cost is more than 48 hours.
- Run all statistic-based log parsers on Loghub-2k
cd benchmark/
./run_statistic_2k.sh
- Run all statistic-based log parsers on Loghub-2.0
cd benchmark/
./run_statistic_full.sh
-
Run Semantic-based log parsers: LogPPT & UniParser
Since these methods are quite different with other log parsers, and they requires a GPU to support efficient parsing, we seperate their environments from other log parsers. Please refer to the README file of LogPPT or UniParser to use one-click script to parse and evaluate each log parsers respectively.
If you use our labeled datasets for public research, please cite the following two papers:
- Loghub: Jieming Zhu, Shilin He, Pinjia He, Jinyang Liu, Michael R. Lyu. Loghub: A Large Collection of System Log Datasets for AI-driven Log Analytics. ISSRE, 2023.
- Loghub-2.0: Zhihan Jiang, Jinyang Liu, Junjie Huang, Yichen Li, Yintong Huo, Jiazhen Gu, Zhuangbin Chen, Jieming Zhu, Michael R. Lyu. A Large-scale Evaluation for Log Parsing Techniques: How Far are We? ISSTA, 2024.
In addition, if you use the souce code of our work for public research, please kindly cite the following paper for their efforts on logparser implementation:
- Jieming Zhu, Shilin He, Jinyang Liu, Pinjia He, Qi Xie, Zibin Zheng, Michael R. Lyu. Tools and Benchmarks for Automated Log Parsing. ICSE, 2019.