EasyTPP: Towards Open Benchmarking Temporal Point Processes
EasyTPP
is an easy-to-use development and application toolkit for Temporal Point Process (TPP), with key features in configurability, compatibility and reproducibility. We hope this project could benefit both researchers and practitioners with the goal of easily customized development and open benchmarking in TPP.
| Features | Model List | Dataset | Quick Start | Benchmark |Documentation |Todo List | Citation |Acknowledgement | Star History |
- [09-02-2023] We published two non-anthropogenic datasets [earthquake](https://drive.google.com/drive/folders/1ubeIz_CCNjHyuu6-XXD0T-gdOLm12rf4) and [volcano eruption](https://drive.google.com/drive/folders/1KSWbNi8LUwC-dxz1T5sOnd9zwAot95Tp?usp=drive_link)! See Dataset for details. - [05-29-2023] We released ``EasyTPP`` v0.0.1! - [12-27-2022] Our paper [Bellman Meets Hawkes: Model-Based Reinforcement Learning via Temporal Point Processes](https://arxiv.org/abs/2201.12569) was accepted by AAAI'2023! - [10-01-2022] Our paper [HYPRO: A Hybridly Normalized Probabilistic Model for Long-Horizon Prediction of Event Sequences](https://arxiv.org/abs/2210.01753) was accepted by NeurIPS'2022! - [05-01-2022] We started to develop `EasyTPP`.
EasyTPP
implements two equivalent sets of models, which can
be run under Tensorflow (both Tensorflow 1.13.1 and Tensorflow 2.0) and PyTorch 1.7.0+ respectively. While the PyTorch models are more popular among researchers, the compatibility with Tensorflow is important for industrial practitioners.We provide reference implementations of various state-of-the-art TPP papers:
No | Publication | Model | Paper | Implementation |
---|---|---|---|---|
1 | KDD'16 | RMTPP | Recurrent Marked Temporal Point Processes: Embedding Event History to Vector | Tensorflow Torch |
2 | NeurIPS'17 | NHP | The Neural Hawkes Process: A Neurally Self-Modulating Multivariate Point Process | Tensorflow Torch |
3 | NeurIPS'19 | FullyNN | Fully Neural Network based Model for General Temporal Point Processes | Tensorflow Torch |
4 | ICML'20 | SAHP | Self-Attentive Hawkes process | Tensorflow Torch |
5 | ICML'20 | THP | Transformer Hawkes process | Tensorflow Torch |
6 | ICLR'20 | IntensityFree | Intensity-Free Learning of Temporal Point Processes | Tensorflow Torch |
7 | ICLR'21 | ODETPP | Neural Spatio-Temporal Point Processes (simplified) | Tensorflow Torch |
8 | ICLR'22 | AttNHP | Transformer Embeddings of Irregularly Spaced Events and Their Participants | Tensorflow Torch |
We preprocessed one synthetic and five real world datasets from widely-cited works that contain diverse characteristics in terms of their application domains and temporal statistics:
Per users' request, we processed two non-anthropogenic datasets
Earthquake: timestamped earthquake events over the Conterminous U.S from 1996 to 2023, processed from USGS.
Volcano eruption: timestamped volcano eruption events over the world in recent hundreds of years, processed from The Smithsonian Institution.
All datasets are preprocess to the Gatech
format dataset widely used for TPP researchers, and saved at Google Drive with a public access.
We provide an end-to-end example for users to run a standard TPP model with EasyTPP
.
First of all, we can install the package either by using pip or from the source code on Github.
To install the latest stable version:
pip install easy-tpp
To install the latest on GitHub:
git clone https://github.com/ant-research/EasyTemporalPointProcess.git
cd EasyTemporalPointProcess
python setup.py install
We need to put the datasets in a local directory before running a model and the datasets should follow a certain format. See OnlineDoc - Datasets for more details.
Suppose we use the taxi dataset in the example.
Before start training, we need to set up the config file for the pipeline. We provide a preset config file in Example Config. The details of the configuration can be found in OnlineDoc - Training Pipeline.
After the setup of data and config, the directory structure is as follows:
data
|______taxi
|____ train.pkl
|____ dev.pkl
|____ test.pkl
configs
|______experiment_config.yaml
Then we start the training by simply running the script
import argparse
from easy_tpp.config_factory import Config
from easy_tpp.runner import Runner
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--config_dir', type=str, required=False, default='configs/experiment_config.yaml',
help='Dir of configuration yaml to train and evaluate the model.')
parser.add_argument('--experiment_id', type=str, required=False, default='NHP_train',
help='Experiment id in the config file.')
args = parser.parse_args()
config = Config.build_from_yaml_file(args.config_dir, experiment_id=args.experiment_id)
model_runner = Runner.build_from_config(config)
model_runner.run()
if __name__ == '__main__':
main()
A more detailed example can be found at OnlineDoc - QuickStart.
The classes and methods of EasyTPP
have been well documented so that users can generate the documentation by:
cd doc
pip install -r requirements.txt
make html
NOTE:
doc/requirements.txt
is only for documentation by Sphinx, which can be automatically generated by Github actions .github/workflows/docs.yml
. (Trigger by pull request.)The full documentation is available on the website.
In the examples folder, we provide a script to benchmark the TPPs, with Taxi dataset as the input.
To run the script, one should download the Taxi data following the above instructions. The config file is readily setup up. Then run
cd examples
python benchmark_script.py
This project is licensed under the Apache License (Version 2.0). This toolkit also contains some code modified from other repos under other open-source licenses. See the NOTICE file for more information.
If you find EasyTPP
useful for your research or development, please cite the following paper:
@inproceedings{xue2024easytpp,
title={EasyTPP: Towards Open Benchmarking Temporal Point Processes},
author={Siqiao Xue and Xiaoming Shi and Zhixuan Chu and Yan Wang and Hongyan Hao and Fan Zhou and Caigao Jiang and Chen Pan and James Y. Zhang and Qingsong Wen and Jun Zhou and Hongyuan Mei},
booktitle = {International Conference on Learning Representations (ICLR)},
year = {2024},
url ={https://arxiv.org/abs/2307.08097}
}
The project is jointly initiated by Machine Intelligence Group, Alipay and DAMO Academy, Alibaba.
The following repositories are used in EasyTPP
, either in close to original form or as an inspiration: