Mitigating performance saturation in neural marked point processes : architectures and loss functions
Attributed event sequences are commonly encountered in practice. A recent research line focuses on incorporating neural networks with the statistical model -- marked point processes, which is the conventional tool for dealing with attributed event sequences. Neural marked point processes possess...
Saved in:
Main Authors: | , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/152799 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-152799 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1527992021-10-05T05:43:19Z Mitigating performance saturation in neural marked point processes : architectures and loss functions Li, Tianbo Luo, Tianze Ke, Yiping Pan, Sinno Jialin School of Computer Science and Engineering 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (KDD '21) Computational Intelligence Lab Engineering::Computer science and engineering Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Neural Point Processes Hawkes Processes Event Sequential Analysis Attributed event sequences are commonly encountered in practice. A recent research line focuses on incorporating neural networks with the statistical model -- marked point processes, which is the conventional tool for dealing with attributed event sequences. Neural marked point processes possess good interpretability of probabilistic models as well as the representational power of neural networks. However, we find that performance of neural marked point processes is not always increasing as the network architecture becomes more complicated and larger, which is what we call the performance saturation phenomenon. This is due to the fact that the generalization error of neural marked point processes is determined by both the network representational ability and the model specification at the same time. Therefore we can draw two major conclusions: first, simple network structures can perform no worse than complicated ones for some cases; second, using a proper probabilistic assumption is as equally, if not more, important as improving the complexity of the network. Based on this observation, we propose a simple graph-based network structure called GCHP, which utilizes only graph convolutional layers, thus it can be easily accelerated by the parallel mechanism. We directly consider the distribution of interarrival times instead of imposing a specific assumption on the conditional intensity function, and propose to use a likelihood ratio loss with a moment matching mechanism for optimization and model selection. Experimental results show that GCHP can significantly reduce training time and the likelihood ratio loss with interarrival time probability assumptions can greatly improve the model performance. Ministry of Education (MOE) Nanyang Technological University National Research Foundation (NRF) This research/project is supported by: (1) the National Research Foundation, Singapore under its Industry Alignment Fund – Pre-positioning (IAF-PP) Funding Initiative; (2) NTU Singapore Nanyang Assistant Professorship (NAP) grant M4081532.020; (3) Singapore MOE AcRF Tier-1 grant 2018-T1-002-143 (RG131/18 (S)); and (4) Alibaba-NTU Singapore Joint Research Institute, Nanyang Technological University. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not reflect the views of National Research Foundation, Singapore. 2021-10-05T05:42:38Z 2021-10-05T05:42:38Z 2021 Conference Paper Li, T., Luo, T., Ke, Y. & Pan, S. J. (2021). Mitigating performance saturation in neural marked point processes : architectures and loss functions. 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (KDD '21), 986-994. https://dx.doi.org/10.1145/3447548.3467436 9781450383325 https://hdl.handle.net/10356/152799 10.1145/3447548.3467436 2-s2.0-85114927856 986 994 en SDSC-2020-004 M4081532.020 2018-T1-002-143 (RG131/18 (S) © 2021 Copyright held by the owner/author(s). Publication rights licensed to ACM. All rgihts reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Neural Point Processes Hawkes Processes Event Sequential Analysis |
spellingShingle |
Engineering::Computer science and engineering Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Neural Point Processes Hawkes Processes Event Sequential Analysis Li, Tianbo Luo, Tianze Ke, Yiping Pan, Sinno Jialin Mitigating performance saturation in neural marked point processes : architectures and loss functions |
description |
Attributed event sequences are commonly encountered in practice. A recent
research line focuses on incorporating neural networks with the statistical
model -- marked point processes, which is the conventional tool for dealing
with attributed event sequences. Neural marked point processes possess good
interpretability of probabilistic models as well as the representational power
of neural networks. However, we find that performance of neural marked point
processes is not always increasing as the network architecture becomes more
complicated and larger, which is what we call the performance saturation
phenomenon. This is due to the fact that the generalization error of neural
marked point processes is determined by both the network representational
ability and the model specification at the same time. Therefore we can draw two
major conclusions: first, simple network structures can perform no worse than
complicated ones for some cases; second, using a proper probabilistic
assumption is as equally, if not more, important as improving the complexity of
the network. Based on this observation, we propose a simple graph-based network
structure called GCHP, which utilizes only graph convolutional layers, thus it
can be easily accelerated by the parallel mechanism. We directly consider the
distribution of interarrival times instead of imposing a specific assumption on
the conditional intensity function, and propose to use a likelihood ratio loss
with a moment matching mechanism for optimization and model selection.
Experimental results show that GCHP can significantly reduce training time and
the likelihood ratio loss with interarrival time probability assumptions can
greatly improve the model performance. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Li, Tianbo Luo, Tianze Ke, Yiping Pan, Sinno Jialin |
format |
Conference or Workshop Item |
author |
Li, Tianbo Luo, Tianze Ke, Yiping Pan, Sinno Jialin |
author_sort |
Li, Tianbo |
title |
Mitigating performance saturation in neural marked point processes : architectures and loss functions |
title_short |
Mitigating performance saturation in neural marked point processes : architectures and loss functions |
title_full |
Mitigating performance saturation in neural marked point processes : architectures and loss functions |
title_fullStr |
Mitigating performance saturation in neural marked point processes : architectures and loss functions |
title_full_unstemmed |
Mitigating performance saturation in neural marked point processes : architectures and loss functions |
title_sort |
mitigating performance saturation in neural marked point processes : architectures and loss functions |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/152799 |
_version_ |
1713213289785196544 |