Stack operation of tensor networks
The tensor network, as a facterization of tensors, aims at performing the operations that are common for normal tensors, such as addition, contraction and stacking. However, due to its non-unique network structure, only the tensor network contraction is so far well defined. In this paper, we prop...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/160356 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-160356 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1603562023-02-28T20:08:18Z Stack operation of tensor networks Zhang, Tianning Chen, Tianqi Li, Erping Yang, Bo Ang, L. K. School of Physical and Mathematical Sciences Science::Physics Tensor Network Stack Operation The tensor network, as a facterization of tensors, aims at performing the operations that are common for normal tensors, such as addition, contraction and stacking. However, due to its non-unique network structure, only the tensor network contraction is so far well defined. In this paper, we propose a mathematically rigorous definition for the tensor network stack approach, that compress a large amount of tensor networks into a single one without changing their structures and configurations. We illustrate the main ideas with the matrix product states based machine learning as an example. Our results are compared with the for loop and the efficient coding method on both CPU and GPU. Published version This work was supported by US Office of Naval Research Global (N62909-19-1-2047) and SUTD-ZJU Visiting Professor (VP 201303). 2022-07-20T06:06:01Z 2022-07-20T06:06:01Z 2022 Journal Article Zhang, T., Chen, T., Li, E., Yang, B. & Ang, L. K. (2022). Stack operation of tensor networks. Frontiers in Physics, 10, 906399-. https://dx.doi.org/10.3389/fphy.2022.906399 2296-424X https://hdl.handle.net/10356/160356 10.3389/fphy.2022.906399 2-s2.0-85130970420 10 906399 en Frontiers in Physics © 2022 Zhang, Chen, Li, Yang and Ang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Science::Physics Tensor Network Stack Operation |
spellingShingle |
Science::Physics Tensor Network Stack Operation Zhang, Tianning Chen, Tianqi Li, Erping Yang, Bo Ang, L. K. Stack operation of tensor networks |
description |
The tensor network, as a facterization of tensors, aims at performing the
operations that are common for normal tensors, such as addition, contraction
and stacking. However, due to its non-unique network structure, only the tensor
network contraction is so far well defined. In this paper, we propose a
mathematically rigorous definition for the tensor network stack approach, that
compress a large amount of tensor networks into a single one without changing
their structures and configurations. We illustrate the main ideas with the
matrix product states based machine learning as an example. Our results are
compared with the for loop and the efficient coding method on both CPU and GPU. |
author2 |
School of Physical and Mathematical Sciences |
author_facet |
School of Physical and Mathematical Sciences Zhang, Tianning Chen, Tianqi Li, Erping Yang, Bo Ang, L. K. |
format |
Article |
author |
Zhang, Tianning Chen, Tianqi Li, Erping Yang, Bo Ang, L. K. |
author_sort |
Zhang, Tianning |
title |
Stack operation of tensor networks |
title_short |
Stack operation of tensor networks |
title_full |
Stack operation of tensor networks |
title_fullStr |
Stack operation of tensor networks |
title_full_unstemmed |
Stack operation of tensor networks |
title_sort |
stack operation of tensor networks |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/160356 |
_version_ |
1759855819597283328 |