Skip to content

tnbar/awesome-tensorial-neural-networks

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 

Repository files navigation

Awesome Tensorial Neural Networks Awesome

An survey of tensorial neural networks (TNNs) in

  • Network compression via TNNs
  • Information fusion via TNNs
  • Quantum Circuit Simulation via TNNs
  • Training Strategy of TNNs
  • Toolboxes of TNNs

This repository is consistent with our survey paper Tensor Networks Meet Neural Networks: A Survey and Future Perspectives. Please see https://arxiv.org/abs/2302.09019 for more details. And if you find this work helpful, we would appreciate it if you could cite this collection in the following form:

@article{DBLP:journals/corr/abs-2302-09019,
  author       = {Maolin Wang and
                  Yu Pan and
                  Zenglin Xu and
                  Xiangli Yang and
                  Guangxi Li and
                  Andrzej Cichocki},
  title        = {Tensor Networks Meet Neural Networks: {A} Survey and Future Perspectives},
  journal      = {CoRR},
  volume       = {abs/2302.09019},
  year         = {2023}
}

Network compression via TNNs

Tensorial Convolutional Neural Networks

Paper Remarks Conference/Journal Year
Pan et al. "A Unified Weight Initialization Paradigm for Tensorial Convolutional Neural Networks". [link] Proposing a universal weight initialization paradigm, which generalizes Xavier and Kaiming methods and can be widely applicable to arbitrary TCNNs. ICML 2022
Ye Liu and Michael K. Ng. "Deep neural network compression by Tucker decomposition with nonlinear response". [link] Compressing deep neural network with low multilinear rank Tucker format. Knowledge-Based Systems 2022
Liu et al. "TT-TSVD: A Multi-modal Tensor Train Decomposition with Its Application in Convolutional Neural Networks for Smart Healthcare".[link] A tensor train-tensor singular value decomposition (TT-TSVD) algorithm for data reduction and compression of the convolutional neural networks. TOMM 2022
Ye et al. "Block-term tensor neural networks". [link] Exploring the correlations in the weight matrices, and approximating the weight matrices with the low-rank Block-Term Tucker tensors. Neural Networks 2020
Kossaifi et al. "Tensor regression networks". [link] Introducing Tensor Contraction Layers (TCLs) that reduce the dimensionality. JMLR 2020
Wu et al. "Hybrid tensor decomposition in neural network compression". [link] Introducing the hierarchical Tucker (HT) to investigate its capability in neural network compression. Neural Networks 2020
Kossaifi et al. "Factorized higher-order cnns with an application to spatio-temporal emotion estimation". [link] Proposing coined CP-HigherOrder Convolution (HO-CPConv), to spatio-temporal facial emotion analysis. CVPR 2020
Phan et al. "Stable low-rank tensor decomposition for compression of convolutional neural network".[link] A stable decomposition method CPD-EPC is proposed with a minimal sensitivity design for both CP convolutional layers and hybrid Tucker2-CP convolutional layers. ECCV 2020
Wang et al. "Concatenated tensor networks for deep multi-task learning". [link] Introducing a novel Concatenated Tensor Network structure, in particular, Projected Entangled Pair States (PEPS) like structure, into multi-task deep models. ICONIP 2020
Kossaifi et al. "T-net: Parametrizing fully convolutional nets with a single high-order tensor". [link] Proposing to fully parametrize Convolutional Neural Networks (CNNs) with a single highorder, low-rank tucker tensor format. CVPR 2019
Hayashi et al. "Einconv: Exploring Unexplored Tensor Network Decompositions for Convolutional Neural Networks". [link] Characterizing a decomposition class specific to CNNs by adopting a flexible graphical notation. NeurIPS 2019
Wang et al. "Wide compression: Tensor ring nets". [link] Significantly compressing both the fully connected layers and the convolutional layers of deep networks via Introducing Tensor Ring format. CVPR 2018
Yang et al. "Deep multi-task representation learning: A tensor factorisation approach". [link] Proposing deep multi-task Tucker models and Tensor Train modesl that learn cross-task sharing structure. ICLR 2017
Garipov et al. "Ultimate tensorization: compressing convolutional and fc layers alike". [link] Compressing convolutional layers via Tensor Train format. Arxiv preprint 2016
Novikov et al. "Tensorizing neural networks". [link] Converting the dense weight matrices of the fully-connected layers in CNNs to the Tensor Train format. NeurIPS 2015
Lebedev et al. "Speeding-up convolutional neural networks using fine-tuned CP-decomposition". [link] Decomposing the 4D convolution kernel tensor via CP-decomposition. ICLR 2015
Denton et al. "Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation". [link] Speeding up the test-time evaluation of large convolutional networks via CP-decomposition. NeurIPS 2014

Tensorial Recurrent Neural Networks

Paper Remarks Conference/Journal Year
Yin et al. "Towards extremely compact rnns for video recognition with fully decomposed hierarchical tucker structure". [link] Proposing to develop extremely compact RNN models with fully decomposed hierarchical Tucker structure. CVPR 2021
Wang et al. "Kronecker CP decomposition with fast multiplication for compressing RNNs". [link] Compressing RNNs based on a novel Kronecker CANDECOMP/PARAFAC decomposition, which is derived from Kronecker tensor decomposition. TNNLS 2021
Kossaifi et al. "Tensor regression networks". [link] Introducing Tensor Contraction Layers (TCLs) that reduce the dimensionality. JMLR 2020
Ye et al. "Block-term tensor neural networks". [link] Exploring the correlations in the weight matrices, and approximating the weight matrices with the low-rank Block-Term Tucker tensors. Neural Networks 2020
Su et al. "Convolutional tensor-train LSTM for spatio-temporal learning". [link] Proposing a novel tensor-train module that performs prediction by combining convolutional features across time. NeurIPS 2020
Wu et al. "Hybrid tensor decomposition in neural network compression". [link] Introducing the hierarchical Tucker (HT) to investigate its capability in neural network compression. Neural Networks 2020
Tjandra et al. "Recurrent Neural Network Compression Based on Low-Rank Tensor Representation". [link] Proposing to use Tensor Train formats to re-parameterize the Gated Recurrent Unit (GRU) RNN. IEICE Transactions on Information and Systems 2019
Pan et al. "Compressing recurrent neural networks with tensor ring for action recognition". [link] Proposing a novel compact LSTM model, named as TR-LSTM, by utilizing the low-rank tensor ring decomposition (TRD) to reformulate the input-to-hidden transformation. AAAI 2019
Jose et al. "Kronecker recurrent units". [link] Achieving a parameter efficiency in RNNs through a Kronecker factored recurrent matrix. ICML 2018
Ye et al. "Learning compact recurrent neural networks with block-term tensor decomposition". [link] Proposing to apply Block-Term tensor decomposition to reduce the parameters of RNNs and improves their training efficiency. CVPR 2018
Yang et al. "Tensor-train recurrent neural networks for video classification". [link] Factorizing the input-to-hidden weight matrix in RNNs using Tensor-Train decomposition. ICML 2017
Kossaifi et al. "Tensor Contraction Layers for Parsimonious Deep Nets". [link] Proposing the Tensor Contraction Layer (TCL), the first attempt to incorporate tensor contractions as end-to-end trainable neural network layers. CVPR-Workshop 2017

Tensorial Transformer

Paper Remarks Conference/Journal Year
Pan et al."Reusing Pretrained Models by Multi-linear Operators for Efficient Training". [link] Utilizing tensor ring matrix product operator (TR-MPO) to grow a small pretrained model to a large counterpart for efficient training. NeurIPS 2023
Vasilescu et al."Causal Deep Learning: Causal Capsules and Tensor Transformers". [link] Forward causal questions are addressed with a neural network architecture composed of causal capsules and a tucker format tensor transformer. Arxiv preprint 2023
Liu et al. "Tuformer: Data-driven Design of Transformers for Improved Generalization or Efficiency". [link] Proposing a novel design by allowing data-driven weights across heads via low rank tensor diagrams. ICLR 2022
Ren et al. "Exploring extreme parameter compression for pre-trained language models". [link] Proposing to use Tucker formats to improve the effectiveness and efficiency during compression of Transformers. ICLR 2022
Li et al. "Hypoformer: Hybrid decomposition transformer for edge-friendly neural machine translation". [link] Compressing and accelerating Transformer via a Hybrid TensorTrain (HTT) decomposition. EMNLP 2022
Liu et al. "Enabling lightweight fine-tuning for pre-trained language model compression based on matrix product operators". [link] Proposing a novel fine-tuning strategy by only updating the parameters from the auxiliary tensors, and design an optimization algorithm for MPO-based approximation over stacked network architectures. ACL/IJCNLP 2021
Ma et al. "A tensorized transformer for language modeling". [link] Proposing a novel self-attention model (namely Multi-linear attention) with Block-Term Tensor Decomposition. NeurIPS 2019

Tensorial Graph Neural Networks

Paper Remarks Conference/Journal Year
Hua et al. "High-Order Pooling for Graph Neural Networks with Tensor Decomposition". [link] Proposing the highly expressive Tensorized Graph Neural Network (tGNN) to model high-order non-linear node interactions. NeurIPS 2022
"Multi-view tensor graph neural networks through reinforced aggregation".[link] A Tucker format structure is applied to extract the graph structure features in the common feature space, was introduced to capture the potential high order correlation information in multi-view graph learning tasks TKDE 2022
Baghershahi et al. "Efficient Relation-aware Neighborhood Aggregation in Graph Neural Networks via Tensor Decomposition". [link] Introducing a general knowledge graph encoder incorporating tensor decomposition in the aggregation function. Arxiv preprint 2022
Jia et al. "Dynamic spatiotemporal graph neural network with tensor network". [link] Exploring the entangled correlations in spatial tensor graph and temporal tensor graph by Projected Entangled Pair States (PEPS). Arxiv preprint 2020

Tensorial Restricted Boltzmann Machine

Paper Remarks Conference/Journal Year
Ju et al. "Tensorizing Restricted Boltzmann Machine". [link] Proposing TT-RBM which both visible and hidden variables are in tensorial form and are connected by a parameter matrix in tensor train formats. TKDD 2019
Chen et al. "Matrix Product Operator Restricted Boltzmann Machines". [link] Proposing the matrix product operator RBM that utilizes a tensor network generalization of Mv/TvRBM. IJCNN 2019
Wang et al. "Tensor ring restricted Boltzmann machines". [link] Proposing a tensor-input RBM model, which employs the tensor-ring (TR) decomposition structure to naturally represent the high-order relationship. IJCNN 2019
Qi et al. "Matrix variate restricted Boltzmann machine". [link] Proposing a bilinear connection between matrix variate visible layer and matrix variate hidden layer. IJCNN 2016
Nguyen et al. "Tensor-variate restricted Boltzmann machines". [link] Generalizing RBMs to capture the multiplicative interaction between data modes and the latent variables via CP decomposition. AAAI 2015

Information Fusion via TNNs

Tensor Fusion Layer-Based Methods

Paper Remarks Conference/Journal Year
Hou et al. "Deep multimodal multilinear fusion with high-order polynomial pooling". [link] Proposing a polynomial tensor pooling (PTP) block for integrating multimodal features by considering high-order moments. NeurIPS 2019
Liu et al. "Efficient low-rank multimodal fusion with modality-specific factors". [link] Proposing the low-rank method, which performs multimodal fusion using low-rank tensors to improve efficiency. ACL 2018
Zadeh et al. "Tensor fusion network for multimodal sentiment analysis". [link] Introducing a novel model, termed Tensor Fusion Network, which learns both intra-modality and inter-modality dynamics. EMNLP 2017

Multimodal Pooling-Based Methods

Paper Remarks Conference/Journal Year
Do et al. "Compact trilinear interaction for visual question answering". [link] Introducing a multimodal tensor-based PARALIND decomposition which efficiently parameterizes trilinear teraction between inputs. CVPR 2019
Fukui et al. "Multimodal compact bilinear pooling for visual question answering and visual grounding". [link] Proposing utilizing Multimodal Compact Bilinear pooling (MCB) to efficiently and expressively combine multimodal features. EMNLP 2016
Kim et al. "Hadamard product for low-rank bilinear pooling". [link] Proposing low-rank bilinear pooling using Hadamard product for an efficient attention mechanism of multimodal learning. Arxiv preprint 2016
Ben-Younes et al. "Mutan: Multimodal tucker fusion for visual question answering". [link] Proposing a multimodal tensor-based Tucker decomposition to efficiently parametrize bilinear interactions between visual and textual representations. CVPR 2017

Quantum Circuit Simulation on TNNs

Classical Data's Quantum State Embedding

Paper Remarks Conference/Journal Year
Miller et al. "Tensor Networks for Probabilistic Sequence Modeling". [link] Introducing a novel generative algorithm giving trained u-MPS the ability to efficiently sample from a wide variety of conditional distributions, each one defined by a regular expression. AISTATS 2021
Li et al. "CNM: An interpretable complex-valued network for matching". [link] Unifing different linguistic units in a single complex-valued vector space. NAACL 2019
Zhang et al. "A quantum many-body wave function inspired language modeling approach". [link] Considering word embeddings as a kind of global dependency information and integrated the quantum-inspired idea in a neural network architecture. CIKM 2018
Stoudenmire et al. "Supervised learning with tensor networks". [link] Introducing a framework for applying quantum-inspired tensor networks to image classification. NeurIPS 2016

Quantum Embedded Data Processing

Paper Remarks Conference/Journal Year
Liu et al. "Tensor networks for unsupervised machine learning".[link] A tensor network model combined matrix product states from quantum many-body physics and autoregressive modeling from machine learning. Physical Review E 2023
Miller et al. "Tensor Networks for Probabilistic Sequence Modeling". [link] Introducing a novel generative algorithm giving trained u-MPS the ability to efficiently sample from a wide variety of conditional distributions, each one defined by a regular expression. AISTATS 2021
Glasser et al. "Expressive power of tensor-network factorizations for probabilistic modeling". [link] Introducing locally purified states (LPS), a new factorization inspired by techniques for the simulation of quantum systems, with provably better expressive power than all other representations considered. NeurIPS 2019
Cheng et al. "Tree tensor networks for generative modeling". [link] Designing the tree tensor network to utilize the 2-dimensional prior of the natural images and develop sweeping learning and sampling algorithms. Physical Review B 2019
Han et al. "Unsupervised generative modeling using matrix product states". [link] Proposing a generative model using matrix product states, which is a tensor network originally proposed for describing (particularly one-dimensional) entangled quantum states. Physical Review X 2018
Edwin Stoudenmire and David J. Schwab. "Supervised learning with tensor networks". [link] Introducing a framework for applying quantum-inspired tensor networks to image classification. NeurIPS 2016

Convolutional Arithmetic Circuits

Paper Remarks Conference/Journal Year
Zhang et al. "A Generalized Language Model in Tensor Space". [link] Proposing a language model named Tensor Space Language Model (TSLM), by utilizing tensor networks and tensor decomposition. AAAI 2019
Levine et al. "Quantum Entanglement in Deep Learning Architectures". [link] Identifying an inherent re-use of information in the network operation as a key trait which distinguishes them from standard Tensor Network based representations. PRL 2019
Zhang et al. "A quantum many-body wave function inspired language modeling approach". [link] Proposing a Quantum Many-body Wave Function (QMWF) inspired language modeling approach. CIKM 2018
Levine et al. "Deep Learning and Quantum Entanglement: Fundamental Connections with Implications to Network Design". [link] Showing an equivalence between the function realized by a deep convolutional arithmetic circuit (ConvAC) and a quantum many-body wave function. ICLR 2018
Cohen et al. "On the Expressive Power of Deep Learning: A Tensor Analysis". [link] Showing that a shallow network corresponds to CP (rank-1) decomposition, whereas a deep network corresponds to Hierarchical Tucker decomposition. COLT 2016
Nadav Cohen and Amnon Shashua. "Convolutional Rectifier Networks as Generalized Tensor Decompositions". [link] Describing a construction based on generalized tensor decompositions, that transforms convolutional arithmetic circuits into convolutional rectifier networks. ICML 2016

Training Strategy

Stable Training

Paper Remarks Conference/Journal Year
Pan et al. "A Unified Weight Initialization Paradigm for Tensorial Convolutional Neural Networks". [link] Proposing a universal weight initialization paradigm, which generalizes Xavier and Kaiming methods and can be widely applicable to arbitrary TCNNs. ICML 2022
Panagakis et al. "Tensor methods in computer vision and deep learning". [link] Proposing a mixed-precision strategy to trade off time cost and numerical stability. Proceedings of IEEE 2021

Rank Selection

Paper Remarks Conference/Journal Year
Sobolev et al. "PARS: Proxy-Based Automatic Rank Selection for Neural Network Compression via Low-Rank Weight Approximation".[link] A proxy-based Bayesian optimization approach to find the best combination of ranks for neural network (NN) compression. Mathematics 2022
Sedighin et al. "Adaptive Rank Selection for Tensor Ring Decomposition".[link] An adaptive rank search framework for TR format in which TR ranks gradually increase in each iteration rather than being predetermined in advance. IEEE Journal of Selected Topics in Signal Processing 2021
Li et al. "Heuristic rank selection with progressively searching tensor ring network". [link] Proposing a novel progressive genetic algorithm named progressively searching tensor ring network search (PSTRN), which has the ability to find optimal rank precisely and efficiently. Complex & Intelligent Systems 2021
Cole Hawkins and Zheng Zhang. "Bayesian tensorized neural networks with automatic rank selection". [link] Proposing approaches for posterior density calculation and maximum a posteriori (MAP) estimation for the end-to-end training of our tensorized neural network. Neurocomputing 2021
Yin et al. "Towards efficient tensor decomposition-based dnn model compression with optimization framework". [link] Proposing a systematic framework for tensor decomposition-based model compression using Alternating Direction Method of Multipliers(ADMM). CVPR 2021
Cheng et al. "A novel rank selection scheme in tensor ring decomposition based on reinforcement learning for deep neural networks". [link] Proposing a novel rank selection scheme, which is inspired by reinforcement learning, to automatically select ranks in recently studied tensor ring decomposition in each convolutional layer. ICASSP 2020
Kim et al. "Compression of deep convolutional neural networks for fast and low power mobile applications". [link] Deriving an approximate rank by employing the Bayesian matrix factorization (BMF) to an unfolding weight tensor. ICLR 2016
Zhao et al. "Bayesian CP factorization of incomplete tensors with automatic rank determination". [link] Formulating CP factorization using a hierarchical probabilistic model and employ a fully Bayesian treatment. TPAMI 2015

Hardware Training

Paper Remarks Conference/Journal Year
Kao et al. "Hardware Acceleration in Large-Scale Tensor Decomposition for Neural Network Compression". [link] Proposing an energy-efficient hardware accelerator that implements randomized CPD in large-scale tensors for neural network compression. MWSCAS 2022
Qu et al. "Hardware-Enabled Efficient Data Processing with Tensor-Train Decomposition". [link] Proposing an algorithm-hardware co-design with customized architecture, namely, TTD Engine to accelerate TTD. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 2021
Deng et al. "TIE: Energy-efficient tensor train-based inference engine for deep neural network". [link] Developing a computation-efficient inference scheme for TT-format DNN. ISCA 2019
Huang et al. "LTNN: An energy-efficient machine learning accelerator on 3D CMOS-RRAM for layer-wise tensorized neural network". [link] Mapping TNNs to a 3D CMOS-RRAM based accelerator with significant bandwidth boosting from vertical I/O connections. SOCC 2017

Toolboxes

Basic Tensor Operation

Name Remarks Backends
Tensorly TensorLy is open-source, actively maintained and easily extensible. TensorLy provides all the utilities to easily use tensor methods from core tensor operations and tensor algebra to tensor decomposition and regression. Python (NumPy, PyTorch, TensorFlow, JAX, Apache MXNet and CuPy)
TensorNetwork TensorNetwork is an open source library for implementing tensor network algorithms. Python (TensorFlow, JAX, PyTorch, and Numpy)
Tensortools TensorTools is a bare bones Python package for fitting and visualizing canonical polyadic (CP) tensor decompositions of higher-order data arrays. Python (NumPy)
TnTorch TnTorch is a PyTorch-powered library for tensor modeling and learning that features transparent support for the the tensor train (TT) model, CANDECOMP/PARAFAC (CP), the Tucker model, and more. Python (Pytorch)
TorchMPS TorchMPS is a framework for working with matrix product state (also known as MPS or tensor train) models within Pytorch. Python (Pytorch)
T3F T3F supports GPU execution, batch processing, automatic differentiation, and versatile functionality for the Riemannian optimization framework. Python (Tensorflow)
TensorD TensorD provides basic decomposition methods, such as Tucker decomposition and CANDECOMP/PARAFAC (CP) decomposition, as well as new decomposition methods developed recently, for example, Pairwise Interaction Tensor Decomposition. Python (Tensorflow)
ITensor ITensor is a system for programming tensor network calculations with an interface modeled on tensor diagram notation, which allows users to focus on the connectivity of a tensor network without manually bookkeeping tensor indices. C++/Julia
TenDeC++ TenDeC++ implements four popular tensor decomposition methods, CANDECOMP/PARAFAC (CP) decomposition, Tucker decomposition, t-SVD, and Tensor-Train (TT) decomposition. C++
TensorToolbox Tensor Toolbox provides a suite of tools for working with multidimensional or N-way arrays. Matlab
TT-Toolbox he TT-Toolbox is a MATLAB implementation of basic operations with tensors in TT-format. Matlab
OSTD Online Stochastic Tensor Decomposition for Background Subtraction in Multispectral Video Sequences. Matlab
Scikit-TT Scikit-TT provides a powerful TT class as well as different modules comprising solvers for algebraic problems, the automatic construction of tensor trains, and data-driven methods. Python

Deep Model Implementation

Name Remarks Backends
Tensorly-Torch TensorLy-Torch is a PyTorch only library that builds on top of TensorLy and provides out-of-the-box tensor layers. It comes with all batteries included and tries to make it as easy as possible to use tensor methods within your deep networks. Python (Pytorch)
TedNet TedNet implements 5 kinds of tensor decomposition (i.e., CANDECOMP/PARAFAC (CP), Block-Term Tucker (BTT), Tucker-2, Tensor Train (TT) and Tensor Ring (TR) on traditional deep neural layers. Python (Pytorch)

Quantum Tensor Simulation

Name Remarks Backends
TensorToolbox Tensor Toolbox provides a suite of tools for working with multidimensional or N-way arrays. Matlab
ITensor ITensor is a system for programming tensor network calculations with an interface modeled on tensor diagram notation, which allows users to focus on the connectivity of a tensor network without manually bookkeeping tensor indices. C++/Julia
Yao Yao is an extensible, efficient open-source framework for quantum algorithm design. Python
lambeq Lambeq is a toolkit for quantum natural language processing. Python
TeD-Q TeD-Q provides an additional layer of annotations to the existing dataset. Python