How to Install and Uninstall python310-opt-einsum Package on openSuSE Tumbleweed
Last updated: November 26,2024
1. Install "python310-opt-einsum" package
This tutorial shows how to install python310-opt-einsum on openSuSE Tumbleweed
$
sudo zypper refresh
Copied
$
sudo zypper install
python310-opt-einsum
Copied
2. Uninstall "python310-opt-einsum" package
This is a short guide on how to uninstall python310-opt-einsum on openSuSE Tumbleweed:
$
sudo zypper remove
python310-opt-einsum
Copied
3. Information about the python310-opt-einsum package on openSuSE Tumbleweed
Information for package python310-opt-einsum:
---------------------------------------------
Repository : openSUSE-Tumbleweed-Oss
Name : python310-opt-einsum
Version : 3.3.0-3.1
Arch : noarch
Vendor : openSUSE
Installed Size : 463.2 KiB
Installed : No
Status : not installed
Source package : python-opt-einsum-3.3.0-3.1.src
Upstream URL : https://github.com/dgasmith/opt_einsum
Summary : Optimizing numpys einsum function
Description :
Optimized einsum can significantly reduce the overall execution time of einsum-like expressions (e.g.,
`np.einsum`,`dask.array.einsum`,`pytorch.einsum`,`tensorflow.einsum`)
by optimizing the expression's contraction order and dispatching many
operations to canonical BLAS, cuBLAS, or other specialized routines. Optimized
einsum is agnostic to the backend and can handle NumPy, Dask, PyTorch,
Tensorflow, CuPy, Sparse, Theano, JAX, and Autograd arrays as well as potentially
any library which conforms to a standard API. See the
[**documentation**](http://optimized-einsum.readthedocs.io) for more
information.
---------------------------------------------
Repository : openSUSE-Tumbleweed-Oss
Name : python310-opt-einsum
Version : 3.3.0-3.1
Arch : noarch
Vendor : openSUSE
Installed Size : 463.2 KiB
Installed : No
Status : not installed
Source package : python-opt-einsum-3.3.0-3.1.src
Upstream URL : https://github.com/dgasmith/opt_einsum
Summary : Optimizing numpys einsum function
Description :
Optimized einsum can significantly reduce the overall execution time of einsum-like expressions (e.g.,
`np.einsum`,`dask.array.einsum`,`pytorch.einsum`,`tensorflow.einsum`)
by optimizing the expression's contraction order and dispatching many
operations to canonical BLAS, cuBLAS, or other specialized routines. Optimized
einsum is agnostic to the backend and can handle NumPy, Dask, PyTorch,
Tensorflow, CuPy, Sparse, Theano, JAX, and Autograd arrays as well as potentially
any library which conforms to a standard API. See the
[**documentation**](http://optimized-einsum.readthedocs.io) for more
information.