How to Install and Uninstall python38-w3lib Package on openSuSE Tumbleweed
Last updated: November 07,2024
Deprecated! Installation of this package may no longer be supported.
1. Install "python38-w3lib" package
Please follow the guidance below to install python38-w3lib on openSuSE Tumbleweed
$
sudo zypper refresh
Copied
$
sudo zypper install
python38-w3lib
Copied
2. Uninstall "python38-w3lib" package
Please follow the instructions below to uninstall python38-w3lib on openSuSE Tumbleweed:
$
sudo zypper remove
python38-w3lib
Copied
3. Information about the python38-w3lib package on openSuSE Tumbleweed
Information for package python38-w3lib:
---------------------------------------
Repository : openSUSE-Tumbleweed-Oss
Name : python38-w3lib
Version : 1.22.0-2.4
Arch : noarch
Vendor : openSUSE
Installed Size : 139,4 KiB
Installed : No
Status : not installed
Source package : python-w3lib-1.22.0-2.4.src
Summary : Library of Web-Related Functions
Description :
This is a Python library of web-related functions, such as:
* remove comments, or tags from HTML snippets
* extract base url from HTML snippets
* translate entites on HTML strings
* encoding mulitpart/form-data
* convert raw HTTP headers to dicts and vice-versa
* construct HTTP auth header
* converting HTML pages to unicode
* RFC-compliant url joining
* sanitize urls (like browsers do)
* extract arguments from urls
---------------------------------------
Repository : openSUSE-Tumbleweed-Oss
Name : python38-w3lib
Version : 1.22.0-2.4
Arch : noarch
Vendor : openSUSE
Installed Size : 139,4 KiB
Installed : No
Status : not installed
Source package : python-w3lib-1.22.0-2.4.src
Summary : Library of Web-Related Functions
Description :
This is a Python library of web-related functions, such as:
* remove comments, or tags from HTML snippets
* extract base url from HTML snippets
* translate entites on HTML strings
* encoding mulitpart/form-data
* convert raw HTTP headers to dicts and vice-versa
* construct HTTP auth header
* converting HTML pages to unicode
* RFC-compliant url joining
* sanitize urls (like browsers do)
* extract arguments from urls