Index | index by Group | index by Distribution | index by Vendor | index by creation date | index by Name | Mirrors | Help | Search |
Name: libglusterfs0 | Distribution: SUSE Linux Framework One |
Version: 11.1 | Vendor: SUSE LLC <https://www.suse.com/> |
Release: slfo.1.1.1 | Build date: Thu Aug 22 20:11:49 2024 |
Group: System/Libraries | Build host: reproducible |
Size: 931597 | Source RPM: glusterfs-11.1-slfo.1.1.1.src.rpm |
Packager: https://www.suse.com/ | |
Url: https://www.gluster.org/ | |
Summary: GlusterFS's core library |
GlusterFS is a clustered file-system capable of scaling to several petabytes.
GPL-2.0-only OR LGPL-3.0-or-later
* Thu Aug 22 2024 pgajdos@suse.com - remove dependency on /usr/bin/python3 using %python3_fix_shebang{,_path} macros, [bsc#1212476] * Fri Dec 15 2023 jengelh@inai.de - Update to release 11.1 * Fix upgrade issue by reverting posix change related to storage.reserve value * Fix possible data loss during rebalance if there is any linkfile on the system - Delete 0001-socket-only-issue-IPPROTO_IPV6-setsockopt-calls-on-A.patch (merged), * Wed Jul 05 2023 jengelh@inai.de - Add 0001-glusterfs-fix-heap-buffer-overflow-on-startup.patch * Sat Jun 10 2023 jengelh@inai.de - Disable IO_uring for now [boo#1210894] * Tue May 23 2023 jengelh@inai.de - Add 0001-socket-only-issue-IPPROTO_IPV6-setsockopt-calls-on-A.patch [boo#1210894] * Tue Feb 21 2023 jengelh@inai.de - Update to release 11 [boo#1208517] [boo#1208519] * Major performance impovement of ~36% with rmdir operations * Extension of ZFS support for snapshots * Qouta implimentation based on namespace * Major cleanups and readdir/readdirp improvements * Thu Jul 14 2022 kastl@b1-systems.de - add explicit Requires for the dependencies, so we make sure the dependencies are installed in the same version * Fri May 20 2022 jengelh@inai.de - Update to release 10.2 * Some 165 bugfixes with none particularly sticking out * Tue Jan 18 2022 dsterba@suse.cz - add python-rpm-macros as dependency, fix build on on Leap 15.x * Tue Sep 28 2021 jengelh@inai.de - Update to release 9.3 * New reset-brick command * Ability to get node level status of a cluster * Multi-threaded self-heal for Disperse volumes * Lock revocation feature * On-demand scrubbing for bitrot detection * Real time Cluster notifications using Events APIs - Move mount helper to /usr/sbin [boo#1191062] * Thu Apr 29 2021 jengelh@inai.de - Update to release 9.1 * Implemented seek file operation for open-behind * Added io_uring support for Gluster * Added Support for 5K volumes * Enabled granular-entry-heal by default * Added support for monitoring the epoll/rpc layer * Improve rebalance of sparse files - Drop nocommon.patch (merged) * Thu Jun 04 2020 callumjfarmer13@gmail.com - Fixes for %_libexecdir changing to /usr/libexec * Thu Jan 09 2020 jengelh@inai.de - Update to release 7.1 * Fix rebalance causing IO Error - File descriptor in bad state * Fix geo-replication sessions going faulty * Fix memory leak in glusterfsd process * Fix GlusterFS brick process crash - Add nocommon.patch [boo#1160392] * Thu Dec 26 2019 ngompa13@gmail.com - Fix the name of the Python bindings subpackage - Properly mark Python bindings subpackage as noarch * Fri Dec 13 2019 jengelh@inai.de - Update to release 7.0 * Latency based read child to improve read workload latency in a cluster, especially in a cloud setup. Also provides a load balancing with the outstanding pending request. * Thin-arbiter integration with GD1. * If a client older than glusterfs-4.x (i.e. 3.x clients) accesses a volume which has the `fips-mode-rchecksum` volume option enabled, it can cause erroneous checksum computation/ unwanted behaviour during afr self-heal. * Wed Apr 24 2019 mliska@suse.cz - Move definition of _lto_cflags into %build. * Tue Apr 23 2019 mliska@suse.cz - Disable LTO (boo#1133088). * Sun Mar 17 2019 seanlew@opensuse.org - Update to new upstream version 5.5 * No changelog yet available in release notes * Sat Mar 02 2019 seanlew@opensuse.org - Update to new upstream version 5.4 * FUSE Client memory leak fixed (#1623107) * fail to sync non-ascii utf8 file and dir names (#1648642) * Tracker bug for all leases related issues (#1651323) * gluster-blockd segfaults because of a null-deref (#1659563) * Memory leak: dict_t leak in rda_opendir (#1659676) * dht_revalidate may not heal attrs on the brick root (#1660736) * Fix tests/bugs/shard/zero-flag.t (#1660932) * NL cache: fix typos (#1662200) * Sat Jan 19 2019 seanlew@opensuse.org - Update to new upstream version 5.3 (4.0 is EOL) * 5.3 brings hundreds of improvements, see complete relnote at https://docs.gluster.org/en/latest/release-notes/5.0/ * Fixed a number of bugs: CVE-2018-1088, CVE-2018-1112 [boo#1090084], CVE-2018-10904 [boo#1107018], CVE-2018-10907 [boo#1107019], CVE-2018-10911 [boo#1107020], CVE-2018-10913 [boo#1107021], CVE-2018-10914 [boo#1107022], CVE-2018-10923 [boo#1107023], CVE-2018-10924 [boo#1107024], CVE-2018-10926 [boo#1107025], CVE-2018-10927 [boo#1107026], CVE-2018-10928 [boo#1107027], CVE-2018-10928 [boo#1107027], CVE-2018-10929 [boo#1107028], CVE-2018-10930 [boo#1107029], boo#1105776 . - Remove patches no longer needed in 5.x: * 0001-build-restore-ability-to-control-verbosity-settings.patch * glusterfs-tirpc.diff * Thu May 03 2018 jengelh@inai.de - Update to new upstream release 4.0.2 * GlusterD2 (GD2) is a new management daemon for Gluster-4.0, glustercli is the command-line interface for it. It is a preview release. GD2 is not backwards compatible with the older GlusterD. Heterogeneous clusters running both GD2 and GlusterD are not possible. * Not all commands from GlusterD have been implemented for GD2. * GD2 does not (yet) work well in 2-node clusters. * Self Heal feature integrated for the new Volumes created using Glusterd2. * Thu Dec 14 2017 jengelh@inai.de - Add 0001-build-restore-ability-to-control-verbosity-settings.patch - Replace xdr_sizeof.c with an empty file * Mon Dec 04 2017 jengelh@inai.de - Update to new release 3.12.3 * -- Features introduced in 3.10 * Brick multiplexing * Support to display op-version information from clients * Support for rebalance time to completion estimation * Separation of tier as its own service * Statedump support for gfapi based applications * Disabled creation of trash directory by default * Implemented parallel readdirp with distribute xlator * -- Features introduced in 3.12: * Ability to mount sub-directories using the Gluster FUSE protocol * GFID to path conversion is enabled by default * Provided an option to set a limit on number of bricks multiplexed in a processes * Enhanced the option to export statfs data for bricks sharing the same backend filesystem * Provided a means to resolve GFID split-brain using the gluster CLI * Introduction of new api glfs_fd_set_lkowner() to set lock owner - Remove glusterfs-date.diff (no longer applies) * Mon Nov 13 2017 kukuk@suse.de - glusterfs-tirpc.diff: fallback to libtirpc if there is no sunrpc support in glibc * Fri Nov 03 2017 mpluskal@suse.com - Explicitly require python2 as python3 is not yet supported * Wed Oct 18 2017 vcizek@suse.com - Add missing Requires: libacl-devel to the devel subpackage * glfs.h #includes <sys/acl.h> * Wed Aug 23 2017 vcizek@suse.com - Update to version 3.8.15 * fixes build with OpenSSL 1.1 (bsc#1055466) * Mandatory lock support for Multiprotocol environment * Gluster/NFS disabled by default * Geo-replication for Sharded Volumes * Tiering aware Geo-replication * Enhance Quota enable/disable in glusterd * Automagic unsplit-brain by [ctime|mtime|size|majority] for AFR * Granular Entry self-heal for AFR * Tons of bugfixes * for more details see the NEWS file - spec cleanup - Install License files/docu * Tue Oct 27 2015 scorot@free.fr - Update to version 3.6.6 * No ChangeLog available * Tue Aug 04 2015 scorot@free.fr - Update to version 3.6.4 * No ChangeLog available - Fix Url * Thu Apr 09 2015 afaerber@suse.de - Update to version 3.6.2 * glusterfs-api.pc had version 0.0.0 in v3.6.1, now 4.3.6.2 This is still a backwards jump from v3.6.0 (7.0.0) and v3.5.3 (6). * Package glfsheal again * Thu Jan 15 2015 p.drouand@gmail.com - Update to version 3.6.1 * api: versioned symbols in libgfapi.so for compatibility * Use versioned symbols to keep libgfapi at libgfapi.so.0.0.0 * Some nits uncovered: - there are a couple functions declared that do not have an associated definition, e.g. glfs_truncate(), glfs_caller_specific_init() - there are seven private/internal functions used by heal/src/glfsheal and the gfapi master xlator (glfs-master.c): glfs_loc_touchup(), glfs_active_subvol(), and glfs_subvol_done(), glfs_init_done(), glfs_resolve_at(), glfs_free_from_ctx(), and glfs_new_from_ctx(); which are not declared in glfs.h; - for this initial pass at versioned symbols, we use the earliest version of all public symbols, i.e. those for which there are declarations in glfs.h or glfs-handles.h. Further investigation as we do backports to 3.6, 3.4, and 3.4 will be required to determine if older implementations need to be preserved (forward ported) and their associated alias(es) and symbol version(s) defined. * Mon Aug 04 2014 scorot@free.fr - Update to new upstream release 3.5.2 * NFS server crashes in _socket_read_vectored_request * Can't write to quota enable folder * nfs: reset command does not alter the result for nfs options earlier set * features/gfid-access: stat on .gfid virtual directory return EINVAL * creating symlinks generates errors on stripe volume * Self-heal errors with "afr crawl failed for child 0 with ret -1" while performing rolling upgrade. * [AFR] I/O fails when one of the replica nodes go down * Fix inode leaks in gfid-access xlator * NFS subdir authentication doesn't correctly handle multi-(homed,protocol,etc) network addresses * nfs-utils should be installed as dependency while installing glusterfs-server * Excessive logging in quotad.log of the kind 'null client' * client_t clienttable cliententries are never expanded when all entries are used * AFR : self-heal of few files not happening when a AWS EC2 Instance is back online after a restart * Dist-geo-rep : deletion of files on master, geo-rep fails to propagate to slaves. * Allow the usage of the wildcard character '*' to the options "nfs.rpc-auth-allow" and "nfs.rpc-auth-reject" * glfsheal: Improve the way in which we check the presence of replica volumes * Resource cleanup doesn't happen for clients on servers after disconnect * mounting a volume over NFS (TCP) with MOUNT over UDP fails * backport 'gluster volume status --xml' issues * Glustershd memory usage too high * Tue Jul 29 2014 scorot@free.fr - Update to new upstream release 3.5.1 * A new volume option server.manage-gids has been added. This option should be used when users of a volume are in more than approximately 93 groups (Bug 1096425). * Duplicate Request Cache for NFS has now been disabled by default, this may reduce performance for certain workloads, but improves the overall stability and memory footprint for most users. * Others changes are mostly bug fixes. - disable systemd pre an post scripts for old product and then fix build on SLE 11 * Mon May 05 2014 jengelh@inai.de - Update to new upstream release 3.5.0 * AFR_CLI_enhancements: Improved logging with more clarity and statistical information. It allows visibility into why a self-heal process was initiated and which files are affected, for example. Prior to this enhancement, clearly identifying split-brain issues from the logs was often difficult, and there was no facility to identify which files were affected by a split brain issue automatically. Remediating split brain without quorum will still require some manual effort, but with the tools provided, this will become much simpler. * Exposing Volume Capabilities: Provides client-side insight into whether a volume is using the BD translator and, if so, which capabilities are being utilized. * File Snapshot: Provides a mechanism for snapshotting individual files. The most prevalent use case for this feature will be to snapshot running VMs, allowing for point-in-time capture. This also allows a mechanism to revert VMs to a previous state directly from Gluster, without needing to use external tools. * GFID Access: A new method for accessing data directly by GFID. With this method, the data can be directly consumed in changelog translator, which is logging ‘gfid’ internally, very efficiently. * On-Wire Compression + Decompression: Use of this feature reduces the overall network overhead for Gluster operations from a client. * Prevent NFS restart on Volume change (Part 1): Previously, any volume change (volume option, volume start, volume stop, volume delete, brick add, etc.) would restart the NFS server, which led to service disruptions. This feature allow modifying certain NFS-based volume options without such interruptions occurring. Part 1 is anything not requiring a graph change. * Quota Scalability: Massively increase the amount of quota configurations from a few hundred to 65536 per volume. * readdir_ahead: Gluster now provides read-ahead support for directories to improve sequential directory read performance. * zerofill: Enhancement to allow zeroing out of VM disk images, which is useful in first time provisioning or for overwriting an existing disk. * Brick Failure Detection: Detecting failures on the filesystem that a brick uses makes it possible to handle errors that are caused from outside of the Gluster environment. * Disk encryption: Implement the previous work done in HekaFS into Gluster. This allows a volume (or per-tenant part of a volume) to be encrypted “at rest” on the server using keys only available on the client. [Note: Only content of regular files is encrypted. File names are not encrypted! Also, encryption does not work in NFS mounts.] * Geo-Replication Enhancement: Previously, the geo-replication process, gsyncd, was a single point of failure as it only ran on one node in the cluster. If the node running gsyncd failed, the entire geo-replication process was offline until the issue was addressed. In this latest incarnation, the improvement is extended even further by foregoing use of xattrs to identify change candidates and directly consuming from the volume changelog, which will improve performance twofold: one, by keeping a running list of only those files that may need to be synced; and two, the changelog is maintained in memory, which will allow near instant access to which data needs to be changed and where by the gsync daemon. * Thu Feb 28 2013 jengelh@inai.de - Update to new upstream release 3.4.0alpha (rpm: 3.4.0~qa9) * automake-1.13 support - Enable AIO support * Tue Nov 27 2012 jengelh@inai.de - Use `glusterd -N` in glusterd.service to run in foreground as required * Tue Nov 27 2012 cfarrell@suse.com - license update: GPL-2.0 or LGPL-3.0+ * Fri Nov 09 2012 jengelh@inai.de - Update to new upstream release 3.4.0qa2 * No changelog provided by upstream - Remove glusterfs-init.diff, merged upstream - Provide systemd service file * Wed Oct 31 2012 jengelh@inai.de - Update to new upstream release 3.3.1 * mount.glusterfs: Add support for {attribute,entry}-timeout options * cli: Proper xml output for "gluster peer status" * self-heald: Fix inode leak * storage/posix: implement native linux AIO support * Mon Sep 24 2012 jengelh@inai.de - Update to new upstream release 3.3.0 * New: Unified File & Object access * New: Hadoop hooks - HDFS compatibility layer * New volume type: Repstr - replicated + striped (+ distributed) volumes * Fri Dec 02 2011 coolo@suse.com - add automake as buildrequire to avoid implicit dependency * Wed Oct 05 2011 jengelh@medozas.de - Initial package for build.opensuse.org
/usr/lib64/libglusterfs.so.0 /usr/lib64/libglusterfs.so.0.0.1
Generated by rpm2html 1.8.1
Fabrice Bellet, Fri Oct 4 00:15:39 2024