[Gluster-users] Disk size and virtual size drive me crazy!
Gilberto Ferreira
gilberto.nunes32 at gmail.com
Fri Nov 29 19:47:18 UTC 2024
No! I didn't! I wasn't aware of this option.
I will try.
Thanks
Em sex., 29 de nov. de 2024 às 16:43, Strahil Nikolov <hunter86_bg at yahoo.com>
escreveu:
> Have you figured it out ?
>
> Have you tried setting storage.reserve to 0 ?
>
> Best Regards,
> Strahil Nikolov
>
> On Thu, Nov 21, 2024 at 0:39, Gilberto Ferreira
> <gilberto.nunes32 at gmail.com> wrote:
>
> 11.1
> ---
> Gilberto Nunes Ferreira
> +55 (47) 99676-7530
> Proxmox VE
> VinChin Backup & Restore
>
> Em qua., 20 de nov. de 2024, 19:28, Strahil Nikolov <hunter86_bg at yahoo.com>
> escreveu:
>
> What's your gluster version ?
>
> Best Regards,
> Strahil Nikolov
>
> В понеделник, 11 ноември 2024 г. в 20:57:50 ч. Гринуич+2, Gilberto
> Ferreira <gilberto.nunes32 at gmail.com> написа:
>
>
> Hi there.
>
> I can't understand why I am having this different values:
>
> proxmox01:/vms/images# df
> Sist. Arq. Tam. Usado Disp. Uso% Montado em
> udev 252G 0 252G 0% /dev
> tmpfs 51G 9,4M 51G 1% /run
> /dev/sda4 433G 20G 413G 5% /
> tmpfs 252G 63M 252G 1% /dev/shm
> tmpfs 5,0M 0 5,0M 0% /run/lock
> efivarfs 496K 335K 157K 69% /sys/firmware/efi/efivars
> /dev/sda2 1,8G 204M 1,5G 12% /boot
> /dev/sda1 1,9G 12M 1,9G 1% /boot/efi
> /dev/sdb 932G 728G 204G 79% /disco1TB-0
> /dev/sdc 932G 718G 214G 78% /disco1TB-1
> /dev/sde 932G 720G 212G 78% /disco1TB-2
> /dev/sdd 1,9T 1,5T 387G 80% /disco2TB-0
> tmpfs 51G 4,0K 51G 1% /run/user/0
> *gluster1:VMS 4,6T 3,6T 970G 80% /vms*
> /dev/fuse 128M 36K 128M 1% /etc/pve
> proxmox01:/vms/images# cd 103
> proxmox01:/vms/images/103# ls
> vm-103-disk-0.qcow2 vm-103-disk-1.qcow2
> proxmox01:/vms/images/103# ls -lh
> total 21T
>
> *-rw-r----- 1 root root 101G nov 11 15:53 vm-103-disk-0.qcow2-rw-r----- 1
> root root 210G nov 11 15:45 vm-103-disk-1.qcow2*
> proxmox01:/vms/images/103# qemu-img info vm-103-disk-0.qcow2
> image: vm-103-disk-0.qcow2
> file format: qcow2
>
> *virtual size: 100 GiB (107374182400 bytes)disk size: 3.78 TiB*
> cluster_size: 65536
> Format specific information:
> compat: 1.1
> compression type: zlib
> lazy refcounts: false
> refcount bits: 16
> corrupt: false
> extended l2: false
> Child node '/file':
> filename: vm-103-disk-0.qcow2
> protocol type: file
>
> * file length: 100 GiB (107390828544 bytes) disk size: 3.78 TiB*
> proxmox01:/vms/images/103# qemu-img info vm-103-disk-1.qcow2
> image: vm-103-disk-1.qcow2
> file format: qcow2
>
> *virtual size: 2 TiB (2199023255552 bytes)disk size: 16.3 TiB*
> cluster_size: 65536
> Format specific information:
> compat: 1.1
> compression type: zlib
> lazy refcounts: false
> refcount bits: 16
> corrupt: false
> extended l2: false
> Child node '/file':
> filename: vm-103-disk-1.qcow2
> protocol type: file
>
> * file length: 210 GiB (225117732864 bytes) disk size: 16.3 TiB*
> proxmox01:/vms/images/103#
>
> Here is the vol info.
>
> proxmox01:/vms/images/103# gluster vol info
>
> Volume Name: VMS
> Type: Distributed-Replicate
> Volume ID: a98f7944-4308-499f-994e-9029f3be56c0
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 4 x 2 = 8
> Transport-type: tcp
> Bricks:
> Brick1: gluster1:/disco2TB-0/vms
> Brick2: gluster2:/disco2TB-0/vms
> Brick3: gluster1:/disco1TB-0/vms
> Brick4: gluster2:/disco1TB-0/vms
> Brick5: gluster1:/disco1TB-1/vms
> Brick6: gluster2:/disco1TB-1/vms
> Brick7: gluster1:/disco1TB-2/vms
> Brick8: gluster2:/disco1TB-2/vms
> Options Reconfigured:
> cluster.lookup-optimize: off
> server.keepalive-count: 5
> server.keepalive-interval: 2
> server.keepalive-time: 10
> server.tcp-user-timeout: 20
> server.event-threads: 4
> client.event-threads: 4
> cluster.choose-local: off
> cluster.shd-wait-qlength: 10000
> cluster.shd-max-threads: 8
> cluster.locking-scheme: granular
> cluster.server-quorum-type: none
> cluster.quorum-type: fixed
> network.remote-dio: disable
> performance.client-io-threads: on
> performance.strict-o-direct: on
> performance.low-prio-threads: 32
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> performance.flush-behind: off
> performance.write-behind: off
> cluster.data-self-heal-algorithm: full
> cluster.favorite-child-policy: mtime
> network.ping-timeout: 20
> cluster.quorum-count: 1
> cluster.quorum-reads: false
> cluster.self-heal-daemon: enable
> cluster.heal-timeout: 5
> user.cifs: off
> features.shard: on
> cluster.granular-entry-heal: enable
> storage.fips-mode-rchecksum: on
> transport.address-family: inet
> nfs.disable: on
> ---
>
>
> Gilberto Nunes Ferreira
>
>
>
>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20241129/346e0434/attachment.html>
More information about the Gluster-users
mailing list