[Bugs] [Bug 1263746] New: Data Tiering:Setting only promote frequency and no demote frequency causes crash

bugzilla at redhat.com bugzilla at redhat.com
Wed Sep 16 15:04:42 UTC 2015


https://bugzilla.redhat.com/show_bug.cgi?id=1263746

            Bug ID: 1263746
           Summary: Data Tiering:Setting only promote frequency and no
                    demote frequency causes crash
           Product: GlusterFS
           Version: 3.7.5
         Component: tiering
          Severity: urgent
          Priority: urgent
          Assignee: bugs at gluster.org
          Reporter: dlambrig at redhat.com
        QA Contact: bugs at gluster.org
                CC: bugs at gluster.org, dlambrig at redhat.com,
                    nbalacha at redhat.com, nchilaka at redhat.com
        Depends On: 1263200, 1263204



+++ This bug was initially created as a clone of Bug #1263204 +++

+++ This bug was initially created as a clone of Bug #1263200 +++

Description of problem:
=======================
I created a regular volume and created some files cf1,cf2,cf3 and started with
linux kernel files

I then attached a tier 
Set the ctr enable and promote freq to 10sec.
I then tried to modify files cf* using touch command.

But the files were not at all getting promoted (seperate bz#1262885)

I kept the volume idle for some few hours and then I saw that a crash was hit.


Note: As part of another unrelated issue, I was hitting glusterd crashes, I did
a work-around suggested by dev. to modify/add below option in glusterd.vol file
and restart glusterd.
   option ping-timeout 0
   option event-threads 1

I restarted glusterd a couple of time post this modification


Another problem after crash is wrong info of ec cold volume. It shows a two way
distributed EC cold tier as 12 way distributed.

Before crash:
[root at zod ~]# gluster v info 9301

Volume Name: 9301
Type: Tier
Volume ID: 0314fa86-49dc-4fbe-925f-8080157a9c8b
Status: Started
Number of Bricks: 16
Transport-type: tcp
Hot Tier :
Hot Tier Type : Distributed-Replicate
Number of Bricks: 2 x 2 = 4
Brick1: yarrow:/rhs/brick6/9301_hot
Brick2: zod:/rhs/brick6/9301_hot
Brick3: yarrow:/rhs/brick7/9301_hot
Brick4: zod:/rhs/brick7/9301_hot
Cold Tier:
Cold Tier Type : Distributed-Disperse
Number of Bricks: 2 x (4 + 2) = 12
Brick5: zod:/rhs/brick1/9301
Brick6: yarrow:/rhs/brick1/9301
Brick7: zod:/rhs/brick2/9301
Brick8: yarrow:/rhs/brick2/9301
Brick9: zod:/rhs/brick3/9301
Brick10: yarrow:/rhs/brick3/9301
Brick11: zod:/rhs/brick4/9301
Brick12: yarrow:/rhs/brick4/9301
Brick13: zod:/rhs/brick5/9301
Brick14: yarrow:/rhs/brick5/9301
Brick15: yarrow:/rhs/brick6/9301
Brick16: zod:/rhs/brick6/9301
Options Reconfigured:
cluster.tier-promote-frequency: 10
features.ctr-enabled: on
performance.io-cache: off
performance.quick-read: off
performance.readdir-ahead: on




After crash:
Volume Name: 9301
Type: Tier
Volume ID: 0314fa86-49dc-4fbe-925f-8080157a9c8b
Status: Started
Number of Bricks: 16
Transport-type: tcp
Hot Tier :
Hot Tier Type : Distributed-Replicate
Number of Bricks: 2 x 2 = 4
Brick1: yarrow:/rhs/brick6/9301_hot
Brick2: zod:/rhs/brick6/9301_hot
Brick3: yarrow:/rhs/brick7/9301_hot
Brick4: zod:/rhs/brick7/9301_hot
Cold Tier:
Cold Tier Type : Distributed-Disperse
Number of Bricks: 12 x (4 + 2) = 12
Brick5: zod:/rhs/brick1/9301
Brick6: yarrow:/rhs/brick1/9301
Brick7: zod:/rhs/brick2/9301
Brick8: yarrow:/rhs/brick2/9301
Brick9: zod:/rhs/brick3/9301
Brick10: yarrow:/rhs/brick3/9301
Brick11: zod:/rhs/brick4/9301
Brick12: yarrow:/rhs/brick4/9301
Brick13: zod:/rhs/brick5/9301
Brick14: yarrow:/rhs/brick5/9301
Brick15: yarrow:/rhs/brick6/9301
Brick16: zod:/rhs/brick6/9301
Options Reconfigured:
performance.readdir-ahead: on
performance.quick-read: off
performance.io-cache: off
features.ctr-enabled: on
cluster.tier-promote-frequency: 10
cluster.read-freq-threshold: 0
cluster.write-freq-threshold: 0




core.2839: ELF 64-bit LSB core file x86-64, version 1 (SYSV), SVR4-style, from
'/usr/sbin/glusterfs -s localhost --volfile-id rebalance/9301 --xlator-option
*d'
[root at zod /]# gdb /usr/sbin/glusterfs core.2839
GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-64.el7
Copyright (C) 2013 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.  Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-redhat-linux-gnu".
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>...
Reading symbols from /usr/sbin/glusterfsd...Reading symbols from
/usr/lib/debug/usr/sbin/glusterfsd.debug...done.
done.

warning: core file may not match specified executable file.
[New LWP 13112]
[New LWP 2844]
[New LWP 2860]
[New LWP 2841]
[New LWP 2842]
[New LWP 2843]
[New LWP 2840]
[New LWP 2863]
[New LWP 2866]
[New LWP 2854]
[New LWP 2859]
[New LWP 2865]
[New LWP 2862]
[New LWP 2864]
[New LWP 2839]
[New LWP 2861]
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib64/libthread_db.so.1".
Core was generated by `/usr/sbin/glusterfs -s localhost --volfile-id
rebalance/9301 --xlator-option *d'.
Program terminated with signal 11, Segmentation fault.
#0  tier_build_migration_qfile (is_promotion=_gf_true, 
    query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
607            list_for_each_entry (local_brick, args->brick_list, list) {
Missing separate debuginfos, use: debuginfo-install glibc-2.17-78.el7.x86_64
keyutils-libs-1.5.8-3.el7.x86_64 krb5-libs-1.12.2-14.el7.x86_64
libcom_err-1.42.9-7.el7.x86_64 libgcc-4.8.3-9.el7.x86_64
libselinux-2.2.2-6.el7.x86_64 libuuid-2.23.2-22.el7_1.1.x86_64
openssl-libs-1.0.1e-42.el7_1.9.x86_64 pcre-8.32-14.el7.x86_64
sqlite-3.7.17-6.el7_1.1.x86_64 sssd-client-1.12.2-58.el7_1.14.x86_64
xz-libs-5.1.2-9alpha.el7.x86_64 zlib-1.2.7-13.el7.x86_64
(gdb) bt
#0  tier_build_migration_qfile (is_promotion=_gf_true, 
    query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
#1  tier_promote (args=0x7f3fee27bca0) at tier.c:704
#2  0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3  0x00007f40035171ad in clone () from /lib64/libc.so.6

--- Additional comment from nchilaka on 2015-09-15 06:22:34 EDT ---

[root at rhsqe-repo sosreports]# hostname
rhsqe-repo.lab.eng.blr.redhat.com
[root at rhsqe-repo sosreports]# ls /home/repo/sosreports/bug.1263200

--- Additional comment from Nithya Balachandran on 2015-09-15 06:25:23 EDT ---

nalysis of the coredump:

[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib64/libthread_db.so.1".
Core was generated by `/usr/sbin/glusterfs -s localhost --volfile-id
rebalance/9301 --xlator-option *d'.
Program terminated with signal 11, Segmentation fault.
#0  tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
607             list_for_each_entry (local_brick, args->brick_list, list) {

#0  tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
#1  tier_promote (args=0x7f3fee27bca0) at tier.c:704
#2  0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3  0x00007f40035171ad in clone () from /lib64/libc.so.6
(gdb) f 0
#0  tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
607             list_for_each_entry (local_brick, args->brick_list, list) {
(gdb) p *args
$1 = {this = 0x0, defrag = 0x0, brick_list = 0x0, freq_time = 0, return_value =
0}


All members of the args structure are NULL, causing the tier process to crash
when it tries to access args->brick_list.

(gdb) t a a bt

Thread 16 (Thread 0x7f3fdbfff700 (LWP 2861)):
#0  0x00007f4003bd4705 in pthread_cond_wait@@GLIBC_2.3.2 () from
/lib64/libpthread.so.0
#1  0x00007f3ff6da434a in gf_defrag_task (opaque=0x7f3ff0028c10) at
dht-rebalance.c:1801
#2  0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3  0x00007f40035171ad in clone () from /lib64/libc.so.6


...



Thread 6 (Thread 0x7f3ffa380700 (LWP 2843)):
#0  0x00007f40034de48d in nanosleep () from /lib64/libc.so.6
#1  0x00007f40034de324 in sleep () from /lib64/libc.so.6
#2  0x00007f3ff6911f2a in tier_start (this=0x7f3ff0020920,
defrag=0x7f3ff0028c10) at tier.c:860
#3  0x00007f3ff6da6d34 in gf_defrag_start_crawl (data=0x7f3ff0020920) at
dht-rebalance.c:2841
#4  0x00007f4004da9d72 in synctask_wrap (old_task=<optimized out>) at
syncop.c:380
#5  0x00007f40034680f0 in ?? () from /lib64/libc.so.6
#6  0x0000000000000000 in ?? ()

Thread 5 (Thread 0x7f3ffab81700 (LWP 2842)):
#0  0x00007f4003bd4ab2 in pthread_cond_timedwait@@GLIBC_2.3.2 () from
/lib64/libpthread.so.0
#1  0x00007f4004dabe88 in syncenv_task (proc=proc at entry=0x7f4006a675a0) at
syncop.c:607
#2  0x00007f4004dacbc0 in syncenv_processor (thdata=0x7f4006a675a0) at
syncop.c:699
#3  0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#4  0x00007f40035171ad in clone () from /lib64/libc.so.6

...

Thread 2 (Thread 0x7f3ff7ca8700 (LWP 2844)):
#0  0x00007f4003517783 in epoll_wait () from /lib64/libc.so.6
---Type <return> to continue, or q <return> to quit---
#1  0x00007f4004dc9680 in event_dispatch_epoll_worker (data=0x7f4006aa45c0) at
event-epoll.c:668
#2  0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3  0x00007f40035171ad in clone () from /lib64/libc.so.6

Thread 1 (Thread 0x7f3afa855700 (LWP 13112)):
#0  tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
#1  tier_promote (args=0x7f3fee27bca0) at tier.c:704
#2  0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3  0x00007f40035171ad in clone () from /lib64/libc.so.6


Looking at thread 6:

(gdb) t 6
[Switching to thread 6 (Thread 0x7f3ffa380700 (LWP 2843))]
#0  0x00007f40034de48d in nanosleep () from /lib64/libc.so.6
(gdb) bt
#0  0x00007f40034de48d in nanosleep () from /lib64/libc.so.6
#1  0x00007f40034de324 in sleep () from /lib64/libc.so.6
#2  0x00007f3ff6911f2a in tier_start (this=0x7f3ff0020920,
defrag=0x7f3ff0028c10) at tier.c:860
#3  0x00007f3ff6da6d34 in gf_defrag_start_crawl (data=0x7f3ff0020920) at
dht-rebalance.c:2841
#4  0x00007f4004da9d72 in synctask_wrap (old_task=<optimized out>) at
syncop.c:380
#5  0x00007f40034680f0 in ?? () from /lib64/libc.so.6
#6  0x0000000000000000 in ?? ()


(gdb) f 2

#2  0x00007f3ff6911f2a in tier_start (this=0x7f3ff0020920,
defrag=0x7f3ff0028c10) at tier.c:860
860                     sleep(1);
(gdb) l
855     
856             defrag->defrag_status = GF_DEFRAG_STATUS_STARTED;
857     
858             while (1) {
859     
860                     sleep(1);
861     


...

911     
912     
913                     ret_promotion = -1;
914                     ret_demotion = -1;
(gdb) 
915     
916                     if (is_demotion_triggered) {
917                             demotion_args.this = this;
918                             demotion_args.brick_list = &bricklist_hot;
919                             demotion_args.defrag = defrag;
920                             demotion_args.freq_time = freq_demote;
921                             ret_demotion = pthread_create (&demote_thread,
922                                                     NULL, &tier_demote,
923                                                     &demotion_args);
924                             if (ret_demotion) {
(gdb) 
925                                     gf_msg (this->name, GF_LOG_ERROR, 0,
926                                             DHT_MSG_LOG_TIER_ERROR,
927                                             "Failed starting Demotion "
928                                             "thread!");
929                             }
930                     }
931     
932                     if (is_promotion_triggered) {
933                             promotion_args.this = this;
934                             promotion_args.brick_list = &bricklist_cold;
(gdb) 
935                             promotion_args.defrag = defrag;
936                             promotion_args.freq_time = freq_promote * 
PROMOTION_CYCLE_CNT;
937                             ret_promotion = pthread_create
(&promote_thread,
938                                                     NULL, &tier_promote,
939                                                     &promotion_args);
940                             if (ret_promotion) {
941                                     gf_msg (this->name, GF_LOG_ERROR, 0,
942                                             DHT_MSG_LOG_TIER_ERROR,
943                                             "Failed starting Promotion "
944                                             "thread!");
                               }


...


958                    if (is_demotion_triggered && (ret_promotion == 0)) {
959                            pthread_join (promote_thread, NULL);
960                            if (promotion_args.return_value) {
961                                    gf_msg (this->name, GF_LOG_ERROR, 0,
962                                            DHT_MSG_LOG_TIER_ERROR,
963                                            "Promotion failed!");
964                            }
(gdb) 
965                            ret_promotion = promotion_args.return_value;
966                    }
967    
968                    /* Collect previous and current cummulative status */
969                    /* If demotion was not triggered just pass 0 to ret */
970                    ret = (is_demotion_triggered) ? ret_demotion : 0;
971                    /* If promotion was not triggered just pass 0 to ret */
972                    ret = ret | (is_promotion_triggered) ?
973                                    ret_promotion : 0;
974    
(gdb) 
975                    /* reseting promotion and demotion arguments for
976                     * next iteration*/
977                    memset (&demotion_args, 0, sizeof(demotion_args_t));
978                    memset (&promotion_args, 0, sizeof(promotion_args_t));
979    
980            }

(gdb) p is_demotion_triggered
$2 = _gf_false

As is_demotion_triggered is false, the pthread_join (promote_thread, NULL) on
line 959 is never called. The main thread proceeds without waiting for the
promote thread to complete and memsets promotion_args to 0, thus causing the
process to crash when the promote_thread tries to access the args structure.

--- Additional comment from Vijay Bellur on 2015-09-15 06:40:34 EDT ---

REVIEW: http://review.gluster.org/12179 (cluster/tier: Fixed a crash in
tiering) posted (#1) for review on master by N Balachandran
(nbalacha at redhat.com)

--- Additional comment from Vijay Bellur on 2015-09-16 09:25:04 EDT ---

COMMIT: http://review.gluster.org/12179 committed in master by Dan Lambright
(dlambrig at redhat.com) 
------
commit 0cf1c851b9fc6b98b5839b32ef7e3ff38d87f3c3
Author: Nithya Balachandran <nbalacha at redhat.com>
Date:   Tue Sep 15 16:07:32 2015 +0530

    cluster/tier: Fixed a crash in tiering

    An incorrect check was causing the arguments to
    the promote thread to be cleared before the thread
    was done with them. This caused the process to crash
    when it tried to dereference a NULL pointer.

    Change-Id: I8348309ef4dad33b7f648c7a2c2703487e401269
    BUG: 1263204
    Signed-off-by: Nithya Balachandran <nbalacha at redhat.com>
    Reviewed-on: http://review.gluster.org/12179
    Tested-by: NetBSD Build System <jenkins at build.gluster.org>
    Reviewed-by: Dan Lambright <dlambrig at redhat.com>
    Reviewed-by: Joseph Fernandes


Referenced Bugs:

https://bugzilla.redhat.com/show_bug.cgi?id=1263200
[Bug 1263200] Data Tiering:Setting only promote frequency and no demote
frequency causes crash
https://bugzilla.redhat.com/show_bug.cgi?id=1263204
[Bug 1263204] Data Tiering:Setting only promote frequency and no demote
frequency causes crash
-- 
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.


More information about the Bugs mailing list