[Bugs] [Bug 1263204] New: Data Tiering:Setting only promote frequency and no demote frequency causes crash
bugzilla at redhat.com
bugzilla at redhat.com
Tue Sep 15 10:27:56 UTC 2015
https://bugzilla.redhat.com/show_bug.cgi?id=1263204
Bug ID: 1263204
Summary: Data Tiering:Setting only promote frequency and no
demote frequency causes crash
Product: GlusterFS
Version: mainline
Component: tiering
Severity: urgent
Priority: urgent
Assignee: bugs at gluster.org
Reporter: nbalacha at redhat.com
QA Contact: bugs at gluster.org
CC: bugs at gluster.org, nbalacha at redhat.com,
nchilaka at redhat.com
Depends On: 1263200
+++ This bug was initially created as a clone of Bug #1263200 +++
Description of problem:
=======================
I created a regular volume and created some files cf1,cf2,cf3 and started with
linux kernel files
I then attached a tier
Set the ctr enable and promote freq to 10sec.
I then tried to modify files cf* using touch command.
But the files were not at all getting promoted (seperate bz#1262885)
I kept the volume idle for some few hours and then I saw that a crash was hit.
Note: As part of another unrelated issue, I was hitting glusterd crashes, I did
a work-around suggested by dev. to modify/add below option in glusterd.vol file
and restart glusterd.
option ping-timeout 0
option event-threads 1
I restarted glusterd a couple of time post this modification
Another problem after crash is wrong info of ec cold volume. It shows a two way
distributed EC cold tier as 12 way distributed.
Before crash:
[root at zod ~]# gluster v info 9301
Volume Name: 9301
Type: Tier
Volume ID: 0314fa86-49dc-4fbe-925f-8080157a9c8b
Status: Started
Number of Bricks: 16
Transport-type: tcp
Hot Tier :
Hot Tier Type : Distributed-Replicate
Number of Bricks: 2 x 2 = 4
Brick1: yarrow:/rhs/brick6/9301_hot
Brick2: zod:/rhs/brick6/9301_hot
Brick3: yarrow:/rhs/brick7/9301_hot
Brick4: zod:/rhs/brick7/9301_hot
Cold Tier:
Cold Tier Type : Distributed-Disperse
Number of Bricks: 2 x (4 + 2) = 12
Brick5: zod:/rhs/brick1/9301
Brick6: yarrow:/rhs/brick1/9301
Brick7: zod:/rhs/brick2/9301
Brick8: yarrow:/rhs/brick2/9301
Brick9: zod:/rhs/brick3/9301
Brick10: yarrow:/rhs/brick3/9301
Brick11: zod:/rhs/brick4/9301
Brick12: yarrow:/rhs/brick4/9301
Brick13: zod:/rhs/brick5/9301
Brick14: yarrow:/rhs/brick5/9301
Brick15: yarrow:/rhs/brick6/9301
Brick16: zod:/rhs/brick6/9301
Options Reconfigured:
cluster.tier-promote-frequency: 10
features.ctr-enabled: on
performance.io-cache: off
performance.quick-read: off
performance.readdir-ahead: on
After crash:
Volume Name: 9301
Type: Tier
Volume ID: 0314fa86-49dc-4fbe-925f-8080157a9c8b
Status: Started
Number of Bricks: 16
Transport-type: tcp
Hot Tier :
Hot Tier Type : Distributed-Replicate
Number of Bricks: 2 x 2 = 4
Brick1: yarrow:/rhs/brick6/9301_hot
Brick2: zod:/rhs/brick6/9301_hot
Brick3: yarrow:/rhs/brick7/9301_hot
Brick4: zod:/rhs/brick7/9301_hot
Cold Tier:
Cold Tier Type : Distributed-Disperse
Number of Bricks: 12 x (4 + 2) = 12
Brick5: zod:/rhs/brick1/9301
Brick6: yarrow:/rhs/brick1/9301
Brick7: zod:/rhs/brick2/9301
Brick8: yarrow:/rhs/brick2/9301
Brick9: zod:/rhs/brick3/9301
Brick10: yarrow:/rhs/brick3/9301
Brick11: zod:/rhs/brick4/9301
Brick12: yarrow:/rhs/brick4/9301
Brick13: zod:/rhs/brick5/9301
Brick14: yarrow:/rhs/brick5/9301
Brick15: yarrow:/rhs/brick6/9301
Brick16: zod:/rhs/brick6/9301
Options Reconfigured:
performance.readdir-ahead: on
performance.quick-read: off
performance.io-cache: off
features.ctr-enabled: on
cluster.tier-promote-frequency: 10
cluster.read-freq-threshold: 0
cluster.write-freq-threshold: 0
core.2839: ELF 64-bit LSB core file x86-64, version 1 (SYSV), SVR4-style, from
'/usr/sbin/glusterfs -s localhost --volfile-id rebalance/9301 --xlator-option
*d'
[root at zod /]# gdb /usr/sbin/glusterfs core.2839
GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-64.el7
Copyright (C) 2013 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law. Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-redhat-linux-gnu".
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>...
Reading symbols from /usr/sbin/glusterfsd...Reading symbols from
/usr/lib/debug/usr/sbin/glusterfsd.debug...done.
done.
warning: core file may not match specified executable file.
[New LWP 13112]
[New LWP 2844]
[New LWP 2860]
[New LWP 2841]
[New LWP 2842]
[New LWP 2843]
[New LWP 2840]
[New LWP 2863]
[New LWP 2866]
[New LWP 2854]
[New LWP 2859]
[New LWP 2865]
[New LWP 2862]
[New LWP 2864]
[New LWP 2839]
[New LWP 2861]
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib64/libthread_db.so.1".
Core was generated by `/usr/sbin/glusterfs -s localhost --volfile-id
rebalance/9301 --xlator-option *d'.
Program terminated with signal 11, Segmentation fault.
#0 tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
607 list_for_each_entry (local_brick, args->brick_list, list) {
Missing separate debuginfos, use: debuginfo-install glibc-2.17-78.el7.x86_64
keyutils-libs-1.5.8-3.el7.x86_64 krb5-libs-1.12.2-14.el7.x86_64
libcom_err-1.42.9-7.el7.x86_64 libgcc-4.8.3-9.el7.x86_64
libselinux-2.2.2-6.el7.x86_64 libuuid-2.23.2-22.el7_1.1.x86_64
openssl-libs-1.0.1e-42.el7_1.9.x86_64 pcre-8.32-14.el7.x86_64
sqlite-3.7.17-6.el7_1.1.x86_64 sssd-client-1.12.2-58.el7_1.14.x86_64
xz-libs-5.1.2-9alpha.el7.x86_64 zlib-1.2.7-13.el7.x86_64
(gdb) bt
#0 tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
#1 tier_promote (args=0x7f3fee27bca0) at tier.c:704
#2 0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3 0x00007f40035171ad in clone () from /lib64/libc.so.6
--- Additional comment from nchilaka on 2015-09-15 06:22:34 EDT ---
[root at rhsqe-repo sosreports]# hostname
rhsqe-repo.lab.eng.blr.redhat.com
[root at rhsqe-repo sosreports]# ls /home/repo/sosreports/bug.1263200
--- Additional comment from Nithya Balachandran on 2015-09-15 06:25:23 EDT ---
nalysis of the coredump:
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib64/libthread_db.so.1".
Core was generated by `/usr/sbin/glusterfs -s localhost --volfile-id
rebalance/9301 --xlator-option *d'.
Program terminated with signal 11, Segmentation fault.
#0 tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
607 list_for_each_entry (local_brick, args->brick_list, list) {
#0 tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
#1 tier_promote (args=0x7f3fee27bca0) at tier.c:704
#2 0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3 0x00007f40035171ad in clone () from /lib64/libc.so.6
(gdb) f 0
#0 tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
607 list_for_each_entry (local_brick, args->brick_list, list) {
(gdb) p *args
$1 = {this = 0x0, defrag = 0x0, brick_list = 0x0, freq_time = 0, return_value =
0}
All members of the args structure are NULL, causing the tier process to crash
when it tries to access args->brick_list.
(gdb) t a a bt
Thread 16 (Thread 0x7f3fdbfff700 (LWP 2861)):
#0 0x00007f4003bd4705 in pthread_cond_wait@@GLIBC_2.3.2 () from
/lib64/libpthread.so.0
#1 0x00007f3ff6da434a in gf_defrag_task (opaque=0x7f3ff0028c10) at
dht-rebalance.c:1801
#2 0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3 0x00007f40035171ad in clone () from /lib64/libc.so.6
...
Thread 6 (Thread 0x7f3ffa380700 (LWP 2843)):
#0 0x00007f40034de48d in nanosleep () from /lib64/libc.so.6
#1 0x00007f40034de324 in sleep () from /lib64/libc.so.6
#2 0x00007f3ff6911f2a in tier_start (this=0x7f3ff0020920,
defrag=0x7f3ff0028c10) at tier.c:860
#3 0x00007f3ff6da6d34 in gf_defrag_start_crawl (data=0x7f3ff0020920) at
dht-rebalance.c:2841
#4 0x00007f4004da9d72 in synctask_wrap (old_task=<optimized out>) at
syncop.c:380
#5 0x00007f40034680f0 in ?? () from /lib64/libc.so.6
#6 0x0000000000000000 in ?? ()
Thread 5 (Thread 0x7f3ffab81700 (LWP 2842)):
#0 0x00007f4003bd4ab2 in pthread_cond_timedwait@@GLIBC_2.3.2 () from
/lib64/libpthread.so.0
#1 0x00007f4004dabe88 in syncenv_task (proc=proc at entry=0x7f4006a675a0) at
syncop.c:607
#2 0x00007f4004dacbc0 in syncenv_processor (thdata=0x7f4006a675a0) at
syncop.c:699
#3 0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#4 0x00007f40035171ad in clone () from /lib64/libc.so.6
...
Thread 2 (Thread 0x7f3ff7ca8700 (LWP 2844)):
#0 0x00007f4003517783 in epoll_wait () from /lib64/libc.so.6
---Type <return> to continue, or q <return> to quit---
#1 0x00007f4004dc9680 in event_dispatch_epoll_worker (data=0x7f4006aa45c0) at
event-epoll.c:668
#2 0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3 0x00007f40035171ad in clone () from /lib64/libc.so.6
Thread 1 (Thread 0x7f3afa855700 (LWP 13112)):
#0 tier_build_migration_qfile (is_promotion=_gf_true,
query_cbk_args=0x7f3afa854e70, args=0x7f3fee27bca0) at tier.c:607
#1 tier_promote (args=0x7f3fee27bca0) at tier.c:704
#2 0x00007f4003bd0df5 in start_thread () from /lib64/libpthread.so.0
#3 0x00007f40035171ad in clone () from /lib64/libc.so.6
Looking at thread 6:
(gdb) t 6
[Switching to thread 6 (Thread 0x7f3ffa380700 (LWP 2843))]
#0 0x00007f40034de48d in nanosleep () from /lib64/libc.so.6
(gdb) bt
#0 0x00007f40034de48d in nanosleep () from /lib64/libc.so.6
#1 0x00007f40034de324 in sleep () from /lib64/libc.so.6
#2 0x00007f3ff6911f2a in tier_start (this=0x7f3ff0020920,
defrag=0x7f3ff0028c10) at tier.c:860
#3 0x00007f3ff6da6d34 in gf_defrag_start_crawl (data=0x7f3ff0020920) at
dht-rebalance.c:2841
#4 0x00007f4004da9d72 in synctask_wrap (old_task=<optimized out>) at
syncop.c:380
#5 0x00007f40034680f0 in ?? () from /lib64/libc.so.6
#6 0x0000000000000000 in ?? ()
(gdb) f 2
#2 0x00007f3ff6911f2a in tier_start (this=0x7f3ff0020920,
defrag=0x7f3ff0028c10) at tier.c:860
860 sleep(1);
(gdb) l
855
856 defrag->defrag_status = GF_DEFRAG_STATUS_STARTED;
857
858 while (1) {
859
860 sleep(1);
861
...
911
912
913 ret_promotion = -1;
914 ret_demotion = -1;
(gdb)
915
916 if (is_demotion_triggered) {
917 demotion_args.this = this;
918 demotion_args.brick_list = &bricklist_hot;
919 demotion_args.defrag = defrag;
920 demotion_args.freq_time = freq_demote;
921 ret_demotion = pthread_create (&demote_thread,
922 NULL, &tier_demote,
923 &demotion_args);
924 if (ret_demotion) {
(gdb)
925 gf_msg (this->name, GF_LOG_ERROR, 0,
926 DHT_MSG_LOG_TIER_ERROR,
927 "Failed starting Demotion "
928 "thread!");
929 }
930 }
931
932 if (is_promotion_triggered) {
933 promotion_args.this = this;
934 promotion_args.brick_list = &bricklist_cold;
(gdb)
935 promotion_args.defrag = defrag;
936 promotion_args.freq_time = freq_promote *
PROMOTION_CYCLE_CNT;
937 ret_promotion = pthread_create
(&promote_thread,
938 NULL, &tier_promote,
939 &promotion_args);
940 if (ret_promotion) {
941 gf_msg (this->name, GF_LOG_ERROR, 0,
942 DHT_MSG_LOG_TIER_ERROR,
943 "Failed starting Promotion "
944 "thread!");
}
...
958 if (is_demotion_triggered && (ret_promotion == 0)) {
959 pthread_join (promote_thread, NULL);
960 if (promotion_args.return_value) {
961 gf_msg (this->name, GF_LOG_ERROR, 0,
962 DHT_MSG_LOG_TIER_ERROR,
963 "Promotion failed!");
964 }
(gdb)
965 ret_promotion = promotion_args.return_value;
966 }
967
968 /* Collect previous and current cummulative status */
969 /* If demotion was not triggered just pass 0 to ret */
970 ret = (is_demotion_triggered) ? ret_demotion : 0;
971 /* If promotion was not triggered just pass 0 to ret */
972 ret = ret | (is_promotion_triggered) ?
973 ret_promotion : 0;
974
(gdb)
975 /* reseting promotion and demotion arguments for
976 * next iteration*/
977 memset (&demotion_args, 0, sizeof(demotion_args_t));
978 memset (&promotion_args, 0, sizeof(promotion_args_t));
979
980 }
(gdb) p is_demotion_triggered
$2 = _gf_false
As is_demotion_triggered is false, the pthread_join (promote_thread, NULL) on
line 959 is never called. The main thread proceeds without waiting for the
promote thread to complete and memsets promotion_args to 0, thus causing the
process to crash when the promote_thread tries to access the args structure.
Referenced Bugs:
https://bugzilla.redhat.com/show_bug.cgi?id=1263200
[Bug 1263200] Data Tiering:Setting only promote frequency and no demote
frequency causes crash
--
You are receiving this mail because:
You are the QA Contact for the bug.
You are on the CC list for the bug.
You are the assignee for the bug.
More information about the Bugs
mailing list