[Gluster-users] Re; Strange behaviour with add-brick followed by remove-brick

B.K.Raghuram bkrram at gmail.com
Wed Nov 6 05:23:57 UTC 2013


Here are the steps that I did to reproduce the problem. Essentially,
if you try to remove a brick that is not the same as the localhost
then it seems to migrate the files on the localhost brick instead and
hence there is a lot of data loss.. If instead, I try to remove the
localhost brick, it works fine. Can we try and get this fix into 3.4.2
as this seems to be the only way to replace a brick, given that
replace-brick is being removed!

[root at s5n9 ~]# gluster volume create v1 transport tcp
s5n9.testing.lan:/data/v1 s5n10.testing.lan:/data/v1
volume create: v1: success: please start the volume to access data
[root at s5n9 ~]# gluster volume start v1
volume start: v1: success
[root at s5n9 ~]# gluster volume info v1

Volume Name: v1
Type: Distribute
Volume ID: 6402b139-2957-4d62-810b-b70e6f9ba922
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: s5n9.testing.lan:/data/v1
Brick2: s5n10.testing.lan:/data/v1

***********Now NFS mounted the volume onto my laptop and with a script
created 300 files in the mount. Distribution results below **********
[root at s5n9 ~]# ls -l /data/v1 | wc -l
160
[root at s5n10 ~]# ls -l /data/v1 | wc -l
142

[root at s5n9 ~]# gluster volume add-brick v1 s6n11.testing.lan:/data/v1
volume add-brick: success
[root at s5n9 ~]# gluster volume remove-brick v1 s5n10.testing.lan:/data/v1 start
volume remove-brick start: success
ID: 8f3c37d6-2f24-4418-b75a-751dcb6f2b98
[root at s5n9 ~]# gluster volume remove-brick v1 s5n10.testing.lan:/data/v1 status
                                    Node Rebalanced-files
size       scanned      failures       skipped         status run-time
in secs
                               ---------      -----------
-----------   -----------   -----------   -----------   ------------
--------------
                               localhost                0
0Bytes             0             0    not started             0.00
                       s6n12.testing.lan                0
0Bytes             0             0    not started             0.00
                       s6n11.testing.lan                0
0Bytes             0             0    not started             0.00
                       s5n10.testing.lan                0
0Bytes           300             0      completed             1.00


[root at s5n9 ~]# gluster volume remove-brick v1 s5n10.testing.lan:/data/v1 commit
Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y
volume remove-brick commit: success

[root at s5n9 ~]# gluster volume info v1

Volume Name: v1
Type: Distribute
Volume ID: 6402b139-2957-4d62-810b-b70e6f9ba922
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: s5n9.testing.lan:/data/v1
Brick2: s6n11.testing.lan:/data/v1


[root at s5n9 ~]# ls -l /data/v1 | wc -l
160
[root at s5n10 ~]# ls -l /data/v1 | wc -l
142
[root at s6n11 ~]# ls -l /data/v1 | wc -l
160
[root at s5n9 ~]# ls /data/v1
file10   file110  file131  file144  file156  file173  file19   file206
 file224  file238  file250  file264  file279  file291  file31  file44
file62  file86
file100  file114  file132  file146  file159  file174  file192  file209
 file225  file24   file252  file265  file28   file292  file32  file46
file63  file87
file101  file116  file134  file147  file16   file18   file196  file210
 file228  file240  file254  file266  file281  file293  file37  file47
file66  file9
file102  file12   file135  file148  file161  file181  file198  file212
 file229  file241  file255  file267  file284  file294  file38  file48
file69  file91
file103  file121  file136  file149  file165  file183  file200  file215
 file231  file243  file256  file268  file285  file295  file4   file50
file7   file93
file104  file122  file137  file150  file17   file184  file201  file216
 file233  file245  file258  file271  file286  file296  file40  file53
file71  file97
file105  file124  file138  file152  file170  file186  file202  file218
 file234  file246  file261  file273  file287  file297  file41  file54
file73
file107  file125  file140  file153  file171  file188  file203  file220
 file236  file248  file262  file275  file288  file298  file42  file55
file75
file11   file13   file141  file154  file172  file189  file204  file222
 file237  file25   file263  file278  file290  file3    file43  file58
file80

[root at s6n11 ~]# ls  /data/v1
file10   file110  file131  file144  file156  file173  file19   file206
 file224  file238  file250  file264  file279  file291  file31  file44
file62  file86
file100  file114  file132  file146  file159  file174  file192  file209
 file225  file24   file252  file265  file28   file292  file32  file46
file63  file87
file101  file116  file134  file147  file16   file18   file196  file210
 file228  file240  file254  file266  file281  file293  file37  file47
file66  file9
file102  file12   file135  file148  file161  file181  file198  file212
 file229  file241  file255  file267  file284  file294  file38  file48
file69  file91
file103  file121  file136  file149  file165  file183  file200  file215
 file231  file243  file256  file268  file285  file295  file4   file50
file7   file93
file104  file122  file137  file150  file17   file184  file201  file216
 file233  file245  file258  file271  file286  file296  file40  file53
file71  file97
file105  file124  file138  file152  file170  file186  file202  file218
 file234  file246  file261  file273  file287  file297  file41  file54
file73
file107  file125  file140  file153  file171  file188  file203  file220
 file236  file248  file262  file275  file288  file298  file42  file55
file75
file11   file13   file141  file154  file172  file189  file204  file222
 file237  file25   file263  file278  file290  file3    file43  file58
file80


******* An ls of the mountpoint after this whole process only shows
159 files - the ones that are on s5n9. So everything that was on s5n10
is gone!! ****



More information about the Gluster-users mailing list