[Gluster-users] Gluster swift

Andrew Holway a.holway at syseleven.de
Thu Nov 29 20:03:35 UTC 2012


Hi,

glusterfs.x86_64                        3.3.1-3.el6                    @epel-glusterfs
glusterfs-fuse.x86_64                   3.3.1-3.el6                    @epel-glusterfs
glusterfs-server.x86_64                 3.3.1-3.el6                    @epel-glusterfs
glusterfs-swift.noarch                  3.3.1-3.el6                    @epel-glusterfs-swift
glusterfs-swift-account.noarch          3.3.1-3.el6                    @epel-glusterfs-swift
glusterfs-swift-container.noarch        3.3.1-3.el6                    @epel-glusterfs-swift
glusterfs-swift-object.noarch           3.3.1-3.el6                    @epel-glusterfs-swift
glusterfs-swift-plugin.noarch           3.3.1-3.el6                    @epel-glusterfs-swift
glusterfs-swift-proxy.noarch            3.3.1-3.el6                    @epel-glusterfs-swift
glusterfs.i686                          3.2.7-1.el6                    epel     
glusterfs-debuginfo.x86_64              3.3.1-3.el6                    epel-glusterfs
glusterfs-devel.i686                    3.2.7-1.el6                    epel     
glusterfs-devel.x86_64                  3.3.1-3.el6                    epel-glusterfs
glusterfs-geo-replication.x86_64        3.3.1-3.el6                    epel-glusterfs
glusterfs-rdma.x86_64                   3.3.1-3.el6                    epel-glusterfs
glusterfs-swift-doc.noarch              3.3.1-3.el6                    epel-glusterfs-swift
glusterfs-vim.x86_64                    3.2.7-1.el6                    epel     

Ta,

Andrew

On Nov 29, 2012, at 8:52 PM, Peter Portante wrote:

> Hi Andrew,
> 
> What version of Gluster are you using?
> 
> -peter
> 
> 
> ----- Original Message -----
>> From: "Andrew Holway" <a.holway at syseleven.de>
>> To: gluster-users at gluster.org
>> Cc: kaleb at keithley.org
>> Sent: Thursday, November 29, 2012 1:02:15 PM
>> Subject: Re: [Gluster-users] Gluster swift
>> 
>> In addition,
>> 
>> Requests to view the contents of containers that have been filled in
>> this manner fail.
>> 
>> [root at cheese25 free]# ls -l | wc
>>   3654   32879  193695
>> [root at cheese25 free]#
>> 
>> [root at bright60 lots_of_little_files]# curl --verbose -H
>> 'X-Auth-Token: AUTH_tk289d8ebe3ff44c97a9721970a4251f02'
>> https://cheese25:443/v1/AUTH_gv0/free/ -k
>> * About to connect() to cheese25 port 443 (#0)
>> *   Trying 10.141.105.25... connected
>> * Connected to cheese25 (10.141.105.25) port 443 (#0)
>> * Initializing NSS with certpath: sql:/etc/pki/nssdb
>> * warning: ignoring value of ssl.verifyhost
>> * skipping SSL peer certificate verification
>> * SSL connection using TLS_RSA_WITH_AES_256_CBC_SHA
>> * Server certificate:
>> * 	subject: CN=cheese25,O=Default Company Ltd,L=Default City,C=XX
>> * 	start date: Nov 29 16:27:49 2012 GMT
>> * 	expire date: Dec 29 16:27:49 2012 GMT
>> * 	common name: cheese25
>> * 	issuer: CN=cheese25,O=Default Company Ltd,L=Default City,C=XX
>>> GET /v1/AUTH_gv0/free/ HTTP/1.1
>>> User-Agent: curl/7.19.7 (x86_64-redhat-linux-gnu) libcurl/7.19.7
>>> NSS/3.13.1.0 zlib/1.2.3 libidn/1.18 libssh2/1.2.2
>>> Host: cheese25
>>> Accept: */*
>>> X-Auth-Token: AUTH_tk289d8ebe3ff44c97a9721970a4251f02
>>> 
>> < HTTP/1.1 503 Internal Server Error
>> < Content-Type: text/html; charset=UTF-8
>> < Content-Length: 0
>> < Date: Thu, 29 Nov 2012 18:00:04 GMT
>> <
>> * Connection #0 to host cheese25 left intact
>> * Closing connection #0
>> 
>> But, non full volumes are ok :)
>> 
>> [root at bright60 lots_of_little_files]# curl --verbose -H
>> 'X-Auth-Token: AUTH_tk289d8ebe3ff44c97a9721970a4251f02'
>> https://cheese25:443/v1/AUTH_gv0/stuff/ -k
>> * About to connect() to cheese25 port 443 (#0)
>> *   Trying 10.141.105.25... connected
>> * Connected to cheese25 (10.141.105.25) port 443 (#0)
>> * Initializing NSS with certpath: sql:/etc/pki/nssdb
>> * warning: ignoring value of ssl.verifyhost
>> * skipping SSL peer certificate verification
>> * SSL connection using TLS_RSA_WITH_AES_256_CBC_SHA
>> * Server certificate:
>> * 	subject: CN=cheese25,O=Default Company Ltd,L=Default City,C=XX
>> * 	start date: Nov 29 16:27:49 2012 GMT
>> * 	expire date: Dec 29 16:27:49 2012 GMT
>> * 	common name: cheese25
>> * 	issuer: CN=cheese25,O=Default Company Ltd,L=Default City,C=XX
>>> GET /v1/AUTH_gv0/stuff/ HTTP/1.1
>>> User-Agent: curl/7.19.7 (x86_64-redhat-linux-gnu) libcurl/7.19.7
>>> NSS/3.13.1.0 zlib/1.2.3 libidn/1.18 libssh2/1.2.2
>>> Host: cheese25
>>> Accept: */*
>>> X-Auth-Token: AUTH_tk289d8ebe3ff44c97a9721970a4251f02
>>> 
>> < HTTP/1.1 200 OK
>> < X-Container-Object-Count: 15
>> < X-Container-Bytes-Used: 0
>> < Accept-Ranges: bytes
>> < Content-Length: 84
>> < Content-Type: text/plain; charset=utf-8
>> < Date: Thu, 29 Nov 2012 18:00:49 GMT
>> <
>> a1
>> a10
>> a100
>> a1000
>> a10000
>> a1001
>> a1002
>> a1003
>> a1004
>> a1005
>> a1006
>> a1007
>> a1008
>> a1009
>> a101
>> * Connection #0 to host cheese25 left intact
>> * Closing connection #0
>> 
>> Thanks,
>> 
>> Andrew
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> On Nov 29, 2012, at 6:38 PM, Andrew Holway wrote:
>> 
>>> Hi,
>>> 
>>> Im mooperd on irc.
>>> 
>>> After lots of swearing and learning I think I am getting the hang
>>> of it. Today, I created 10000 files and squirted them into UFO.
>>> 
>>> for file in $(ls); do curl -X PUT -T $file -H 'X-Auth-Token:
>>> AUTH_tk289d8ebe3ff44c97a9721970a4251f02'
>>> https://cheese25:443/v1/AUTH_gv0/new_container/ -k; done
>>> 
>>> It works perfectly until you get to about 3500 files....then.
>>> 
>>> </body>
>>> </html><html>
>>> <head>
>>> <title>201 Created</title>
>>> </head>
>>> <body>
>>> <h1>201 Created</h1>
>>> <br /><br />
>>> 
>>> 
>>> 
>>> </body>
>>> </html><html>
>>> <head>
>>> <title>201 Created</title>
>>> </head>
>>> <body>
>>> <h1>201 Created</h1>
>>> <br /><br />
>>> 
>>> 
>>> 
>>> </body>
>>> </html><html>
>>> <head>
>>> <title>404 Not Found</title>
>>> </head>
>>> <body>
>>> <h1>404 Not Found</h1>
>>> The resource could not be found.<br /><br />
>>> 
>>> 
>>> This seems to be a hard limit for the number of files in a dir. Any
>>> ideas?
>>> 
>>> Thanks,
>>> 
>>> Andrew
>>> 
>>> 
>>> 
>>> 
>>> 
>> 
>> 
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>> 





More information about the Gluster-users mailing list