[Gluster-users] A very special announcement from Gluster.org

Philip flips01 at googlemail.com
Sat Jun 2 18:21:38 UTC 2012

I haven't but I will give it a try! Maybe you should also reconsider the
way you are building the debs. Building debs for a stable software on/for a
unstable operating system isn't smart is it?

2012/6/2 Sachidananda URS <surs at redhat.com>

> Hi Philip,
> Did you try installing libssl from source to meet the dependency?
> -sac
> Sent from my iPhone
> On 02-Jun-2012, at 13:57, Philip <flips01 at googlemail.com> wrote:
> It is still not possible to install the 3.3 deb on a stable release of
> debian because squeeze has no libssl1.0.0.
> 2012/5/31 John Mark Walker <johnmark at redhat.com>
>> Today, we’re announcing the next generation of GlusterFS<http://www.gluster.org/>,
>> version 3.3. The release has been a year in the making and marks several
>> firsts: the first post-acquisition release under Red Hat, our first major
>> act as an openly-governed project <http://www.gluster.org/roadmaps/>and
>> our first foray beyond NAS. We’ve also taken our first steps towards
>> merging big data and unstructured data storage, giving users and developers
>> new ways of managing their data scalability challenges.
>> GlusterFS is an open source, fully distributed storage solution for the
>> world’s ever-increasing volume of unstructured data. It is a software-only,
>> highly available, scale-out, centrally managed storage pool that can be
>> backed by POSIX filesystems that support extended attributes, such as
>> Ext3/4, XFS, BTRFS and many more.
>> This release provides many of the most commonly requested features
>> including proactive self-healing, quorum enforcement, and granular locking
>> for self-healing, as well as many additional bug fixes and enhancements.
>> Some of the more noteworthy features include:
>>    - Unified File and Object storage – Blending OpenStack’s Object
>>    Storage API  <http://openstack.org/projects/storage/> with GlusterFS
>>    provides simultaneous read and write access to data as files or as objects.
>>    - HDFS compatibility – Gives Hadoop administrators the ability to run
>>    MapReduce jobs on unstructured data on GlusterFS and access the data with
>>    well-known tools and shell scripts.
>>    - Proactive self-healing – GlusterFS volumes will now automatically
>>    restore file integrity after a replica recovers from failure.
>>    - Granular locking – Allows large files to be accessed even during
>>    self-healing, a feature that is particularly important for VM images.
>>    - Replication improvements – With quorum enforcement you can be
>>    confident that  your data has been written in at least the configured
>>    number of places before the file operation returns, allowing a
>>    user-configurable adjustment to fault tolerance vs performance.
>> *
>> *Visit http://www.gluster.org <http://gluster.org/> to download.
>> Packages are available for most distributions, including Fedora, Debian,
>> RHEL, Ubuntu and CentOS.
>> Get involved! Join us on #gluster on freenode, join our mailing list<http://www.gluster.org/interact/mailinglists/>,
>> ‘like’ our Facebook page <http://facebook.com/GlusterInc>, follow us on
>> Twitter <http://twitter.com/glusterorg>, or check out our LinkedIn group<http://www.linkedin.com/groups?gid=99784>
>> .
>> GlusterFS is an open source project sponsored by Red Hat<http://www.redhat.com/>®,
>> who uses it in its line of Red Hat Storage<http://www.redhat.com/storage/>
>>  products.
>> (this post published at
>> http://www.gluster.org/2012/05/introducing-glusterfs-3-3/ )
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20120602/69ce39db/attachment.html>

More information about the Gluster-users mailing list