Gluster blog stories provide high-level spotlights on our users all over the world
Probably everybody who cares already knows that GlusterFS 3.3 has been released. I find it amusing that quorum enforcement is listed as one of the marquee features, since it’s really such a trivial bit of code compared to the other features on the list, or to internal but significant changes such as the new GFID-based back end, but I guess it’s importance to users rather than effort that define the list. I think it’s also worth mentioning that the release includes a ton of minor fixes that came from running static analysis, and tweaks from the performance team, and other things that are individually small but add up to some pretty major improvements. Maybe some time soon I’ll run some benchmarks of 3.2 vs. 3.3 to show just how dramatic the differences can be. That’s not what I’m here to talk about, though. Now that 3.3 is out the door, a lot of deferred changes have started moving through the queue. Here are some examples of things that are well on their way to becoming part of 3.4, and that I might have mentioned here or in talking to people.
|User-specified DHT layouts that won’t get stomped when you rebalance||already merged|
|Selection of AFR “read child” via hash to avoid hot spots||already merged|
|Server-side resolution of auxiliary GIDs, to support more than 16||already merged|
|Reliable selection of local AFR “read child”||in active review|
|SSL (for I/O path) and socket multi-threading||refreshed and in review|
The other HekaFS features are separately getting a bit unstuck. Kaleb is charging ahead with essential infrastructure for the namespace and ID isolation. Edward is cleaning up the at-rest encryption code for submission. So, what am I going to be doing? I have a few more tweaks in mind around replication and distribution (including those I’m maintaining in my own GitHub trees), but mostly I expect to be working on the infamous …and a pony replication. In my not-humble-at-all opinion, that’s the one feature that will really put some distance between GlusterFS and the also-rans. It’s still the #1 idea for making people’s eyes light up, both in public presentations (come see me at Red Hat Summit!) and in private conversations with customers who have petabytes already . . . and that’s before you even consider its applicability to migrating data into or out of public clouds. After nearly three years of being a good little boy and working on stuff that honestly didn’t interest me nearly as much because the need was more immediate, it looks like I’ll finally be free – encouraged, even – to dive into the project I really came here to do. Expect more here as I work out various details over the coming weeks.
2020 has not been a year we would have been able to predict. With a worldwide pandemic and lives thrown out of gear, as we head into 2021, we are thankful that our community and project continued to receive new developers, users and make small gains. For that and a...
It has been a while since we provided an update to the Gluster community. Across the world various nations, states and localities have put together sets of guidelines around shelter-in-place and quarantine. We request our community members to stay safe, to care for their loved ones, to continue to be...
The initial rounds of conversation around the planning of content for release 8 has helped the project identify one key thing – the need to stagger out features and enhancements over multiple releases. Thus, while release 8 is unlikely to be feature heavy as previous releases, it will be the...