Skip to Main Content
IBM System Storage Ideas Portal


This portal is to open public enhancement requests against IBM System Storage products. To view all of your ideas submitted to IBM, create and manage groups of Ideas, or create an idea explicitly set to be either visible by all (public) or visible only to you and IBM (private), use the IBM Unified Ideas Portal (https://ideas.ibm.com).


Shape the future of IBM!

We invite you to shape the future of IBM, including product roadmaps, by submitting ideas that matter to you the most. Here's how it works:

Search existing ideas

Start by searching and reviewing ideas and requests to enhance a product or service. Take a look at ideas others have posted, and add a comment, vote, or subscribe to updates on them if they matter to you. If you can't find what you are looking for,

Post your ideas
  1. Post an idea.

  2. Get feedback from the IBM team and other customers to refine your idea.

  3. Follow the idea through the IBM Ideas process.


Specific links you will want to bookmark for future use

Welcome to the IBM Ideas Portal (https://www.ibm.com/ideas) - Use this site to find out additional information and details about the IBM Ideas process and statuses.

IBM Unified Ideas Portal (https://ideas.ibm.com) - Use this site to view all of your ideas, create new ideas for any IBM product, or search for ideas across all of IBM.

ideasibm@us.ibm.com - Use this email to suggest enhancements to the Ideas process or request help from IBM for submitting your Ideas.

Status Delivered
Created by Guest
Created on Apr 19, 2021

Extend readReplicaPolicy and "failure group" to nodes, to help localize reads in a stretched cluster.

With a 2-site "stretched cluster", readReplicaPolicy can be used to choose which replica is used for reading, except... our options don't really handle complex network topologies. The "fastest" setting is often not a good choice because network latency is still comparable to HDD seek times.

This RFE proposes that just as disks can be tagged with a failure group, that nodes also be able to be tagged with failure groups. Then also add a new setting for readReplicaPolicy, perhaps "failureGroup", to cause a replica with the same failure group tag as the node to be preferred for reading. (If not available, perhaps fall back to "local".)

This function would extend to multi-cluster, so the failure group of a node in a consuming cluster would prefer the replica with the same failure group in the owning (storage) cluster. This would require some cross-cluster coordination, but there are already other things that require this (for example uid/gid, or the fabric number in verbsPorts).

Potentially the failure group attribute could also be used with non-replicated writes to a file system, to prefer writing locally (that could allow replication to be deferred until a later mmapplypolicy, sparing applications from some write latency at the cost of possibly losing recent work).

Potentially the failure group attribute could also be used to optimize commands that need to scan metadata and the file system, such as mmrestripefs and mmapplypolicy (see also RFE 149675).

Potentially some other attribute, "site", could be added to both disks and nodes - if a site had more than one disk failure group, and we had sufficient replication, this would be useful. But we don't typically see a replication factor of 3, and failure group already seems to capture the important cases, it seems simplest to continue to overload the meaning of "failure group".

Potentially an AFM fileset could have the afmGateway parameter extended to use a failure group for the preferred gateway node - this could optimize network access between application nodes and the gateway (presumably the application is primarily using nodes in the same "failure group" as specified in the fileset).

To really fill out this feature, include system node classes to encompass nodes in a particular failure group.

Idea priority Medium
  • Guest
    Reply
    |
    Sep 1, 2022
    This was delivered in 5.1.4. See https://www.ibm.com/docs/en/spectrum-scale/5.1.4?topic=summary-changes
  • Guest
    Reply
    |
    Mar 13, 2022
    .