I've already explained how no method encompasses all situations nor should they. The basic CI method accurately covers 95% of conditions faced.
Curve shape is a good question. This is the argument made in the Contrast Index paper. Gamma only uses the straight-line portion of the curve, which is not only difficult to determine in some cases, but also doesn't entirely represent the portion of the curve in usage. The paper doesn't cover shoulders though. The Theory of the Photographic Process does note that "the gradient of the characteristic curve is of such importance, especially in connection with the problems of tone reproduction, that it is often desirable to use a derivative curve...This graphic form is useful when it is desired to determine precisely the exposure value corresponding to some particular slope of the D-log E curve...for many purposes it presents the data in more convenient form and gives a more vivid mental picture of the relation between gradient and exposure."
Should variations in the upper end affect the overall average? According to tone reproduction theory, the perception of a quality print allows for compression of the shadows and highlights, but requires the midtones to be over a gradient of ~1.11 of that of the original subject. Tone reproduction theory also states that there isn't a perfect correlation between the negative density range and paper LER. According to Loyd Jones, "for the soft papers, the density scales of the negative (NDR) should in most cases exceed the sensitometric exposure scale of the paper (LER) (see Tone Reproduction Curve attachment). According to Theory of the Photographic Process, "For scenes having unusually long log luminance ranges (2.5 to 3.0), the tone reproduction curves for the first-choice prints were slightly to the left of the curve shown (in the example)...The gradient in the middletone region was always greater than 1.00 (usually 1.10 - 1.20) for the preferred prints of all scenes studied." So, there are no absolutes. A perfect match isn't necessary. Average gradient is a tool to describe how the subject is generally translated into density.
If you measure the entire range from the example, you get (top to bottom) 0.51, 0.41, 0.39. Which in a fixed flare developmental model is from -1 to -3. However, if you measure only up to where the two bottom curves begin to shoulder off at Zone X, you get 0.48, 0.47, 0.45, or -1 1/2 to -2. Then there is the question of flare. Flare for an average scene is around 1 to 1 1/3 stops. Flare's rule of thumb is that flare increases by 1/3 stop per stop increase in the luminance range. A fourteen stop luminance range could have around 3 stops of flare. This would mean measuring the density range of the curves at Zone XI instead of Zone XIV. Which one would best translate into a quality print? Good question. I don't test for such extremes. I don't even know if there is a paper that has a LER of 2.14 from the Acros curve. If there is, what would happen to the all important midtone contrast? I've attached a four quadrant reproduction curve with a ten stop luminance range and two stops flare. The film has a CI 0.43 and the NDR closely matches the paper's LER. Also, there's an example with a slightly higher paper grade. With the higher contrast, the midtone reproduction gradient has just reached 1.00.
These are all interesting questions. In extreme situations, does measuring the entire effective log-H range produce more desirable prints or does limiting the range of measurement in some way work best?
Whatever the answer, it doesn't negate the effectiveness of any of the average gradient methods in most situations generally encountered.