AUSTRALIAN COMPETITION TRIBUNAL
IN THE AUSTRALIAN COMPETITION TRIBUNAL
ACT 4 of 2015
APPLICATIONS UNDER S 71B OF THE NATIONAL ELECTRICITY LAW FOR A REVIEW OF DISTRIBUTION DETERMINATION MADE BY THE AUSTRALIAN ENERGY REGULATOR IN RELATION TO AUSGRID PURSUANT TO RULE 6.11.1 OF THE NATIONAL ELECTRICITY RULES
MR R DAVEY, MEMBER
DR D ABRAHAM, MEMBER
DATE OF DETERMINATION:
DARWIN (VIA VIDEO LINK TO SYDNEY, MELBOURNE, BRISBANE AND ADELAIDE)
THE TRIBUNAL DETERMINES THAT:
1. Pursuant to s 71P(2)(c) of the National Electricity Law, the Final Decision Ausgrid distribution determination 2015-16 to 2018-19, April 2015, including attachments, (the Final Decision) is set aside and remitted to the Australian Energy Regulator (AER) to make the decision again in accordance with the following directions:
(a) the AER is to make the constituent decision on opex under r 6.12.1(4) of the National Electricity Rules in accordance with these reasons for decision including assessing whether the forecast opex proposed by the applicant reasonably reflects each of the operating expenditure criteria in r 6.5.6(c) of the National Electricity Rules including using a broader range of modelling, and benchmarking against Australian businesses, and including a “bottom up” review of Ausgrid’s forecast operating expenditure;
(b) the AER is to make the constituent decision on return on debt in relation to the introduction of the trailing average approach in accordance with these reasons for decision;
(c) the AER is to make the constituent decision on estimated cost of corporate income tax (gamma) in accordance with these reasons for decision, including by reference to an estimated cost of corporate income tax based on a gamma of 0.25; and
(d) the AER is to consider, and to the extent to which it considers appropriate to vary the Final Decision in such other respects as the Australian Energy Regulator considers appropriate having regard to s 16(1)(d) of the National Electricity Law in the light of such variations as are made to the Final Decision by reason of (a)-(c) hereof.
ACT 1 of 2015
APPLICATIONS UNDER S 71B OF THE NATIONAL ELECTRICITY LAW FOR A REVIEW OF DISTRIBUTION DETERMINATION MADE BY THE AUSTRALIAN ENERGY REGULATOR IN RELATION TO PUBLIC INTEREST ADVOCACY CENTRE LTD PURSUANT TO RULE 6.11.1 OF THE NATIONAL ELECTRICITY RULES
PUBLIC INTEREST ADVOCACY CENTRE LTD
ACT 4 of 2015
APPLICATION UNDER S 71B OF THE NATIONAL ELECTRICITY LAW FOR A REVIEW OF DISTRIBUTION DETERMINATION MADE BY THE AUSTRALIAN ENERGY REGULATOR IN RELATION TO AUSGRID PURSUANT TO RULE 6.11.1 OF THE NATIONAL ELECTRICITY RULES
MANSFIELD J, PRESIDENT
MR R DAVEY, MEMBER
DR D ABRAHAM, MEMBER
26 february 2016
DARWIN (via video link to sydney, MELBOURNE, BRISBANE AND ADELAIDE)
REASONS FOR DETERMINATION
1 These reasons relate to two of seven applications made to review decisions of the Australian Energy Regulator (AER) made on 30 April 2015 under the National Electricity Law (NEL). The issues relating to those decisions, and to a further decision of the AER made on 3 June 2015 under the National Gas Law (NGL) have been heard together.
2 That is because a number of the issues arising in relation to these two applications are common to issues in relation to the six other review applications referred to below, although there are a few issues particular to one or more of the applications, and in some respects the matters raised on certain issues differed slightly. It was common ground that the substantial commonality of issues raised in the eight applications made it preferable for them to be heard together.
3 These reasons deal only with the applications to review the AER decision concerning Ausgrid. They will serve as the “lead” reasons insofar as the Tribunal’s general considerations, on the significant matters of common concern, and its consideration of aspects of particular topics may not need to be repeated in full by the Tribunal in its consideration of the other applications.
4 All the applications were made in respect of a regulatory decision-making process by the AER and a regulatory review process by the Tribunal which were each significantly different from the regulatory decision-making process and the review process previously existing. The parties understandably spent considerable time addressing those differences and their significance. These reasons contain the Tribunal’s consideration of those submissions. Where appropriate that consideration will be incorporated by reference into its determinations in relation to the other applications, rather than be repeated.
5 On 30 April 2015, the AER published a distribution determination final decision under r 6.11.1 of Chapter 6 of the National Electricity Rules (NER) in relation to each of Ausgrid, Endeavour Energy (Endeavour) and Essential Energy (Essential). Each of the final decisions include an overview and attachments, the overviews being entitled: Final Decision Ausgrid distribution determination 2015-16 to 2018-19, Overview, April 2015; Final Decision Endeavour Energy distribution determination 2015-16 to 2018-19, Overview, April 2015; and Final Decision Essential Energy distribution determination 2015-16 to 2018-19, Overview, April 2015. Ausgrid, Endeavour and Essential are State owned corporations incorporated under the Energy Services Corporations Act 1995 (NSW) and are referred to collectively as Networks NSW. Each is the owner and operator of a monopoly electricity distribution network in New South Wales.
6 On the same day, the AER published a distribution determination final decision including an overview and attachments under the same provision in relation to ActewAGL Distribution (ActewAGL): Final Decision ActewAGL distribution determination 2015-16 to 2018-19 Overview, April 2015. ActewAGL is the owner of the electricity distribution network in the Australian Capital Territory.
7 The AER determined as follows:
(1) Ausgrid can recover $6576.4m ($ nominal) from its customers over the 2015-2019 regulatory control period.
(2) Endeavour can recover $3182.8m ($ nominal) from its customers over the 2015-2019 regulatory control period.
(3) Essential can recover $3826.1m ($ nominal) from its customers over the 2015-2019 regulatory control period.
(4) ActewAGL can recover $590.9m ($ nominal) from its customers over the 2015-2019 regulatory control period.
Each of those entities may be referred to as a Distribution Network Service Provider (DNSP).
8 Distribution charges represent a significant component of the annual bill for customers of the DNSPs and so for consumers of electricity in their respective distribution areas. The AER estimated that its decisions would (noting that it was providing estimates only and that there are other factors that will affect a consumer’s electricity bill, such as the wholesale price of electricity) have the following estimated impact:
( ) Ausgrid: For residential customers, a reduction in their average annual electricity bills by $165 (or 8 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period. For small business customers, a reduction in their average annual electricity bills by $264 (or 8 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period.
(a) Endeavour: For residential customers, a reduction in their average annual electricity bills by $106 (or 5.3 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period. For small business customers, a reduction in their average annual electricity bills by $152 (5.3 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period.
(b) Essential: For residential customers, a reduction in their average annual electricity bills by $313 (or 11.9 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period. For small business customers, a reduction in their average annual electricity bills by $528 (or 11.9 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period.
(c) ActewAGL: For residential customers, a reduction in their average annual electricity bills by $112 (or 5.8 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period. For small business customers, a reduction in their average annual electricity bills by $168 (or 5.8 percent) in 2015-16 and remaining relatively stable over the rest of the regulatory control period.
9 On 3 June 2015, the AER published a full access arrangement final decision in relation to Jemena Gas Networks (NSW) Ltd (JGN), pursuant to rr 62 and 64 of the National Gas Rules (NGR): Final Decision – Jemena Gas Networks (NSW) Ltd Access Arrangement 2015-20 Overview, June 2015, which, like the final distribution decisions mentioned above, includes a number of attachments.
10 By that decision, the AER determined that JGN could recover $2,229.0m ($ nominal) from its customers over five years commencing 1 July 2015.
11 As with the DNSPs, the revenue that the AER determined affects the distribution component of a consumer’s final gas bill. For residential and small business customers, the AER estimated that:
(a) average annual gas bills would fall by around 9.2 percent in 2015-16, translating into a $96 reduction in bills for residential customers and a $462 reduction for small business customers;
(b) bills will continue to fall over the following three years; and
(c) there would be a small increase of 1 percent in 2019-20.
12 Each of the distribution determinations that the AER made on 30 April 2015 is a “reviewable regulatory decision” within the meaning of subcl (a) of the definition of that term in s 71A of the NEL. The AER’s access arrangement decision in relation to JGN is also a “reviewable regulatory decision” within the meaning of subcl (d) of the definition in s 244 of the NGL. It is convenient to refer to each of those five decisions as a Final Decision.
13 The term “Final Decision” is used to distinguish between the separate decision-making stages specified in the Rules. Under the NEL and NER a network service provider is required to submit to the AER a regulatory proposal that the AER must consider. The AER must then publish a draft decision, allowing the network service provider the opportunity to submit a revised regulatory proposal that the AER must consider. The AER must then publish a draft decision, allowing the network service provider the opportunity to submit a revised regulatory proposal prior to a Final Decision being made. A like process must be followed under the NGL and NGR by a service provider submitting, and the AER considering, an access arrangement revision proposal prior to a Final Decision being made. In respect of each of Networks NSW, ActewAGL and JGN, it is convenient to refer to each of those proposals and draft decisions as a “Regulatory Proposal”, “Draft Decision” or “Revised Regulatory Proposal” respectively.
14 By applications made on 21 May 2015 pursuant to s 71B of NEL, ActewAGL and Networks NSW applied to the Tribunal for leave to review the AER’s respective Final Decisions concerning them. The Public Interest Advocacy Centre Ltd (PIAC) also applied for leave to review each of the AER’s three Final Decisions relating to Networks NSW. On 24 June 2015, JGN filed an application for leave to review the AER’s Final Decision concerning it.
15 It is convenient hereafter to refer collectively to Networks NSW, ActewAGL and JGN as the Network Applicants from time to time.
16 On 17 July 2015, the Tribunal granted leave in respect of each of the applications relating to the four Final Decisions made on 30 April 2015: Applications by Public Interest Advocacy Centre Ltd, Ausgrid, Endeavour Energy and Essential Energy  ACompT 2 and Application by ActewAGL Distribution  ACompT 3. On 4 August 2015, the Tribunal granted leave in relation to the Final Decision relating to JGN: Application by Jemena Gas Networks (NSW) Limited  ACompT 4.
17 In addition to their applications for review, PIAC and Networks NSW sought and were each granted leave to intervene in the other respective review applications concerning the AER’s Final Decisions relating to Networks NSW. There were other interveners granted leave to appear before the Tribunal in the review applications generally, clearly in the case of other DNSPs because regulatory review determinations were in the process of being made by the AER. A list of the interveners is:
PIAC (in relation to the Networks NSW applications);
Networks NSW (in relation to the PIAC applications);
AusNet Services (Distribution) Pty Ltd, AusNet Services (Transmission) Ltd, Australian Gas Networks Ltd, Citipower Pty Ltd, Powercor Australia Ltd, SA Power Networks and United Energy Distribution Pty Ltd (Vic/SA Network Interveners);
Ergon Energy Corporation Ltd (Ergon); and
the Commonwealth Minister for Industry and Science. On 21 September 2015, the title of the Commonwealth Minister with responsibility for energy matters changed from the Minister for Industry and Science to the Minister for Resources, Energy and Northern Australia. For consistency, hereafter he is referred to as the Minister.
The Minister’s intervention was confined to making submissions on the proper construction and application of the relevant provisions of the NEL, the NER, the NGL and the NGR. Although the Ministers of participating jurisdictions are entitled to intervene, and were notified by the Tribunal of the several applications for review of the Final Decisions (once leave to apply for review had been given), none of the relevant Ministers did intervene.
18 The national electricity market was established following the enactment of the National Electricity (South Australia) Act 1996 (SA), in broad terms providing for the competitive trading and regulation of the generation, transmission, distribution and supply of electricity (then) in south-eastern Australia. The national electricity market was to be a competitive wholesale market comprising a comprehensive set of trading arrangements. The initial version of the NEL was a schedule to that Act.
19 It is clear enough, from that legislation, its context, and the events preceding it, that it was a consequence of the competition policy reforms and the pro-competitive policy mindset following the Hilmer reforms in the early 1990s. It was designed, wherever possible, to introduce competition into the provision of electricity to consumers through structural reform, by ensuring that government enterprises competed in an appropriate form, and in the case of monopoly infrastructure to provide for regulated access to monopoly infrastructure with independent authorities to oversee prices. That Act was amended by the National Electricity (South Australia) (New National Electricity Law) Amendment Act 2005 (SA), substituting the current version (since amended) of the NEL, and providing for the conferral of functions and powers in respect of the national electricity governance on the Australian Energy Market Commission (AEMC) (established under the Australian Energy Market Commission Establishment Act 2004 (Cth)) and upon the AER (established under the Trade Practices Act 1974 (Cth) (now the Competition and Consumer Act 2010 (Cth)). It was at the time contemplated, and it has now been effected, that the AER would operate as the economic regulator of both electricity and gas transmission and distribution networks for all jurisdictions other than Western Australia. The previous State based regulatory structure was removed.
20 The National Electricity (South Australia) National Electricity Law (Miscellaneous Amendments) Amendment Act 2007 (SA) really introduced the present structure under which, subject to amendments to the NER and to the NGR by the AEMC in 2012 and to legislative amendments to the NEL and the NGL in 2013, the AER and the Tribunal are conducting their present functions.
21 The Act referred to in the preceding paragraph set out to establish a single national regulatory framework for electricity networks, and introduced important changes to the AER’s powers, including the prescription of the national electricity objective (the NEO), and the revenue and pricing principles (RPP), to guide the AER in making regulatory decisions, and in other respects. It also introduced new merits review provisions under which the Tribunal (also a creation of the then Trade Practices Act 1974) was given the responsibility of reviewing certain decisions of the AER.
22 It was said in the Second Reading Speech by the Minister for Mineral Resources and Development (South Australia) that the proposed Act would lower the barriers to competition: Legislative Council, South Australia, 16 October 2007 Hansard p 883. It was specifically noted (Hansard p 886) that the NEO did not extend to “broader social and environmental objectives”. The Minister’s Second Reading Speech said:
The purpose of the National Electricity Law is to establish a framework to ensure the efficient operation of the national electricity market, efficient investment in, and the effective regulation of electricity networks. As previously noted, the national electricity objective also guides the Australian Energy Market Commission and the Australian Energy Regulator in performing their functions. This should be guided by an objective of efficiency that is in the long term interests of consumers. Environmental and social objectives are better dealt with in other legislative instruments and policies which sit outside the National Electricity Law.
23 It was the same legislation which introduced the RPP, which, it was said (p 887) are fundamental to ensuring achievement of the intention of enhancing efficiency in the national electricity market. The principles were said to maintain a framework for an efficient network investment irrespective of the evolution of the regulatory regime (by changes to the NER). The same was proposed for the new NGL and the national gas objective (NGO).
24 The AEMC from July 2005 was responsible for developing the NER. Similarly, it later became responsible for developing the NGR.
25 As noted, that legislation introduced also the limited merits review by the Tribunal of certain regulatory decisions under the NEL (and contemplated under the NGL). It was intended that network service providers, users and consumer associations could seek review of primary transmission and distribution determinations made by the AER. The review was confined to specified grounds of review, but more importantly there were two elements limiting the scope of review: the review could only address issues which were raised and addressed previously to the AER, and secondly the review could only address those issues on the material presented to the AER. There was to be no review of an AER decision on arguments not made to, or on material not presented to, the AER.
26 At about the same time, extensive amendments to Chapter 6 of the NER were introduced by the AEMC to guide the AER.
27 The National Electricity (South Australia) (National Electricity Law – Australian Energy Market Operator) Amendment Act 2009 (SA) then effected further amendments by which the Australian Energy Market Operator (AEMO) came to be the national energy market operator for both the electricity and gas markets, so as to reinforce the national character of energy market governance. It is not necessary to discuss in detail the effect of those amendments.
28 It is within that structure that, for present purposes, it can fairly be said that the first regulatory cycle (2008-13) of decision making by the AER became the subject of determinations for transmission and distribution networks under the NEL, and under the NGL, by the AER. Given the complexity of the task, it is hardly surprising that the Standing Committee on Energy and Resources (SCER) and the AEMC determined that significant reforms to the provisions of the NEL and the NER (and in turn the NGL and the NGR) should be introduced in the light of that initial experience.
29 On 15 November 2012, the AEMC published its Final Position Paper National Electricity Amendment (Economic Regulation of Network Service Providers) Rule 2012 National Gas Amendment (Price and Regulation of Gas Services) Rule 2012 (the Final Postiion Paper). On 29 November 2012, the AEMC published its Rule Determination National Electricity Amendment (Economic Regulation of Network Service Providers) Rule 2012 National Gas Amendment (Price and Revenue Regulation of Gas Services) Rule (the 2012 Rule Amendments). It will be necessary to refer to those particular amendments in the course of considering particular matters raised on this and the related applications. It should be noted, however, that the 2012 Rule Amendments required the AER to publish guidelines specifying the approach that the AER proposed to use to assess forecasts of operating expenditure (opex) and capital expenditure (capex), and to set out the methodologies that it proposed to use in estimating the allowed rate of return, and the methods, financial models, market data and other evidence that it proposed to take into account in estimating the return on equity and the return on debt. This requirement resulted in the AER’s Better Regulation Rate of Return Guideline, December 2013, (the RoR Guideline) and its Better Regulation Rate of Return Guideline Explanatory Statement, December 2013 (the RoR Explanatory Statement). Significant amendments were made to r 6.5.2 of the NER which are discussed in relation to the topics of return on debt and return on equity which were each the subject of matters of debate in the course of this application and the related applications. Similarly significant amendments were made to r 6.5.6 in relation to how the AER should address the opex factors and capex factors.
30 To allow for a transition to the new rules, the Savings and Transitional Rules in Division 2 of Part ZW of Chapter 11 of the NER provided for a two stage process for the regulation of ACT and NSW DNSPs over the five year period commencing on 1 July 2014 (the 2014-19 period), comprising:
(a) the transitional regulatory control period from 1 July 2014 and ending on 30 June 2015 (transitional regulatory control period); and
(b) the subsequent regulatory control period from 1 July 2015 and ending on 30June 2019 (subsequent regulatory control period).
31 In addition, the Statutes Amendment (National Electricity and Gas Laws – Limited Merits Review) Act 2013 (SA) (the 2013 Legislative Amendments) refined the responsibility of the AER under s 16(1)(b) of the NEL and amended the Limited Merits Review Regime in Pt 6, Div 3A, of the NEL by facilitating the participation of “reviewable regulatory decision process participants”. It also, in a complementary way, added s 16(1)(c) and 16(1)(d) to the NEL to ensure the AER had a focus on the NEO and the NGO and s 71P (and related provisions) modifying the Tribunal’s power to vary or set aside a determination to circumstances where a substituted decision would, or would be likely to, better serve the NEO or the NGO. The detailed nature of those amendments is set out below so far as they are relevant.
32 As observed above, during 2013, the AER in response to those changes adopted its RoR Guideline which it described in some detail in its Final Decision in relation to Ausgrid (see the Overview at pp 55-56) and in its other Final Decisions. It consulted widely with stakeholders to develop a number of guidelines as required. The guidelines included the Expenditure Forecast Assessment Guideline (EFA Guideline), concerning a networks forecast opex proposal, and the RoR Guideline referred to above. The process of producing the Better Regulation guidelines involved active consultation with interested entities, including with PIAC which (as noted above) has itself sought to review the three Networks NSW decisions of the AER.
33 The NEL was substantially amended by the 2013 Legislative Amendments. Relevant to the role of the AER, apart from having a definition of “constituent components” in s 2(1) of the NEL, for present purposes the significant alteration was made by the substitution of s 16(1)(b) and the addition of s 16(1)(c) and, perhaps more importantly, the addition of s 16(1)(d).
34 Section 16(1)(b) requires extensive consultation and notification obligations to be fulfilled by the AER for the purposes of interested persons (including network service users or prospective users of the relevant services, and user or consumer associations that have an interest in the determination) being given an opportunity to address the issues being considered by the AER, and s 16(1)(c) requires the AER to specify in its decision the manner in which the constituent components of the decision relate to each other and how that inter-relationship has been taken into account.
35 Section 16(1)(d) provides that the AER must:
if the AER is making a reviewable regulatory decision and there are 2 or more possible reviewable regulatory decisions that will or are likely to contribute to the achievement of the national electricity objective –
(i) make the decision that the AER is satisfied will or is likely to contribute to the achievement of the national electricity objective to the greatest degree (the preferable reviewable regulatory decision); and
(ii) specify reasons as to the basis on which the AER is satisfied that the decision is the preferable reviewable regulatory decision.
36 That obligation on the AER in exercising its economic regulatory power was of particular relevance to the three applications by PIAC.
37 Effectively, like obligations were prescribed under the NGL by the deletion and substitution of s 28(1) of the NGL.
38 In relation to the Tribunal, and merits review under Div 3A of the NEL, there were also extensive and presently relevant amendments. Section 71A was amended by inserting some additional definitions (to which it will be appropriate to refer as necessary having regard to the substantive amendments).
39 Section 71C, specifying the grounds for review, did not change.
40 However, the additional recognition of the need for the regulatory decision, whether by the AER or on review by the Tribunal, to reflect the “materially preferable NEO decision” (as defined in s 71P(2a)), and the desirability of ensuring that the long term of interests of consumers are properly identified and addressed through the consumer groups was reinforced. It is worth repeating s 71C(1) setting out the (unchanged) grounds of review, and reciting s 71C(1a). They provide:
(1) An application under s 71B(1) may be made only on 1 or more of the following grounds:
(a) the AER made an error of fact in its finding of facts, and that error of fact was material to the making of the decision;
(b) the AER made more than 1 error of fact in its findings of facts, and that those errors of fact, in combination, were material to the making of the decision;
(c) the exercise of the AER’s discretion was incorrect, having regard to all the circumstances;
(d) the AER’s decision was unreasonable, having regard to all the circumstances.
(1a) An application under section 71B(1) must also specify the manner in which a determination made by the Tribunal varying the reviewable regulatory decision, or setting aside the reviewable regulatory decision and a fresh decision being made by the AER following remission of the matter to the AER by the Tribunal, on the basis of 1 or more grounds raised in the application, either separately or collectively, would, or would be likely to, result in a materially preferable NEO decision.
41 Section 71E prescribes the circumstances in which the Tribunal must not grant leave to apply for review. Those circumstances were extended by adding the additional criterion that it must appear to the Tribunal that the applicant for review has established a prima facie case that any decision or determination by the Tribunal varying the reviewable regulatory decision or setting it aside and remitting the matter back to the AER to make the decision again, on the basis of one or more grounds of review raised in the application:
... either separately or collectively, would, or would be likely to, result in a materially preferable NEO decision.
42 Section 71M was amended by adding s 71M(1a) requiring an intervener who raises a new ground of review to provide appropriate particulars of that ground, including how, if accepted or made out, it would, or would be likely to, result in a materially preferable NEO decision.
43 Section 71O was deleted and substituted in its entirety. It deals with the matters that may or not be raised in a review. It firstly ensures that the AER is not confined in a review application so as to prevent it from raising issues which might be considered under s 71P(2a) and (2b). It provides that the applicant, if it is a regulated network service provider, may only raise for review matters that have been raised and maintained in submissions to the AER. The same restriction applies to a regulated network service provider whose commercial interests might be materially affected by that decision. Any other affected or interested person or body may not raise in relation to an issue a matter that was not raised by that body in submissions to the AER (the requirement of having been raised and maintained is not precisely adopted).
44 Section 71O(2)(d) provides that, subject to those restrictions, the applicant or an intervener who has raised a new ground of review under s 71M is also entitled to raise any matter relevant to the issues to be considered under s 71P(2a) and (2b) (set out below) and otherwise any person or body may not raise any matter relevant to those issues unless it is in response to a matter raised by the AER, the applicant or an intervener.
45 The obligation of the Tribunal to make a determination under s 71P has also been substantially refined. It is necessary to set out ss 71P(2a) and (2b) in full to understand their significance. Those provisions in broad terms mirror the obligations imposed upon the AER by s 16(1)(d), and there are parallel obligations upon the Tribunal as upon the AER imposed by s 71P(2c) requiring the Tribunal to explain how it has taken into account the inter-relationship between constituent components of the reviewable regulatory decision and why it has proceeded to make the order which it has determined to make in that light.
46 Section 71P(1)-(2b) provides:
(1) If, following an application, the Tribunal grants leave in accordance with section 71B(1), the Tribunal must make a determination in respect of the application.
See section 71Q for the time limit which the Tribunal must make its determination.
(2) Subject to subsection (2a), a determination under this section may –
(a) affirm the reviewable regulatory decision; or
(b) vary the reviewable regulatory decision; or
(c) set aside the reviewable regulatory decision and remit the matter back to the AER to make the decision again in accordance with any direction or recommendation of the Tribunal.
(2a) Despite subsection (2), the Tribunal may only make a determination –
(a) to vary the reviewable regulatory decision under subsection (2)(b); or
(b) to set aside the reviewable regulatory decision and remit the matter back to the AER under subsection (2)(c).
(c) the Tribunal is satisfied that to do so will, or is likely to, result in a decision that is materially preferable to the reviewable regulatory decision in making a contribution to the achievement of the national electricity objective (a materially preferable NEO decision) (and if the Tribunal is not so satisfied the Tribunal must affirm the decision); and
(d) in the case of a determination to vary the reviewable regulatory decision – the Tribunal is satisfied that to do so will not require the Tribunal to undertake an assessment of such complexity that the preferable course of action would be to set aside the reviewable regulatory decision and remit the matter to the AER to make the decision again.
(2b) In connection with the operation of subsection (2a) (and without limiting any other matter that may be relevant under this Law) –
(a) the Tribunal must consider how the constituent components of the reviewable regulatory decision interrelate with each other and with the matters raised as a ground for review; and
(b) without limiting paragraph (a), the Tribunal must take into account the revenue and pricing principles (in the same manner in which the AER is to take into account these principles under section 16); and
(c) the Tribunal must, in assessing the extent of contribution to the achievement of the national electricity objective, consider the reviewable regulatory decision as a whole; and
(d) the following matters must not, in themselves, determine the question about whether a materially preferable NEO decision exists:
(i) the establishment of a ground for review under section 71C(1);
(ii) consequences for, or impacts on, the average annual regulated revenue of a regulated network service provider;
(iii) that the amount that is specified in or derived from the reviewable regulatory decision exceeds the amount specified in section 71F(2).
47 As noted above, the review by the Tribunal of any decision of the AER was, and in broad terms still is, confined in the following ways:
the requirement of the Tribunal to be satisfied of certain pre-conditions before leave to apply for review is given by it;
the limitation upon the material which the Tribunal may have regard to in making its determination as to whether a ground of review has been made out on the material that was before the AER; and
the requirement that the issue before the Tribunal upon which a ground of review is sought to be established was properly advanced before the AER.
Those pre-conditions have been maintained, and to a degree refined.
48 In relation to the second of those matters, s 71R defined and still defines review related matter in the same way. In one significant respect, the material relevant to a review has been extended to include matters arising as a result of the consultation required by s 71R(1)(b). That obliges the Tribunal, before making a determination, to take reasonable steps to consult with network service users of the relevant services, and any user or consumer associations or user or consumer interest groups that the Tribunal considers to have an interest in the determination (excluding a user or consumer association or interest group that is a party to the review). In addition, the opportunity for the Tribunal to seek additional information relevant to the relief which it might otherwise contemplate granting has been extended by amendments to s 71R(3) and by the addition of ss 71R(5a) and (5b).
49 In substance, parallel amendments were made to the merits review of AER determinations under the NGL under Part 5 of the NGL, including the extension of the required circumstances in which leave to review may be given under s 248, and by the insertion of s 259(4a) and (4b) equating to s 71P(2a) and (2b) of the NEL, and the deletion and substitution of s 261(1) and insertion of s 261(3a) and (3b) equating to s 71R(1) and s 71R(5a) and (5b) of the NEL.
50 The consultation process referred to in s 71R(1)(b) of the NEL and s 261(1)(b) of the NGL is an additional procedural step which the Tribunal must take and, ideally, be accommodated within the target time prescribed by s 71Q of the NEL and s 260 of the NGL. The Tribunal, having given leave to apply for review in these and the related matters on 17 July 2015 (other than the JGN application where leave to apply for review was given on 30 July 2015) sought information from the AER as to all of the interest groups or persons who might have an interest in the review by the Tribunal under s 71R(1)(b) of the NEL and s 261(1)(b) of the NGL.
51 The Tribunal then conducted an extensive communication process directly with each of those entities or persons to invite them to indicate whether they wished to consult with the Tribunal in relation to any of the Final Decisions, as to the nature of their proposed participation, and as to how the consultation might best be carried out. In the light of that material, the Tribunal consulted with all of those persons on 6 and 7 August 2015. To ensure a satisfactory process, the Tribunal issued a Consultation Agenda under which it provided for those who wished to speak to the Tribunal on that occasion either personally or on behalf of an organisation, to do so. It arranged for the speakers to be listed randomly, so that there was no bias in the sequence of presenting particular perspectives, other than (of course) endeavouring to accommodate the personal circumstances and convenience of each of the proposed participants. During those consultations, members of the Tribunal sought clarification, and sometimes supplementation of comments or submissions or further development in the views expressed so that they were better understood or appreciated by the Tribunal. The transcript of that consultation process has been included by the Tribunal on its website relating to each of these applications. A list of those persons or entities who chose to make submissions, during the consultation process or in writing as a complement or supplement to oral submissions, or only by making written submissions or comments is also listed on the Tribunal’s website.
52 In the course of the consultation process, a number of significant issues of concern to consumers and consumer interests were identified. It is fair to say that price was a significant concern. It is also fair to say that there were a number of persons who participated, and whose concern was to ensure the quality, safety, reliability and security of the supply of electricity either because of their particular circumstances or their particular geographical location, or for other reasons. The balance, as the submissions exposed, is a very difficult one.
53 The applicants to the several applications (including PIAC), the AER and the interveners did not participate in the consultation process. That was appropriate, of course, because they each participated in the hearing before the Tribunal. The matters which emerged in the course of the consultation process, apart from informing the Tribunal about the concerns or views expressed, also provided the foundation for matters the Tribunal raised with the applicants, the AER and the interveners during the hearing. They also served as the focus for questions of the Minister, during the hearing, as to how the various concerns or matters raised were to be taken into account by the Tribunal in reaching its decisions on the several applications. Those matters are addressed later in these reasons for decision.
54 It is a mark of the sophistication of the participants in the consultation process that the range of matters discussed was not extensive. Of course, many focused on the price of electricity and the impact of the present and potential price and the ability of the less well-off in the community either to afford access to the electricity network at all, or at least to do so only at considerable personal cost. There was material showing the number of disconnections over time. The issue of price was not simply raised by consumers or representatives of consumers in a lower socio-economic setting, but by some smaller commercial enterprises, primary producers and others.
55 The Tribunal does not intend to do injustice to the process by listing, without setting out in detail, the views presented.
56 It is, in the view of the Tribunal, helpful to note the broad themes presented during the consultation process as they were identified by the Tribunal. That list was then circulated to the participants for comment. As it was not then suggested that it was inaccurate or incomplete, it is set out below:
• Consumer engagement and understanding
o The Tribunal’s approach to engagement
o Consumer education and access to information
o Participation in the AER processes
• Impact of electricity prices on consumers
o General consumer impact
o Vulnerable customers
o Rural and regional customers
o Price stability (including pace of any change)
o Long-term interests of consumers relating to price
• The regulatory framework
o Success of previous regimes
o Policy observations regarding the framework
• Balancing the NEO and NGO for the long-term interests of consumers
o The meaning of “long term interests of consumers”
o The price and reliability trade off
o The disconnection 'death spiral'
• Operating expenditure
o AER benchmarking approach
o Adherence to operating expenditure guidelines
o Impact of industrial agreements
o Vegetation management and bushfire risks
• Rate of return
o Adherence to the Guideline
o Estimation of return on debt
o Estimation of return on equity
o Ultimate pricing impact of the rate of return
• Demand and energy forecasts
o Inflation of demand and energy forecasts
• Demand management and innovation
o Expenditure on innovation
• Materially preferable NEO/NGO decision
57 As noted, the price of electricity was the most significant issue raised during the consultation, but there were also significant numbers of consumers whose emphasis was more on the reliability of the supply of electricity in their particular circumstances.
58 In the matters concerning the Final Decisions of the AER regarding the Networks NSW businesses, in large measure (and as set out later in these reasons) the role of PIAC was a particularly helpful one. Its three applications sought to have the relevant AER Final Decisions set aside, and to have substituted determinations through the Tribunal which would substantially lessen the amounts recoverable by the Networks NSW businesses over the regulatory period 2015-19, as well as presenting the viewpoint that the matters raised by Networks NSW to have the recoverable amounts increased were erroneous. Many of the views put forward by those on the consultation process were therefore reflected or represented by the contentions of PIAC during the hearing.
59 There is one particular feature of the consultation process views which it is appropriate to comment on at this point.
60 The Tribunal was assisted in the course of the hearing by the intervention of the Minister. It was appropriate, of course, for the Minister to intervene, given the significant changes to the legislation so far as they relate to the Tribunal’s role. The Tribunal is appreciative of those submissions.
61 The Minister, consistently with the submissions of the AER and of the interveners, took the view that the “consumers” referred to in the NEO and the NGO must be treated as a generic group, so that the Tribunal could not and should not address the particular circumstances of particular consumers or consumer interests. The Tribunal adopts that approach. When the National Electricity (South Australia) National Electricity Law (Miscellaneous Amendments) Act 2007 (SA) was first introduced, the point was then made that the objective of the national electricity market generally was to achieve an efficient and so far as possible competitive market for the supply and consumption of electricity and where that could not be achieved that the regulatory structure for access to monopoly services should endeavour to reflect that structure. It was noted (see  above) that social and environmental objectives should be a matter of separate policy of the legislature reflected in different ways. The Tribunal, as the Minister submitted, should take, and does take, that view.
62 Nevertheless, the consultation process did identify and inform the Tribunal significantly as to the matters of concern to significant sections of the consumer community and how some consumer representatives regarded the long-term interests of consumers were best served.
63 As also noted, in relation to this decision (and those concerning Essential and Endeavour), the Tribunal has had the benefit of the applications by PIAC and its helpful submissions. That has enabled the Tribunal to be acutely aware of its obligation ultimately to ensure that its decisions in relation to these applications are those which, in its view, best serve the long-term interests of consumers in terms of the NEO. It has also, by the grounds of review raised by PIAC alleging error on the part of the AER, led to a focus on those particular matters where, it is said, the AER itself has failed to respond appropriately to s 16(1)(d) of the NEL or has otherwise fallen into error to the detriment of consumers. The particular errors asserted by PIAC are addressed in the course of this decision, and to the extent to which they require separate consideration, in the course of considering the reviews of the Final Decisions of the AER in relation to Essential and Endeavour.
64 Given the role of PIAC, and the relevance of its submissions to the Tribunal’s functions and responsibilities under the legislation, the Tribunal has not needed in these matters to address separately the matters which emerged in the course of the consultation process. The role and submissions of PIAC have encompassed those matters.
65 At this point in the reasons of the Tribunal, it is not appropriate to discuss except in a conceptual way how it should approach the question of whether, in the event of finding error on the part of the AER, that is that a ground of review or grounds of review are made out, how it should address the requirements of s 71P(2a) and (2b) in the NEL or s 259(4a) and (4b) of the NGL.
66 It is clear enough from those legislative provisions that, if it is established that in one or more respects the AER has fallen into reviewable error, that is that a ground or grounds of review have been made out, it will not be a materially preferable decision simply to provide for the correction of that ground of review. Section 71P(2b)(d)(i) and s 259(4b)(d)(i) respectively make that plain.
67 It will be necessary to address how the constituent components of the reviewable regulatory decision inter-relate. To the extent to which they inter-relate, and in the light of the revenue and pricing principles, it would be necessary to determine whether, then, the materially preferable decision is to allow the decision of the AER to stand, or to vary it, or to remit the matter to the AER to address a particular aspect or aspects further.
68 To identify how constituent components of a reviewable regulatory decision do inter-relate, the Tribunal will routinely have the benefit of submissions from the AER, and in this instance has had the benefit also of the submissions of PIAC, the Minister as well as the DNSPs. It is clear that several of the issues raised by the parties giving rise to grounds of review do inter-relate, so that the Tribunal, if it finds for instance that significant grounds of review are made out in relation to the allowance for opex, would have to consider the nature and extent to which, if at all, there are consequences for other elements of the AER decision concerning the Service Target Performance Incentive Scheme (STPIS), the Efficiency Benefit Sharing Scheme (EBSS) and the X-factor (those terms are described below), and possibly metering services. Those factors are, to some extent (as acknowledged by all parties) inter-related. There is also some scope for an asserted inter-relationship between the allowance for capex and the allowance for opex, and perhaps more obviously some scope for inter-relationship between the allowance for return on equity and the allowance for return on debt, as they combine to estimate the rate of return. Those inter-relationships were not significantly developed in the course of submissions, except in a general way.
69 The Tribunal would expect that, routinely, if there were direct and measurable relationships between elements making up the AER determination, the AER would indicate how the adjustment of one would or should lead to the adjustment of the other. The AER in each of its decisions has not, and did not need to, explain those inter-relationships in detail although it has made in its general “Overview” section some comments about what might be seen to have been allowances made with some generosity towards the particular DNSP (and about which PIAC has complained) and which in a general sense might be set off against any error or a ground of review which might be exposed by a DNSP.
70 The Tribunal, when it has come to consider this aspect, has primarily looked to the inter-relationships as they have been identified and quantified by the AER, PIAC and in other submissions. It has, in addition, sought to step back and to look at the wider picture where there is no such obvious inter-relationship and no direct evidence quantifying the respective way in which one element of the building blocks as prescribed by the NER or NGR affect other elements. It is in that light, as discussed at the end of these reasons, that the Tribunal has proceeded.
71 In that context, it is also appropriate to observe that the task imposed on the AER is a protean one.
72 It has referred to that task in its written submissions. Following the 2013 Legislative Amendments, the AER in exercising its regulatory function under the NER or the NGR is required to “perform or exercise that function or power in a manner that will or is likely to contribute to the achievement of” the NEO or the NGO: see s 16(1) of the NEL and s 28(1) of the NGL.
73 In relation to a reviewable regulatory decision under the NEL or the NGR, the AER is required, inter alia (see s 16(1)(b) to (d) of the NEL and s 28(1)(b) of the NGL):
(a) to specify the manner in which the constituent components of the decision relate to each other, and the manner in which that interrelationship has been taken into account in the making of the decision; and
(b) if there are 2 or more possible reviewable regulatory decisions that will or are likely to contribute to the achievement of the NEO or NGO:
(i) make the decision that the AER is satisfied will or is likely to contribute to the achievement of the NEO or NGO “to the greatest degree” (the preferable reviewable regulatory decision (in the case of the NEL) or the preferable designated reviewable regulatory decision (in the case of the NGL)); and
(ii) specify reasons as to the basis on which the AER is satisfied that the decision is the preferable reviewable regulatory decision or the preferable designated reviewable regulatory decision.
74 The NEO is set out in s 7 of the NEL:
The objective of this Law is to promote efficient investment in, and efficient operation and use of, electricity services for the long term interests of consumers of electricity with respect to –
(a) price, quality, safety, reliability and security of supply of electricity; and
(b) the reliability, safety and security of the national electricity system.
75 The NGO is set out in s 23 of the NGL:
The objective of this Law is to promote efficient investment in, and efficient operation and use of, natural gas services for the long term interests of consumers of natural gas with respect to price, quality, safety, reliability and security of supply of natural gas.
76 Apart from the NEO and the NGO, the AER is required to take into account, in the prescribed circumstances, the RPP which are set out in s 7A of the NEL (RPP), and s 24 of the NGL respectively: s 16(2) of the NEL and s 28(2) of the NGL. None of those provisions, nor the NEO or the NGL, were amended by the 2013 Legislative Amendments.
77 The ultimate objective reflected in the NEO and NGO is to direct the manner in which the national electricity market and the national natural gas market are regulated, that is, in the long term interests of consumers of electricity and natural gas respectively with respect to the matters specified. The provisions proceed on the legislative premise that their long term interests are served through the promotion of efficient investment in, and efficient operation and use of, electricity and natural gas services. This promotion is to be done “for” the long term interests of consumers. It does not involve a balance as between efficient investment, operation and use on the one hand and the long term interest of consumers on the other. Rather, the necessary legislative premise is that the long term interests of consumers will be served by regulation that advances economic efficiency.
78 In broad terms, it can be said that the economic foundations of the regulatory regime are well understood. In Application by ElectraNet Pty Limited (No 3)  ACompT 3 at , the Tribunal said:
The national electricity objective provides the overarching economic objective for regulation under the Law: the promotion of efficient investment in the long term interests of consumers. Consumers will benefit in the long run if resources are used efficiently, i.e. resources are allocated to the delivery of goods and services in accordance with consumer preferences at least cost. As reflected in the revenue and pricing principles, this in turn requires prices to reflect the long run cost of supply and to support efficient investment, providing investors with a return which covers the opportunity cost of capital required to deliver the services.
79 As noted above, reference to the Second Reading Speech on the introduction of the National Electricity (South Australia) (New National Electricity Law) Amendment Bill (South Australian House of Assembly Hansard, 9 February 2005, p 1451) states (at p 1452):
The national electricity market objective in the new National Electricity Law is to promote efficient investment in, and efficient use of, electricity services for the long term interests of consumers of electricity with respect to price, quality, reliability and security of supply of electricity, and the safety, reliability and security of the national electricity system. The market objective is an economic concept and should be interpreted as such. For example, investment in and use of electricity services will be efficient when services are supplied in the long run at least cost, resources including infrastructure are used to deliver the greatest possible benefit and there is innovation and investment in response to changes in consumer needs and productive opportunities. The long term interests of consumers of electricity requires the economic welfare of consumers, over the long term, to be maximised. If the National Electricity Market is efficient in an economic sense the long term economic interests of consumers in respect of price, quality, reliability, safety and security of electricity services will be maximised. ... Applying an objective of economic efficiency recognises that, in a general sense, the national electricity market should be competitive, that any person wishing to enter the market should not be treated more or less favourably than persons already participating in the market, and that particular energy sources or technologies should not be treated more or less favourably than other energy technologies.
80 Thus, in Application by Envestra Limited (No 2)  ACompT 3 (Envestra (No 2)), the Tribunal summarised with approval certain submissions by the AER, which were not challenged in that case (at ):
The AER submitted that rule 91 requires the AER to permit service providers a reasonable opportunity to recover what the AER considers "legitimate costs". Legitimacy, according to the AER is informed by the NGO and, in particular, means costs that would be incurred in a "workably competitive market". The requirement for replication of a workably competitive market outcome is said to be derived from the intent of the regulatory framework. This phrase appears to come from the Australian Energy Market Commission, Rule Determination, National Electricity Amendment (the Economic Regulation of Transmission Services) Rule 2006 No. 18, published on 16 November 2006. In this determination, the Australian Energy Market Commission, at page 93, describes the fundamental objective of regulation as being:
to reproduce, to the extent possible, the production and pricing outcomes that would occur in a workably competitive market in circumstances where the development of a competitive market is not economically feasible...
81 A similar point was made in the High Court with respect to a very similar progenitor under the then applicable gas regime, in East Australian Pipeline Pty Ltd v Australian Competition and Consumer Commission (2007) 233 CLR 229 at  (East Australian Pipeline):
The context and purpose of the Code is well understood, not least because the objectives of the legislation are articulated in the legislation itself in considerable detail. The Code as a whole provides for a regulatory regime of a kind which is "a surrogate for the rewards and disciplines normally provided by a competitive market". Competitive pressures in a market stimulate efficiency of production and resource allocation, they stimulate efficient investment decisions and they minimise costs. No party disputed the fact that the regulatory process set out in the legislation was directed to eliminating monopoly pricing whilst nevertheless providing a rate of return to pipeline owners, commensurate with a competitive market.
82 Those references have a particular significance in this matter as appears in the Tribunal’s consideration of the issues concerning rate of return on debt.
83 It is convenient, at this point, to note how the AER, in the Overview section of its Final Decision, explained how it had sought to fulfil its obligation under s 16(1)(d) of the NEL and s 28(1)(b)(iii) of the NGL to make the decision it is satisfied will or is likely to contribute to the achievement of the NEO/NGO to the greatest degree, that is to make the preferable designated reviewable regulatory decision, and to specify its reasons for doing so. Of course, there is further relevant discussion in the various Attachments to its Final Decisions.
84 In the Overview to the Ausgrid Final Decision, Section 1.2 “Contribution to the achievement of the NEO”, the following first appears (at p 10):
We are satisfied that the total revenue approved in our final decision contributes to the achievement of the NEO to the greatest degree. This is because our total revenue reflects the efficient, sustainable costs of providing network services in Ausgrid’s operating environment and the key drivers of efficient costs facing Ausgrid. Our decision will promote the efficient investment in, and efficient operation and use of, electricity services for the long term interests of consumers, as required by the NEO. We set out our reasons below and in our attachments.
85 There follows a description of the key drivers of costs facing a network service provider, acceptance that the key drivers may change from one regulatory period to the next, the most important factors impacting on Ausgrid’s costs in the 2015-2019 regulatory control period, and its overall conclusion. The AER says that the two constituent components of its decision which drive most of the differences between Ausgrid’s proposed revenue and its Final Decision are rate of return (equity and debt), and opex. Those two differences are then explained in some detail.
86 Section 1.4 “Assessment of options under the NEO” (at pp 19-20) returns to the legislative requirement. It attracted considerable attention, particularly by PIAC in it submissions. It is desirable to set it out in full:
The NER recognises that there may be several decisions that contribute to the achievement of the NEO. Our role is to make a decision that we are satisfied contributes to the achievement of the NEO to the greatest degree.
For at least two reasons, we consider that there will almost always be several decisions that contribute to the achievement of the NEO. First, the NER requires us to make forecasts, which are predictions about unknown future circumstances. As a result, there will likely always be more than one plausible forecast. Second, there is substantial debate amongst stakeholders about the costs we must forecast, with both sides often supported by expert opinion. As a result, for several components of our decision there may be several plausible answers or several point estimates within a range. This has the potential to create a multitude of potential overall decisions. In this decision we have approached this from a practical perspective, accepting that it is not possible to consider every possible permutation specifically. Where there are several plausible answers, we have selected what we are satisfied is the best outcome, under the NEL and NER.
In many cases, our approach results in an outcome towards the end of the range of options materially favourable to Ausgrid (for example, our choice of equity beta). While it can be difficult to quantify the exact revenue impact of these individual decisions, we have identified where we have done so in our attachments. Some of these decisions include:
• selecting at the top of the range for the equity beta
• setting the return on debt by reference to data for a BBB broad band credit rating, when the benchmark is BBB+
• the cash flow timing assumptions in the post-tax revenue model
• the point at which we have set the benchmark for opex
• the allowances we have made for operating environment factors in our benchmarking analysis.
We set out our detailed reasons in the attachments. They demonstrate that the constituent components of our decision comply with the NER’s requirements. At an overall level our decision reflects the key reasons set out above, which indicate that Ausgrid should recover less revenue than it has proposed or recovered in recent years. Our decision reflects these at both the constituent component and overall revenue levels.
Given our approach, we are satisfied that our decision will or is likely to contribute to the achievement of the NEO to the greatest degree.
87 The legislative and regulatory background referred to above highlights the complex task of the Tribunal since the 2013 Legislative Amendments.
88 Although the Minister submitted that the 2013 Legislative Amendments may have required a reconstruction of certain provisions of the NEL and the NGR, even unamended provisions, the Tribunal did not discern from the Minister’s submissions that it was necessary to refine or re-define the way the earlier and unamended provisions of the NEL or the NGR have been applied or construed.
89 Rather, it can be accepted, the 2013 Legislative Amendments:
(a) clarify or emphasise the alignment of both the decision making role of the AER by s 16(1)(d)(i) of the NEL and s 28(1)(b)(iii)(A) of the NGL and the decision making role of the Tribunal by s 71P(2a)(c) of the NEL and s 259(4a)(c) of the NGL, with the achievement of the NGO and the NGL objectives; and
(b) require the Tribunal, when deciding whether to vary the AER decision under review or to remit it to the AER for further consideration to do so only if it is satisfied that that will result in an improved decision fulfilling better the NEO or the NGO in the long term interests of consumers. The Minister’s submission, which the Tribunal accepts, is that a regulatory regime such as the NEL and NGL serves as a “surrogate for the rewards and disciplines normally provided by a competitive market” (as originally cited at  above in East Australian Pipeline (2007) 233 CLR 229 at ). As a surrogate, of course, it can only approximate those rewards and disciplines.
90 As the AER and the Minister point out, this feature of the regulatory regimes acknowledges the existence of a range of possible decisions which equally, or approximately equally, promote economic efficiency. But that is not to say that such decisions will equally, or even approximately equally, promote the long term interests of consumers. As was said in introducing the 2013 Legislative Amendments:
The national electricity objective and national gas objective explicitly target economically efficient outcomes that are in the long term interests of consumers, but the nature of decisions in the energy sector are such that there may be several possible economically efficient decisions, with different implications for the long term interests of consumers. (emphasis added)
See: South Australia, House of Assembly, Hansard, 26 September 2013 at 7172 (The Hon J R Rau).
91 Consequently, the correction of error or errors in a decision under review will not necessarily lead to a materially preferable decision. Whether there is a preferable decision to the decision made by the AER depends upon an assessment of the decision as a whole, and a comparison of that decision with a putative alternative decision; it does not depend simply on an assessment of errors in individual components of the decision under review. That reflects the Minister’s comments that the 2013 Legislative Amendments:
Require the [Tribunal] to undertake a holistic assessment of whether the setting aside or varying of the reviewable regulatory decision, or remission of the matter back to the original decision maker, will or is likely to deliver a materially preferable outcome in the long term interests of consumers.
See: South Australia, House of Assembly, Hansard, 26 September 2013 at 7173 (The Hon J R Rau).
92 The 2013 Legislative Amendments reflect a deliberate policy decision to change the NEL and NGL and, in particular, to change the scope of the Tribunal’s limited merits review function. They introduce a series of steps which require the Tribunal, even if it is satisfied of one or more grounds of review arising from one particular aspect of the AER’s decision, to consider whether and how the potential consequences of that ground being established may be reduced, counterbalanced or rendered immaterial following the processes mandated by ss 71P(2a), 71P(2b)(a) and 71P(2b)(c) of the NEL and ss 259(4a), 259(4b)(a) and 259(4b)(c) of the NGL.
93 Nevertheless, as the Minister said, it is axiomatic in the principles of regulatory economics, that promoting allocative, productive and dynamic efficiency generally serves the long term interests of consumers. However, the 2013 Legislative Amendments contemplate that there can be more than one available decision that is economically efficient – and certainly more than one available decision that is roughly so, having regard to the unavoidable approximations involved.
94 The role of the AER and the Tribunal in giving effect to the NEO and NGO is to promote the “long term interests of consumers” with respect to the matters stipulated. This will always involve an attempt to promote efficient investment in, and operation and use of, services, but will also require taking into account other factors as appropriate. The Minister gave emphasis to taking into account the appropriate “long term” character of the consumer interests that are to be promoted. He also said that it may also, in an appropriate case, require taking into account the distribution to consumers of the benefits of efficiencies, even though this may not bear on the economic efficiency of the decision. Of course, the benefits of efficiencies are to be awarded to consumers – that is, to use the Minister’s word “axiomatic”. The Tribunal has previously acknowledged that “in some circumstances” it may, in effect, be preferable for the benefits of economic efficiencies to be passed to consumers in their long term interests rather than wholly retained or captured by the regulated entity: Envestra (No 2)  ACompT 3 at .
95 It is nevertheless accepted that the decision which is to be made must comply with, or meet, the requirements of the relevant Rules – whether the NER or the NGR. The AEMC has specified the means by which the AER (and the Tribunal) is to reach its decision. An “holistic” assessment of the AER decision by the Tribunal cannot entitle it to ignore the relevant Rules or the RPP in s 7A of the NEL or s 24 of the NGL which inform proper application of the relevant Rules made by the AEMC. They are made precisely because, in both the national electricity network and the national gas network, the underlying objective is the establishment and maintenance of competition in the long term interests of consumers. Where there is monopoly infrastructure (such as the transmission and distribution networks), the regulatory process is to reach an outcome by the relevant Rules set down by the AEMC, and then by the AER applying the Rules as appropriate to simulate as best as can be done under the relevant Rules the outcome for the DNSPs by its reviewable regulatory decisions which would reflect the outcome of a competitive market.
96 The AER made that point in its submissions. It pointed out that, in a workably competitive market, an inefficient business with higher costs than its efficient competitors still receives the market price which is set by the efficient competitors. As all businesses in workably competitive markets receive the market price, an inefficient business will only be able to obtain a price which is lower than its costs. It could not ask its customers to pay higher prices (generally, even for a transitional period) to fund the costs of it moving away from inefficient practices and, accordingly, will be unable to achieve the same returns to shareholders as an efficient business.
97 It is desirable to say something about the expression “will, or is likely to, result in ....” in s 71P(2a)(c).
98 The submissions acknowledge that the word “likely” may be given a range of, or shades of, meaning: see eg Tillmans Butcheries Pty Ltd v Australian Meat Industry Employees Union (1979) 27 ALR 367 per Deane J at 380.
99 As the Minister has noted, if the words “is likely to” are used in contradistinction to the preceding word “will”, they connote a lower standard of satisfaction than “will”. Several provisions of the Competition and Consumer Act 2010 (Cth), directed to proscribing restrictive trade practices, are enlivened in relation to conduct that “has, or is likely to have, the effect of substantially lessening competition” (ss 45 and 47) or “have the effect, or be likely to have the effect, of substantially lessening competition” (s 50). A number of cases suggest that “likely”, in this context, connotes a “real chance” – something more than a mere possibility, but less than a likelihood on the balance of probabilities: see especially Monroe Topple & Associates Pty Ltd v Institute of Chartered Accountants (2002) 122 FCR 110; Seven Network Limited v News Limited (2009) 182 FCR 160 at ; Australian Gas Light Co v ACCC (2003) 137 FCR 317 at .
100 In the present context, the legislature has identified a particular goal – the regulatory decision that advances to the greatest degree the NEO or NGO (s 16(1)(d) of the NEL and s 28(1)(b)(iii)(A) of the NGL). It has charged the AER with responsibility to achieve that goal, and empowered the Tribunal to set aside or vary the AER’s attempt only if “satisfied” that to do so will or is likely to result in a materially preferable decision.
101 In that context, the Tribunal considers it appropriate to proceed on the basis that the phrase “will, or is likely to” should be construed as a compendious expression of a standard of likelihood that is equivalent to “more likely than not”. It considered that that approach best achieves the purpose or object of the NEL and NGL. It reflects consistency in the intended standard of satisfaction on the part of the Tribunal with the standard which the AER was required to apply. It is also consistent with the Second Reading Speech referred to above, in which the amendments were explained as being directed towards “ensur[ing] that the limited merits review only results in changes to decisions under review where the [Tribunal] concludes that there is a materially preferable decision in the long term interests of consumers”.
102 Understandably, both at the stage of considering whether leave to apply for review and on the reviews, there was considerable debate about the character of the available grounds of review and whether, in the particular circumstances a ground of review had been made out. Concern was expressed by the AER about the breadth of expression adopted by the various DNSPs to invoke the grounds of review.
103 With one qualification, the Tribunal considers that it is preferable to defer addressing that debate until the point of considering separately the particular issues which the various applicants raised, and how the asserted error was then described. It did not discern, in the course of the three intense weeks of competing submissions, that ultimately the respective applicants had expressed their contentions in a way which caused any surprise or unfairness to the AER. That is so, even though in some instances, the relevant applicant sought to engage one or more of the available grounds of review as applicable to a particular contention of error. It is preferable to deal with those matters as they come to arise in their context.
104 As will be seen, the Tribunal is well alive to the terms of s 71C of the NEL and s 246 of the NGL. The onus is on a particular applicant to satisfy the Tribunal that the AER determination, in the respect being debated, is in error within one or more of the available grounds: Application by DBNGP (WA) Transmission Pty Ltd (No 3)  ACompT 14 at -. See also the Tribunal’s remarks in Application by EnergyAustralia  ACompT 8 at ; Application by WA Gas Networks (No 3)  ACompT 12 at  (WA Gas Networks).
105 The one possible reservation to those comments is by PIAC’s submission that s 16(1)(d) of the NEL as introduced by the 2013 Legislative Amendments has widened the grounds of review available under s 71C(1) of the NEL.
106 It may be accepted that the amendment of one provision in an Act may impliedly alter the meaning of an unamended provision of that Act: Commissioner of Stamps (SA) v Telegraph Investment Co Pty Ltd (1995) 184 CLR 453 at 463.
107 However, the Tribunal does not consider that s 16(1)(d) expands in any way the grounds of review in s 71C(1) of the NEL. There is no textual or contextual reason why it should do so, and the extraneous material to which it is permissible to refer does not suggest that that was intended.
108 The PIAC submissions focused upon particular steps in the AER decision making process to demonstrate, in the terms of s 71C, the ground of review being made out. Whilst there is clearly a heavy responsibility on the AER to make the decision which contributes to the NEO “to the greatest degree”, the PIAC contention that it had not done so was not at large, but in the context of the particular asserted grounds of review. A ground or grounds of review, if made out, enlivens the further steps to be addressed by the Tribunal under s 71P of the NEL. The Tribunal has not found it either necessary or helpful, in considering whether PIAC has made out the grounds of review for which it contends, to give any expanded meaning to the available grounds of review. It has taken them as they appear, and as they have been explained in other decisions. On the other hand, the Tribunal has not taken the view that the satisfaction of the AER that it has met the requirement of s 16(1)(d) precludes the Tribunal from itself having to address those steps in s 71P once it has been satisfied that a ground or grounds of review have been made out. If, as a result of that process, the Tribunal determines (as it has done in some respects) to set aside the AER determination and to remit the matter to the AER, it has not needed to separately conduct a qualitative assessment of the AER’s satisfaction under s 16(1)(d) because (it has been satisfied) that its assessment was in part the product or consequence of the error now exposed by the established ground of review.
109 The review applicants (variously supported by the interveners) take issue with aspects of the following building blocks in the AER’s decisions:
Opex: PIAC; Networks NSW; ActewAGL
X-factor: Networks NSW
Efficiency Benefit Sharing Scheme (EBSS): Networks NSW
Service Target Performance Incentive Scheme (STPIS): ActewAGL
Return on equity: PIAC; Networks NSW; ActewAGL; JGN
Return on debt: PIAC; Networks NSW; ActewAGL; JGN
Gamma: Networks NSW; ActewAGL; JGN
Metering services - opex: ActewAGL
Metering classification: ActewAGL
Metering services: Ausgrid
Market Expansion Capital Expenditure (ME Capex): JGN
110 The Tribunal has approached each review application in that sequence (where it applies to a particular applicant or applicants for review).
111 Under the transitional arrangements, the AER was required to make placeholder distribution determinations for the ACT and NSW DNSPs for the transitional regulatory control period which would apply for one year, from 2014-15 to 2015-16. The AER was then required to carry out a full regulatory determination process and make distribution determinations for the ACT and NSW DNSPs for the subsequent regulatory control period, from 2015-16 to 2018-19.
112 Due to the substantial commonality of issues raised, it was common ground that it would be appropriate for the applicants to prepare common written submissions in relationto those issues or topics which it had substantially in common with other applicants. On this basis, and pursuant to the Tribunal’s directions of 5 August 2015:
(a) the Network Applicants prepared common written submissions on the issues of return on equity and the value of imputation credits;
(b) Networks NSW and ActewAGL prepared common written submissions on return on debt; and
(c) Networks NSW prepared common written submissions on framework, opex, X-factor, EBSS, the application of s 71O of the NEL and materially preferable NEO decision.
113 The applicants were also represented during the hearing by common counsel in respect of those issues or topics which it had in common with the other applicants. Relevantly, common counsel appeared on behalf of each of the Network Applicants in relation to return on equity and gamma, on behalf of Networks NSW and ActewAGL in respect of return on debt and on behalf of each of Networks NSW in relation to framework, opex, X-factor, EBSS, s 71O of the NEL and materially preferable NEO decision. In addition, during the course of the hearing the Network Applicants and the interveners adopted the submissions of other parties where it was appropriate to do so.
114 As noted above, where there is a “shared” issue or topic, the Tribunal has endeavoured to incorporate by reference the general or common consideration or matters addressed above. The particular aspects of the application are of course separately addressed. The Tribunal’s reasons so far as they relate to the particular aspects of the six applications not specifically addressed here are outlined in Applications by Public Interest Advocacy Centre Ltd and Endeavour Energy  ACompT 2; Applications by Public Interest Advocacy Centre Ltd and Essential Energy  ACompT 3; Application by ActewAGL Distribution  ACompT 4 and Application by Jemena Gas Networks (NSW) Ltd  ACompT 5.
115 This topic occupied a substantial part of the hearing, and a considerable volume of the very extensive documentary material in the review-related material: s 71R of the NEL (and s 261 of the NGL).
116 Before turning to the particular opex issues, it is helpful to note more broadly the structure of Ch 6 of the NER, relating to the economic regulation of distribution services. Section 6.1 deals with introductory matters. Rule 6.1.1 affirms that the AER is responsible for the economic regulation of distribution services.
117 Of general relevance is that the structure of Ch 6 set out in r 6.12 includes that Part C sets out the building block approach to the regulation of services of the character provided by Networks NSW and ActewAGL. Part B amongst other things, obliges the AER to make a distribution determination for each DNSP: r 6.2.4. Rule 6.2.8 requires the AER to make and publish, amongst others, the RoR Guideline and the EFA Guidelines. It does not oblige the AER to adhere to those guidelines, but it must explain in its relevant decision why it has departed from them: r 6.2.8(c).
118 Part C of Ch 6 is of immediate and direct relevance. Rule 6.3 defines a building block determination as a component of a distribution determination. The procedure to get to that point is contained in Part E of Ch 6, including for each DNSP to submit a building block proposal as prescribed.
119 Rule 6.3.2 prescribes the contents of a building block determination. Rule 6.4.3 says that the building block generally are, for each regulatory year of a regulatory control period, to provide (as relevant to the present applications) for:
(1) indexation of the regulatory asset base: see r 6.5.1;
(2) a return on capital for that year: see r 6.5.2 – within r 6.5.2 both the return on equity and the return on debt, as well as the RoR Guideline, are addressed;
(3) the estimated cost of corporate income tax of the DNSP for that year: r 6.5.3;
(4) the revenue increments or decrements for that year (relevantly) from the EBSS and the STPIS: rr 6.5.8, 6.5.8A and 67.6.2;
(5) the forecast opex for that year: r 6.5.6.
120 Those, and the other elements of the building blocks, including the forecast capex for the regulatory control period: r 6.5.7, are then detailed and cross-referenced to the relevant rules, as referred to in r 6.5.
121 As indicated, these reasons generally deal only with the application to review the AER’s Final Decision relating to Ausgrid. The following focuses on the attachment to that decision (Attachment 7 – Operating expenses) which provides an overview of the AER’s assessment of Ausgrid’s opex and, in appendixes to the attachment, an analysis of its assessment of that opex.
122 However, because much of the AER’s reasons for not accepting Ausgrid’s forecast opex is common to its decisions in relation to the opex forecasts of the other two Networks NSW businesses, namely Essential and Endeavour, and to ActewAGL, it is convenient to consider their challenges along with those of PIAC and Ergon which also address the principal issue under this heading. It is noted that, as well as the three Networks NSW applications (and the ActewAGL application), PIAC separately challenged the opex allowance for the Networks NSW businesses in its three applications. JGN did not raise this issue.
123 The principal issue that may be drawn from the following overview of the parties’ challenges is whether the AER’s application of what the parties referred to as the EI model discharged its obligations under rr 6.5.6 and 6.12.1(4).
124 That issue was expanded by Mr O’Bryan QC for the AER as follows:
It’s said that the AER has not complied with the regulatory rules; and if made out, that would involve an incorrect exercise of discretion, and it’s also said principally that the EI model that was used by the AER, both in assessing the DNSPs proposed opex and also in estimating required opex was flawed. And that argument is properly characterised as either the AER incorrectly exercising its discretion or making an unreasonable decision. Broadly, it will be necessary for the Tribunal to determine whether the AER’s approach is compliant with the regulatory framework and whether the decision had a reasonable basis.
… … …
…all econometric models are an approximation, they’re a simplification of the real world. They can never reflect absolutely all the on-the-ground features of the real world and that must be recognised. That being recognised, a regulator will take steps, or ought to take steps, acting reasonably, to make allowances for what’s not revealed by the model, and so the question then becomes as what the AER has done in its estimation, has it taken sufficient reasonable steps to make those allowances.
125 The three Networks NSW DNSPs challenge the AER’s estimate (made under r 6.12.1(4)) of their required opex because, in their contention, the AER’s estimates were too low. The figures in this part of the reasons are presented by the parties, some of which did not specify whether they are nominal or real figures. Networks NSW estimate the negative impact on each of its DNSPs to be: Ausgrid $731m, Endeavour $264m and Essential $737m.
126 It is Networks NSW view that the opex actually incurred by a DNSP is the best source of information for the AER of its required opex. Networks NSW submits that the AER has, however, ignored that information and instead relied on an unsound and untested econometric model (developed by Economic Insights Pty Ltd (EI) and which the parties refer to as the EI model) to estimate opex for each of the three Networks NSW DNSPs by reference to other businesses against which they were benchmarked. The Tribunal will adopt the term EI model as used by the parties.
127 Networks NSW also submits that the issue is not whether the EI model is better than alternative models but whether, having regard to the data limitations and other matters, any of the models are fit to be given 100 percent weight in assessing an appropriate level of opex.
128 ActewAGL also challenges the AER’s estimate of its required opex because, in its view, the AER’s estimate is too low. The AER’s decision to not accept ActewAGL’s opex forecast resulted in a $130.6m ($2013-14) reduction in the forecast.
129 ActewAGL identifies what it perceives as three broad areas of deficiency in the AER’s decision.
130 First, the AER’s methodology is inconsistent with the methodology prescribed by rr 6.5.6(c) and 6.12.1(4) of the NER, and thus contrary to law.
131 Secondly, the AER’s benchmarking has such serious technical deficiencies that it has no value as a means of assessing ActewAGL’s efficient costs.
132 Thirdly, the AER is in error by confining itself in its consideration of the question whether ActewAGL’s forecast opex reasonably reflects the efficient costs of a prudent operator, having regard only to exogenous considerations (ie matters that are beyond the control of a DNSP, such as the weather and geography). That is to say, in making its decision the AER has assumed that the NER prohibits it from taking into account the real world consequences of its decision on consumers and others to the extent that those consequences arise from endogenous considerations (ie matters within the control of ActewAGL, such as its previous business decisions). This, ActewAGL submits is:
(a) wrong as a matter of law; and
(b) having regard to what it describes as the serious impact of the AER’s decision on ActewAGL’s ability to deliver safe and reliable supplies of electricity, is a matter that the AER should have considered.
133 PIAC challenges the AER’s estimates of the three Networks NSW DNSPs required opex because, in its view, the AER’s estimate of opex for each of them is too high: by $365m for Ausgrid, $196m for Endeavour and $291 for Essential.
134 PIAC generally endorses the AER’s benchmarking methodology but takes issue with the AER’s adjustments to the EI model that are, in the words of PIAC, “arbitrary and illogical”. It submits that the adjustments “disguised a very substantial further relaxation from a position that, at the draft decision stage, the AER had already described as “cautious” and “conservative”.
135 It is PIAC’s submission that the “relaxation” results in the Networks NSW DNSPs receiving opex allowances:
(a) well in excess of the efficient opex requirements of a prudent operator; and
(b) substantially higher,
than if the AER had applied the results of its benchmarking techniques in an internally logical manner and without a quantitative basis in favour of the DNSPs.
136 Ergon (as an intervener in these matters) challenges the decisions because the AER used a flawed model to arrive at its estimates of opex.
137 Further particulars of what the parties perceive as deficiencies in the EI model, the AER’s application of it and the AER’s benchmarking methodology generally appear below.
138 As observed, the NEL and the NER regulate the revenue that a DNSP may derive from the provision of electricity distribution services and the NEL provides that the AER is responsible for the economic regulation of electricity distribution services, including determination of the DNSP’s annual revenue requirements.
139 The annual revenue requirement for a DNSP for each year of a regulatory control period must be determined using a building block approach. One of those building blocks is the forecast opex for that year (r 6.4.3(a)(7)).
140 Rule 6.4.3(b)(7) provides that a DNSP’s opex for the year is its forecast opex as accepted or substituted by the AER in accordance with r 6.5.6.
141 The parties’ submissions focused on:
(a) whether the AER’s substitution of forecast opex was in accordance with the requirements of r 6.5.6 which, as a result of the 2012 Rule Amendments, is in a form different from that previously applied by the AER in determining a DNSP’s opex allowance; and
(b) the significance of the 2012 Rule Amendments vis á vis r 6.5.6.
142 Briefly, rule 6.5.6 requires that:
(a) a DNSP’s building block proposal must include the total forecast opex it considers is required to achieve each of four opex objectives in r 6.5.6(a)(1)-(4);
(b) the AER must accept the DNSP’s forecast if it is satisfied that it reasonably reflects each of three opex criteria in r 6.5.6(c)(1)-(3);
(c) if the AER is not so satisfied, it must not accept the forecast: r 6.5.6(d);
(d) in deciding whether it is satisfied that a DNSP’s forecast reasonably reflects each of three opex criteria, the AER must have regard to eleven opex factors in r 6.5.6(e): r 6.5.6(c).
143 Rule 6.12.1 provides that a distribution determination is predicated on a number of “constituent decisions” by the AER that are specified in that rule. Rule 6.12.1(4) specifies that one of those constituent decisions is a decision in which the AER either:
(a) acting in accordance with r 6.5.6(c), accepts the DNSP’s forecast opex (ie is satisfied that it reasonably reflects each of the three opex criteria: r 6.12.1(4)(i); or
(b) acting in accordance with r 6.5.6(d), does not accept the DNSP’s opex forecast, in which case the AER must set out its reason for its decision and an estimate of the total the DNSP’s required opex that the AER is satisfied reasonably reflects the three opex criteria in r 6.5.6(c) taking into account the eleven opex factors in r 6.5.6(e): r 6.12.4(ii).
144 The relevant 2012 Rule Amendments vis á vis r 6.5.6 as identified in the parties’ respective submissions may be summarised as follows:
(a) a change to r 6.5.6(c)(2) which, in effect, changed the focus of the rule from the costs that a prudent operator in the circumstances of the relevant DNSP (ie the DNSP whose opex forecast is being assessed by the AER) to the costs of a prudent operator per se;
(b) the deletion of what was opex factor 6.5.6(e)(4) (benchmark opex that would be incurred by an efficient DNSP) and the insertion of a new opex factor 6.5.6(e)(4) (the most recent annual benchmarking report published by the AER under r 6.27 and the benchmark opex that would be incurred by an efficient DNSP);
(c) the insertion of opex factor 6.5.6(e)(7) (the substitution possibilities between opex and capital expenditure);
(d) the insertion of opex factor 6.5.6(e)(12) (any other factor that the AER considers relevant and which it has notified the DNSP prior to the DNSP submitting its revised regulatory proposal);
(e) the deletion of r 6.12.3(f) which constrained the AER’s discretion in developing a substitute estimate under 6.12(4)(ii) by providing, in effect, that if the AER refused to approve an estimate, the substitute estimate must be:
(i) determined on the basis of the DNSP’s regulatory proposal; and
(ii) amended from that basis only to the extent necessary to enable it to be approved in accordance with the NER; and
(f) the amendment of r 6.2.8(a)(1) and the introduction of r 6.4.5 requiring the AER to develop and publish EFA Guidelines specifying the approach that the AER proposes to use to assess a DNSP’s opex and capex forecasts.
145 The substance of what the parties referred to as the EI model as it was applied by the AER appears in two reports that EI prepared for the AER, namely:
(a) the First EI Report: Economic Benchmarking Assessment of Operating Expenses for NSW and ACT Electricity DNSPs, 17 November 2014; and
(b) the Second EI Report: Economic Insights, Response to Consultants’ Reports on Economic Benchmarking of Electricity DNSPs, April 2015, prepared by EI in response to the DNSPs’ criticism of the AER’s reliance on the First EI Report in its draft decisions.
146 The release of the First EI Report:
(a) on 18 November 2014 pre-dated the AER’s publication of its draft decisions for the Networks NSW DNSPs and ActewAGL on 27 November 2014 by but nine days; and
(b) constituted the first signal of EI’s reliance on overseas data in its model and the AER’s acceptance of such data.
147 It appears from the First EI Report that the AER engaged EI to assist it with the application of economic benchmarking and to advise it on:
(a) whether the AER should make adjustments to base opex for Networks NSW and ActewAGL based on the results from economic benchmarking models; and
(b) the productivity change to be applied to forecast opex for these DNSPs.
148 To that end, EI developed:
(a) three econometric benchmarking models; and
(b) two Productivity Index Number (PIN) benchmarking models.
149 An econometric model seeks to:
(a) estimate a relationship between opex and output or explanatory variables, such as those used in the EI model (as explained below, customer numbers, circuit length, maximum ratcheted demand and the proportion of underground circuits); and
(b) use the variation in costs not explained by the output or explanatory variables to derive an estimate of inefficiency for each benchmarked DNSP.
150 An econometric model will produce only an estimate of the relationships between opex and the specified output variables. Thus, taking Ausgrid as an example and using its customer numbers, circuit length, maximum ratcheted demand and share of undergrounding for a particular year, the output of an econometric model would not exactly equal Ausgrid’s actual opex for that year.
151 As observed by the Productivity Commission: Electricity Network Regulatory Frameworks, Vol 1, 9 April 2013, p.182:
Benchmarking models do not actually estimate inefficiency, although this is how they are generally interpreted. The results of any benchmarking model show the extent to which the model fails to explain performance … That is, the inefficiency of any business is the difference between the business’s observed performance and that predicted by a set of cost drivers. This can reflect missing cost drivers, data errors, incorrect estimation methods, and invalid assumptions about the functional form and error distributions.
152 Reasons why an econometric model’s estimate of a DNSP’s inefficiency may differ from its actual inefficiency include:
(a) the data used is not accurate (eg suffer from some measurement error or are not comparable);
(b) the sample used may be too small to produce accurate estimates;
(c) the variables included in the model may be either inappropriate or incomplete (in the sense they do not reflect all the relevant drivers of opex); or
(d) the assumptions underpinning the econometric model are inappropriate.
153 As noted in in a report prepared by Pacific Economics Group Research LLC (PEG), Statistical Benchmarking for NSW Distributors, 19 January 2015 (the Second PEG Report) for Networks NSW at p 18:
Some of these sources of error may not be detectable based on the model results alone, and therefore must be guarded against through the careful application of economic theory and sector-specific knowledge.
(This report is referred to as “the Second PEG report” because prior to the draft determinations, the AER retained PEG to compile a US data set for the AER which, in the event, it did not use.)
154 The three econometric benchmarking models developed by EI are:
(a) a Cobb Douglas stochastic frontier analysis (CD SFA) opex cost function model (EI selected this CD SFA model as its preferred econometric model. It is, as noted by the AER at p 7-26 of Attachment 7 to the Ausgrid Final Decision and elsewhere, its “preferred model” and is the model the parties refer to as the EI model);
(b) a Cobb Douglas least squares econometric (LSE): an econometric opex cost function using the Cobb Douglas functional form; and
(c) a translog LSE: an econometric opex cost function using the translog functional form.
155 The PIN benchmarking models developed by EI are:
(a) a multilateral total factor productivity (MTFP) model (which assesses the productivity of all inputs, opex and capital, relative to total output); and
(b) a multilateral partial factor productivity (MPFP) model (which assesses the productivity of opex as an input relative to total output).
156 While EI derived its MTFP and MPFP scores using only the Australian DNSPs’ responses to the AER’s regulatory information notices (RINs), each of EI’s econometric models used data derived from 68 DNSPs as follows:
(a) all 13 Australian DNSPs’ responses to the AER’s RINs;
(b) 18 New Zealand DNSPs; and
(c) 37 Ontario DNSPs.
As a result, 19 percent of the data for the EI model was derived from Australian DNSPs, 26 percent from New Zealand and 54 percent from Ontario DNSPs.
157 EI’s explanation for its use of what it described as: “… comparable regulators’ data from New Zealand and Ontario” appears in the First EI Report at pp 28-29 as follows:
After a careful analysis of the economic benchmarking RIN data we concluded that there was insufficient variation in the data set to allow us to reliably estimate even a simple version of an opex cost function model.
… … …
We thus concluded that to obtain robust and reliable results from an econometric opex cost function analysis we needed to look to add additional cross sectional observations which meant drawing on overseas data, provided largely comparable DNSP data were available.
158 EI’s First Report emphasised that the reason for its inclusion of the overseas data was to increase the sample size to obtain what it described as:
(a) more robust estimates of the slope coefficients in the cost function; and
(b) more robust opex efficiency comparisons among the Australian DNSPs.
159 Benchmarking the Australian DNSPs against their international counterparts was not, it said, one of its objectives. Thus, it explained at p 31, it included country-level dummy variables (for New Zealand and Ontario) in its cost functions to:
… control for possible cross-country differences/inconsistencies in accounting definitions, price measures, regulatory and physical operating environments, etc. As a consequence, all cost efficiency scores obtained are relative to Australian best practice and NOT relative to international best practice.
160 As EI could not be certain it had exactly the same opex coverage across the three countries it included country dummy variables for New Zealand and Ontario to pick up differences in opex coverage (as well as systematic differences in operating environment factors such as the impact of harsher winter conditions in Ontario). As explained by EI in the First EI Report at p 31, the country dummy variables also pick up differences in conversion factors not adequately captured by its use of the Organisation for Economic Cooperation and Development’s (OECD) gross domestic product (GDP) purchasing power parities to convert financial variables to Australian dollars.
161 Thus, in very simplified terms, drawing on EI’s explanation:
(a) a country dummy variable for Ontario may be implemented by, say, a value of 1 being assigned to a DNSP based in Ontario and 0 otherwise and will be the coefficient representing the percentage by which a DNSP’s opex is higher or lower if it is located in Ontario, all other things being equal;
(b) while the inclusion of a dummy variable in an econometric model will change the ‘intercept’, it will not change the ‘slope’ of the model, ie where a dummy variable is used (here to indicate the different countries in which a DNSP may be located) for each DNSP the slope coefficient of the relationship (customer numbers, ratcheted maximum demand etc) remains the same. So in the case of the EI Model, the addition of dummy variables for Ontario and New Zealand assumes that the underlying relationship between, say, customer numbers and opex is the same for a DNSP regardless of the country in which it is located.
162 Relying on the Second EI Report, the AER submissions provide the following examples of what the country dummy variables correct for:
(a) differences in opex coverage (eg the inclusion or exclusion of opex associated with very high voltage assets);
(b) differences in DNSP activity coverage (eg whether the DNSP performs meter reading);
(c) operating environment differences, such as the impact of harsher winter conditions in Ontario;
(d) regulatory environment differences;
(e) accounting differences, such as different definitions of reporting categories; and
(f) differences in currency conversion factors not adequately captured by the use of OECD GDP purchasing power parities.
163 EI’s selection of outputs are based on three criteria referred to in the First EI Report at p 10. First, that the output aligns with the rule 6.5.6 opex objectives.
164 Secondly, that the output reflects a service provided to a customer rather than an activity undertaken by a DNSP which does not directly affect what the customer receives. As explained by EI, if an activity is undertaken by a DNSP which does not directly affect what its customers receive is included as an output, there is a risk that the DNSP would have an incentive to over-engage in the activity and not concentrate sufficiently on meeting its customers’ needs at an efficient cost.
165 Thirdly, that the output is significant. That is, as explained by EI, while a DNSP has a wide range of outputs, its costs are dominated by a few key outputs and only those key outputs should be included to keep the analysis manageable and to be consistent with the high level nature of economic benchmarking (eg a call centre’s operations are not normally a large part of a DNSP’s costs and so the centre’s performance is not normally included as an output in a DNSP economic benchmarking study).
166 Attachment 7 to the Ausgrid Draft Decision, provides (at pp 7-56 and 7-59) the following explanation (without footnotes) for EI’s selection of output specifications for its MTFP and MPFP models to which the AER had regard its Final Decision:
Economic Insights' preferred output specification for the MTFP and MPFP includes:
• Customer numbers
• Ratcheted maximum demand
• Circuit line length
• Energy throughput
• Reliability (measured as total customer minutes off supply).
… this specification takes into account the operating environment variable of customer density by including both customers and line length as outputs. It similarly includes some allowance for differences in energy density and demand density by including energy delivered and a measure of maximum demand as outputs. Further this specification includes reliability as an output.
The MTFP analysis uses opex and capital as inputs. In this analysis capital is split into five distinct components – subtransmission overhead lines, distribution overhead lines, subtransmission underground cables, distribution underground cables and transformers and other. Each input is measured in terms of its physical quantity. This measure of inputs aligns with Economic Insights' preferred input specification which is justified in our explanatory statement to our Guideline.
Several submissions on our draft benchmarking report said that we did not allocate an appropriate weight to line length. Economic Insights consider that the weighting for overhead lines is appropriate because it has been developed through a Leontief estimation of the cost function.
Some submissions also noted that Economic Insights' lines and cables input index for MTFP analysis might be multiplicative in nature placing a greater weighting on high voltage lines than is warranted. Economic Insights addressed this concern by creating separate input indexes for subtransmission and distribution lines. The weighting given to high voltage lines will not influence our alternative assessment techniques that examine the productivity of opex. These techniques, unlike MTFP, are not sensitive to the weighting given to individual capital inputs.
… … …
In addition to accounting for these factors in the model specification, Economic Insights tested the effect of the following operating environment factors on the MPFP scores in a second-stage regression analysis:
• customer numbers (to check whether additional scale effects are significant)
• customer, energy and demand network densities
• the share of underground cable length in total circuit kilometres
• the share of single stage transformation capacity in single stage plus the second stage of two stage transformation capacity at the zone substation level, and
• system average interruption duration index (SAIDI)146.
Economic Insights found, using these tests, that none of these variables are statistically significant in their effect on the MPFP scores. This indicates that the MPFP results have appropriately captured the effects of these variables.
167 Having regard to EI’s three output specification criteria outlined above and to output specifications used by PEG’s research in work for the Ontario Energy Board, EI selected the following outputs for the EI model:
(a) customer numbers;
(b) circuit length; and
(c) maximum ratcheted demand (as explained by EI in the First EI Report at p 11, this variable is simply the highest value of peak demand observed in the period up to the year in question for each DNSP which recognises the capacity that has actually been used to satisfy demand and gives the DNSP credit for this capacity in subsequent years, even though annual peak demand may be lower in subsequent years.);
168 The EI model also specifies the proportion of the DNSP circuits that are underground as opposed to aboveground, not as an output variable, but as an operating environment variable. As EI explained it:
Undergrounding: by including an operating environment variable for the proportion of underground cables in total line and cable length in our cost functions, we explicitly allow for the impact of this factor.
169 The following explanation (without footnotes) of the EI model’s specifications appears at p 7-26 of Attachment 7 to the Ausgrid Draft Decision:
The opex cost functions incorporate the significant output variables of customer numbers, circuit length, and ratcheted maximum demand. Unlike the MTFP model the opex cost function models do not include energy delivered and reliability. Economic Insights excluded energy delivered because it was highly correlated with ratcheted maximum demand. The estimated coefficients of either energy delivered or ratcheted maximum demand were generally insignificant in these models. Economic Insights found that the correlation coefficient between these two variables was larger than 0.99 and the behaviour of their coefficients was almost certainly a consequence of multicollinearity problems.
Hence Economic Insights excluded energy delivered. As energy delivered is highly correlated with ratcheted maximum demand the model will pick up the effect of energy delivered.
Reliability was not included because consistent reliability data is not available for the international distributors. We are comfortable with Economic Insights not including reliability in the econometric models. A primary driver of reliability performance is capital expenditure. Expenditure on maintenance may prevent outages. However, individual network outages lead to opex associated with rectifying the outages.
The opex cost function models also include the proportion of underground circuits as an operating environment factor. This is consistent with the MTFP analysis which has separate input indexes for overhead and underground lines. As expected the coefficient of this variable is negative. Underground cables will require less ongoing maintenance than overhead cables. Further, underground cables do not incur vegetation management costs.
170 The Second EI Report was commissioned by the AER to assist it with its application of economic benchmarking and to advise:
(a) whether the AER should make adjustments to base year opex for New South Wales, Australian Capital Territory and Queensland DNSPs based on the results from economic benchmarking models; and
(b) the productivity change to be applied to forecast opex for the NSW, ACT and Queensland DNSPs.
171 Referring to the First EI Report, the Second EI Report notes at p iv:
After choosing a conservative efficiency target based on the weighted average performance of the five top performing DNSPs and making additional allowance for factors not included in the econometric models, downwards adjustments were recommended for the base year opex of each of the NSW and ACT DNSPs.
The NSW and ACT DNSPs’ revised regulatory proposals included a number of supporting consultants’ reports critiquing the analysis in … [the First EI Report] …. These included reports by Pacific Economics Group Research, Frontier Economics, Cambridge Economic Policy Associates (CEPA), Advisian and Huegin. A number of consultants’ reports were also submitted by the Queensland DNSPs including ones by Frontier Economics, Huegin and Synergies.
We have reviewed both the critiques presented by the consultants and the alternative models presented in detail and have found no reason to change the approach adopted in … [the First EI Report] … benchmarking analysis. We do, however, consider there is a case for revising the opex efficiency target. And updated and more detailed information on the impact of operating environment factors not explicitly included in the opex cost function model is now available.
172 Expanding on the observation that there is a case for revising the opex efficiency target, the Second EI Report stated at p x:
… we are of the view there may be a case for setting an even more conservative target than that used in …[the First EI Report ] … . This is particularly the case given that this is the first time economic benchmarking is being used as the primary basis for an Australian regulatory decision.
… … …
Incorporating the more conservative efficiency target and updated information on the impact of operating environment factors not included in the econometric models produces the base year opex reductions listed in table A for the NSW, ACT and Queensland DNSPs. Since Endeavour Energy is already exceeding its (conservatively set) target, no adjustment to its base year opex is required.
173 Table A, as it appeared in the Second EI Report, is reproduced below:
Table A NSW, ACT and Queensland DNSP opex efficiency scores, adjusted efficiency targets and base year opex adjustments to reach the target
Target allowing for additional OEFs
Reduction to base year opex
174 In a nutshell, the Second EI Report resulted in no change to the AER’s application of the EI model and benchmarking methodology other than the AER’s (contentious):
(a) lowering of the comparison point for determining the AER’s alternative estimate of base opex in the EI model from Citi Power to AusNet; and
(b) further adjustments to operating environment factors (OEFs) as outlined below.
The lowering of the comparison point and the further OEF adjustments did result in the AER increasing its opex allowances for the DNSPs. While the increases were not enough to satisfy the DNSPs, they were, in PIAC’s submission, overly generous.
175 In its Draft Decisions, the AER used as the benchmark comparison point (ie each DNSP’s efficiency target) the weighted average of the top quartile of opex efficiency scores generated by the EI model for all DNSPs. This was the average of the efficiency scores of the five most efficient DNSPs, each of which was over 0.75.
176 In its Final Decisions, based on advice from EI, the AER used as a benchmark comparison point the opex efficiency score of the DNSP whose score was at the bottom of the upper third of the scores of all DNSPs (ie the DNSP with the lowest opex efficiency score above 0.75). This, as may be seen by reference to Table 7.4 that is reproduced below, is AusNet. It reduced the benchmark comparison point from 0.86 to 0.77.
177 The AER’s lowering of the comparison point is a matter of some controversy from PIAC’s perspective and is also criticised by the DNSPs as pointing to weaknesses in the AER’s approach to benchmarking. The controversy and criticism are canvassed below.
178 At page 7-180 of Attachment 7 to the Ausgrid Final Decision, the AER notes (without footnotes):
It is important to recognise that service providers do not operate under exactly the same operating environment factors (OEFs). OEFs may have a significant impact on measured efficiency through their impact on a service provider's opex. It is desirable to adjust for material OEF differences to ensure that when comparisons are made across service providers, we are comparing like with like to the greatest extent possible. By identifying the effect of OEFs on costs one can determine the extent to which cost differences are exogenous or due to inefficiency.
Some key OEFs are directly accounted for in Economic Insights’ SFA model. Where this has not been possible, we have considered the quantum of the impact of the OEF on the NSW service providers’ opex relative to the comparison firms. We have then adjusted the SFA efficiency scores based on our findings on the effects of OEFs.
Like paragraphs also appear in Attachment 7 to the other Final Decisions in issue in the DNSPs applications.
179 In its post-modelling adjustment process, the AER assessed the effects of 65 OEFs, being potential differences between DNSPs that were not directly accounted for in the model specification. This included factors nominated by the AER during the decision process and factors put forward by DNSPs following the issue of the Draft Decisions.
180 The assessment was undertaken in two stages. First, the AER assessed each OEF against three OEF criteria, namely, exogeneity, materiality, and duplication as explained by the AER in the following paragraphs.
181 The first, exogeneity, is that an OEF should be outside the control of a DNSP because adjusting for an OEF that a DNSP can control itself may mask inefficient investment or expenditure.
182 Although, as explained further below, a collective adjustment was made for individually immaterial factors, the second criterion, materiality, is that an OEF should create a material difference in a particular DNSP’s opex. An OEF was considered to be material where it would affect a DNSP’s opex by 0.5 percent or more.
183 To avoid double-counting the effects of an OEF, the third criterion, duplication, is that the OEF should not have been accounted for elsewhere.
184 Where an OEF satisfied all three criteria in relation to a particular DNSP, the AER made an adjustment to the DNSP’s target opex to allow for its effects.
185 The second stage involved the AER providing an additional single adjustment for each DNSP to account for factors that satisfied the exogeneity and duplication criteria but did not independently have a material effect on opex.
186 In order to determine whether an OEF was likely to have a material effect on a DNSP’s opex, the AER assessed all available information, including information provided by the DNSPs.
187 The AER identified four OEFs that fulfilled all three of the material OEF criteria for the Networks NSW DNSPs: subtransmission configurations, licence conditions, occupational health and safety regulations and termite exposure.
188 For ActewAGL, the OEFs that satisfied the materiality criteria were backyard reticulation, capitalisation practices, occupational health and safety regulations and standard control services connections.
189 The AER made adjustments to each DNSP’s target efficiency scores to take account of the above mentioned material OEFs.
190 The AER then estimated the collective effect of the OEFs that had been found to be exogenous and non-duplicative, but not individually material (immaterial OEFs).
191 Where the AER considered that an immaterial OEF was likely to disadvantage a DNSP, or where it was uncertain whether the OEF would advantage or disadvantage a DNSP (directionally ambiguous OEFs), the AER allowed 0.5 percent in the DNSP’s favour. Where an immaterial OEF was likely to advantage the DNSP, the AER subtracted 0.5 percent. There was one exception to this procedure: where the AER was able to quantify the effect of an immaterial OEF, it made an adjustment only for that amount.
192 The AER then made a further adjustment to each DNSP’s target efficiency scores to account for the collective effects of the immaterial OEFs.
193 Table A 6 Summary of final decision on OEF adjustments in Attachment 7 to the Ausgrid Final Decision details the OEF adjustments made to each of the Networks NSW DNSPs’ target opex and is reproduced below:
Table A. 6 Summary of final decision on OEF adjustments
Reasons against OEF criteria551
The boundary between distribution and transmission is not determined by service providers
Data from Ausgrid's regulatory accounts suggest that subtransmission assets are up to twice as costly to operate as distribution assets.
Economic Insights' SFA model does not include a variable that accounts for subtransmission assets.
The network planning requirements in the NSW service providers licence conditions are not determined by service providers.
Category analysis and economic benchmarking RIN data suggest that the increased transformer capacity to meet the 2005 and 2007 change in licence conditions may lead to a material increase in maintenance expenditure.
Economic Insights' SFA model does not include a variable that accounts for changes in licence conditions.
OH&S regulations are not set by service providers.
Data from the ABS and a PwC report commissioned by the Victorian Government suggest that differences in OH&S regulations may materially affect service provider's opex.
Economic Insights' SFA model does not include a variable that accounts for differences in OH&S legislation.
The prevalence of termites in a geographic area is beyond service providers’ control.
Data on Powercor’s termite management costs and data from the CSIRO on the range of termites suggest that the Essential Energy may have a material cost disadvantage due to termite exposure.
Economic Insights’ SFA model does not include a variable that accounts for differences in termite exposure.
There are various exogenous, individually immaterial factors not accounted for in Economic Insights' SFA model that may affect service providers' costs relative to the comparison firms. While individually these costs may not lead to material differences in opex, collectively they may.
The AER said in its footnote to that Table that its OEF criteria exogeneity, materiality, and duplication, are explained in detail in its section on its approach to OEFs.
194 Table A 8 Summary of individually immaterial OEF adjustment in Attachment 7 to the Ausgrid Final Decision provides a summary of the quantification of the effect of immaterial factors on each of the three Networks NSW DNSPs. It is also reproduced below:
Table A 8 Summary of individually immaterial OEF adjustment
Cultural heritage obligations
Extreme weather events
Proportion of 11kV and 12kV lines
Rainfall and humidity
Transformer capacity owned by customers
Division of vegetation management responsibility
Source: AER analysis
The totals do not reconcile entirely due to rounding
195 Table A.6 Summary of final decision on OEF adjustments in Attachment 7 to the ActewAGL Final Decision details the OEF adjustments made to its target opex and is reproduced below.
Table A. 6 Summary of final decision on OEF adjustments
Reasons against OEF criteria
Although capitalisation practices are the result of management decisions, differences in capitalisation practices can lead to material differences that are unrelated to efficiency.
ActewAGL's capitalisation practices, with regard to vehicle and IT costs, provide it with a material cost disadvantage relative to the comparison firms.
Economic Insights' SFA model does not include variables that account differences in capitalisation practices between the NEM service providers.
Backyard reticulation has been required by ACT planning approaches.
ActewAGL has provided evidence that backyard reticulation materially increases its costs.
Economic Insights' SFA model does not include variables that account for backyard reticulation between the NEM service providers.
Standard control services connections
The AER determines service providers' service classifications.
Standard control services connections opex accounts for a material amount of ActewAGL's standard control services opex.
Economic Insights' SFA uses network services data. Connection services are not included in network services.
OH&S regulations are not set by service providers.
Data from the ABS and a PwC report commissioned by the Victorian Government suggest that differences in OH&S regulations may materially affect service provider's opex.
Economic Insights' SFA model does not include a variable that accounts for differences in OH&S legislation.
Individually immaterial factors
There are various exogenous, individually immaterial factors not accounted for in Economic Insights' SFA model that may affect service providers' costs relative to the comparison firms. While individually these costs may not lead to material differences in opex, collectively they may.
Source: AER analysis
The AER points out that the OEF criteria exogeneity, materiality, and duplication, are explained in detail in its section on its approach to OEFs.
196 Table A.8 Summary of individually immaterial OEF adjustments in Attachment 7 to the ActewAGL Final Decision details the quantified effect of immaterial factors and is reproduced below.
Table A 8 Summary of individually immaterial OEF adjustment
Cultural heritage obligations
Extreme weather events
Humidity and rainfall
Proportion of 11kV and 12kV lines
Transformer capacity owned by customer
Source: AER analysis
197 The AER’s OEF adjustments are also a matter of controversy. PIAC challenges them as being too generous. The DNSPs challenging them as subjective and arbitrary. The challenges are canvassed below.
198 Figure 7.2: Our assessment approach in Attachment 7 to the Ausgrid Final Decision outlines the AER’s five step approach to forming its alternative estimate of opex as follows:
Step 1 –Start with service provider’s opex
We typically use the service provider’s actual opex in a single year as the starting point for our assessment. We call this the base year. While categories of opex can vary year to year, total opex is relatively recurrent. We typically choose a recent year for our assessment.
Step 2 – Assess base year opex
We assess whether opex the service provider incurred in the base year reasonably reflects the opex criteria. We have a number of techniques including economic benchmarking by which we can test the efficiency of opex in the base year.
Step 3 – Add a rate of change to base opex
As the opex of an efficient service provider tends to change over time due to price changes, output and productivity we trend our estimate of base opex forward over the regulatory control period to take account of these changes. We refer to this as the rate of change.
Step 4 – Add or subtract any step changes
We then adjust base year expenditure to account for any forecast cost changes over the regulatory control period that would meet the opex criteria that are not otherwise captured in base opex or rate of change. This may be due to new regulatory obligations in the forecast period and efficient capex/opex trade-offs. We call these step changes.
Step 5 – Other opex
Finally we add any additional opex components which have not been forecast using this approach. For instance, we forecast debt raising costs based on the costs incurred by a benchmark efficient service provider.
Having established our estimate of total forecast opex we can compare our alternative opex forecast with the service provider’s total forecast opex. If we are not satisfied there is an adequate explanation for the difference between our opex forecast and the service provider's opex forecast, we will use our opex forecast.
The AER refer to it as a “revealed cost method” in its EFA Guideline (and sometimes refer to it as the “base-step-trend method” in its past regulatory decisions).
199 Table 7.3: Assessment of Ausgrid’s base opex in Attachment 7 to the Ausgrid Final Decision outlines the main techniques used by the AER to test the efficiency of Ausgrid’s base opex.
Table 7.3 Assessment of Ausgrid’s base opex
Description of technique
Economic benchmarking measures the efficiency of a service provider in the use of its inputs to produce outputs.
The economic benchmarking techniques we used to test Ausgrid's efficiency included Multilateral Total Factor Productivity, Multilateral Partial Factor Productivity and opex cost function modelling. We compared Ausgrid's efficiency to other service providers in the NEM.
Despite differences in the techniques we used, all benchmarking techniques show Ausgrid does not perform as efficiently as most other service providers in the NEM.
We consider that differences in Ausgrid's operating environment not captured in the benchmarking models do not adequately explain the different benchmarking results between Ausgrid and other service providers.
Labour costs represent a large proportion of all NSW service providers' opex. We engaged Deloitte Access Economics (Deloitte) to review the NSW service providers' labour and workforce practices.
Deloitte found that because of labour and workforce management issues, Ausgrid's base year would not likely represent efficient costs.
Deloitte concludes that:
the NSW service providers have high labour costs because they have too many employees. They all engaged permanent staff in preference to contractors over the 2009–14 period for transitory capex work. Now, due to EBA restrictions on redundancies, they have stranded labour
because the NSW service providers employ a high proportion of their employees through EBAs (more than 75 percent) restrictive EBA clauses have a significant impact on workforce flexibility
the optimum level of outsourcing is likely to be higher than the level the NSW service providers outsourced at over the 2009–14 period; this is a key distinguishing factor from the Victorian service providers
while the NSW service provider have been implementing efficiency improvements, many efficiencies have not been realised until after the 2012–13 base year.
Source: AER analysis
The reference to Deloitte is to Deloitte Access Economics (Deloitte). A footnote to Table 7.3 cited NSW distribution network service providers labour analysis: addendum to 2014 report, April 2015, pp. ii–vii; and Deloitte NSW Distribution Network Service Providers Labour Analysis, November 2014, p. iv.
200 A like Table 7.3 appeared in each of the Attachments 7 to the AER’s Endeavour, Essential and ActewAGL Final Decisions. Subject only to the name of the relevant DNSP, the “Economic benchmarking” paragraphs in each of those tables are the same. Subject to an additional bullet point in the case of Endeavour and an additional paragraph in the case of Essential, so too are the “Review of labour and workforce practices” paragraphs in each of the Endeavour and Essential tables.
201 The additional bullet point in the case of Endeavour is as follows:
• Deloitte considered that Endeavour Energy’s base year opex was likely more efficient than Ausgrid’s and Essential Energy’s because it had commenced implementing efficiency improvements earlier. However, all NSW service providers (including Endeavour Energy) had efficiencies they were yet to realise because the reforms they had implemented to date did not consider potential opportunities to improve efficiency outside of the three NSW businesses. That is, they compared efficiency among themselves, but not to businesses in other jurisdictions.
202 The additional paragraph in the case of Essential is as follows (without footnote):
Further, in response to submissions in its revised proposal about the adverse impact of the dispersed nature of its network on labour costs, Deloitte found that Essential Energy could potentially achieve significant cost savings by implementing a local service agent (LSA) model. Powercor achieved significant efficiencies from implementing an LSA model following privatisation.
203 Table 7.3 in Attachment 7 to the Essential Final Decision also had the additional following entry relating to “Vegetation management”:
Description of technique
Essential Energy's vegetation management costs have increased significantly over the 2009–14 period. Category analysis showed Essential Energy has very high costs compared to most of its peers and Essential Energy’s regulatory proposal included a step down in vegetation management for the forecast period acknowledging that its 2009–14 practices required reform. Therefore, we decided to review Essential Energy’ vegetation management practices in detail.
Our overall findings for vegetation management remain the same as those from our draft decision. That is, Essential Energy's own documentation, including a report it commissioned from Select Solutions, provide evidence that its vegetation management practices in the base year (2012–13) were inefficient.41
Select Solutions' review found that Essential Energy must move to a "significantly more efficient" vegetation management model to reduce the impact of its expenditure on customer prices.42 Select Solutions found several causes of inefficiency, including:
attributing too much vegetation management effort to reactive spot clearing rather than proactive cyclic maintenance
primarily engaging contractors for cutting on a demonstrably less efficient hourly rate basis
less than optimal outsourcing.
We discuss our vegetation management findings in more detail in Appendix A.5.
Source: AER analysis
The footnoted references are to Essential, Regulatory Proposal, 2014, p. 73 and its paper Essential Energy, Vegetation Management Strategy and Implementation Plan for Additional Expenditure – FY 2013 to 14, February 2013; and to Select Solutions, Review of Essential Energy Vegetation Management Strategy–Final Report, 22 March 2013.
204 While the “Economic benchmarking” paragraphs in Table 7.3 in the ActewAGL Final Decision are the same as those in the Networks NSW DSNP’s equivalent tables, the “Review of labour and workforce practices” and “Review of vegetation management” paragraphs in Table 7.3 ActewAGL are different, as reproduced below:
Description of technique
Labour costs represent a large proportion of ActewAGL’s opex (approximately 80 percent). Category analysis showed ActewAGL had high labour costs relative to most of its peers and ActewAGL’s regulatory proposal suggested labour costs were a reason ActewAGL overspent its opex allowance in 2012–13.
Therefore, we decided, with the assistance of EMCa, to conduct a detailed review of ActewAGL’s labour and workforce practices.
EMCa considered that there is evidence that ActewAGL’s work practices, processes and systems in 2012–13 were ineffective. EMCa considered that this lead to inefficient use of labour in the office and field. This inefficiency is characterised by duplication of effort in work planning and scheduling, loss of field productivity through ineffective works management and through ineffective data and information management.
EMCa also considered that ActewAGL’s labour levels were not reasonably efficient in 2012–13, noting that ActewAGL has steadily increased its ASL based on assumed future growth scenarios and adopting an internal resourcing strategy.
EMCa considered that if ActewAGL had outsourced more of its work, it would likely have benefited from increased labour flexibility and reduced operating costs.
EMCa found a lack of compelling evidence to demonstrate that ActewAGL’s labour costs in 2012–13 were reflective of an efficient service provider. EMCa consider this was evident by the relatively high level of internal resources used and the extent to which work was outsourced on an hourly rate bases for the urgent clearance of vegetation..
ActewAGL's vegetation management costs have increased significantly over the 2009–14 period. Category analysis showed ActewAGL has very high costs compared to most of its peers and ActewAGL's regulatory proposal suggested vegetation management was a reason ActewAGL overspent its opex allowance in 2012–13. Therefore, we decided, with the assistance of EMCa, to review ActewAGL's vegetation management practices in detail.
EMCa found that ActewAGL did not act prudently and efficiently to manage costs associated with increased vegetation growth that occurred prior to 2012–13 because its vegetation management practices and its strategic and tactical responses were inadequate.
EMCa also found evidence of inefficient vegetation management costs in 2012–13 due to the manual processes between the office and field and the extent of clearance work that was deemed to be urgent, and which was therefore undertaken with a resultant higher cost. It is EMCa’s view that a service provider acting to efficiently minimise costs would have incurred a lower level of urgent clearance work.
Source: AER analysis.
205 It appears from Attachment 7 to the final decisions that:
(a) the AER used the EI model to adjust Ausgrid, Essential and ActewAGL’s base opex to determine a starting point for a forecast that it considered would reasonably reflect the criteria; and
(b) while the AER considered that it was unable to use Ausgrid, Essential or ActewAGL’s historical opex to arrive at an alternative forecast because such a forecast would not result in a forecast that would reasonably reflect the opex criteria, it was not satisfied that a forecast based on Endeavour’s actual opex in 2012–13 could be regarded as materially inefficient.
206 While in relation to Endeavour, the AER concluded it was not satisfied that a forecast based on Endeavour’s actual opex in 2012–13 could be regarded as materially inefficient, it formed the view that Endeavour’s forecast opex involved two additional items which it analysed (but rejected) as “step changes” (ie additional expenditure not incurred by Endeavour during the base year), namely:
(a) $240.7m ($2013-14) in respect of increased vegetation management; and
(b) $17.3m ($2013-14) in redundancy costs.
207 As is illustrated by the following passage (without footnotes) from Attachment 7 to the Endeavour Final Decision at pp 7-268 to 7-269, in rejecting what it described as Endeavour’s “step changes”, the AER relied on the EI model:
For Endeavour Energy's base year opex, because we are not satisfied that it contains material inefficiency it does not require an adjustment. We, therefore, consider it appropriate to use Endeavour Energy's base opex when developing our alternative forecast. This is a departure from our draft decision. Our benchmarking analysis is nevertheless relevant in our assessment of other components of Endeavour Energy's alternative total opex forecast, such as a consideration of its proposed step changes.
We disagree with the service providers’ submissions that advocate we should abandon our benchmarking techniques and the extent to which we rely upon our benchmarking results. Therefore, we continue to place significant weight on the results of Economic Insights’ preferred [EI] model (Cobb Douglas SFA) in estimating necessary reductions in base opex.
208 More particularly, the AER rejected the increased vegetation management opex which Endeavour proposed to enable it to meet increased outsourced providers’ contract prices to be incurred in targeting improvements to conform with the minimum risk standards it is obliged to meet regarding the clearance distance between its mains and vegetation across its network area.
209 In rejecting the increase the AER observed (at p 7-288 of Attachment 7 to the Endeavour Final Decision) that:
(a) Endeavour had stated that it did not face any change to the minimum risk standards with which it must comply;
(b) without persuasive evidence that a DNSP’s total historical opex was too low to achieve the opex objectives, it did not consider increased contract costs to be a reason to increase the total opex forecast to meet what are unchanged regulatory obligations;
(c) Endeavour does not benchmark well when compared to other DNSPs in the national electricity market; and
(d) it would be inconsistent with the application of Endeavour’s EBSS to include the vegetation management opex in the opex forecast (Endeavour proposed that it retain gains from its EBSS rather than share them with its customers).
210 The AER rejected the proposed redundancy expenditure because it considered that it was needed only because Endeavour was not currently operating as efficiently as it could. In this respect it relied on the April 2015 Deloitte report referred to above (Deloitte Access Economics, NSW distribution network service providers labour analysis: addendum to 2014 report, April 2015) concerning labour and workforce management issues affecting Networks NSW. That Deloitte report in turn relied on the EI Model.
211 Table 7.4: Arriving at our alternative estimate of base opex in the Ausgrid Final Decision outlines the steps that the AER took to arrive at Ausgrid’s base opex.
Table 7.4 Arriving at our alternative estimate of base opex
Step 1 – Start with Ausgrid's average opex over the 2006 to 2013 period.
Ausgrid's network services opex was, on average, $509.3 million ($2013) over the 2006 to 2013 period.
Step 2 - Calculate the raw efficiency scores using our preferred economic benchmarking model
Our preferred economic benchmarking model is Economic Insights’ Cobb Douglas SFA model. We use it to determine all service providers' raw efficiency scores.
Based on Ausgrid's customer numbers, line length, and ratcheted maximum demand over the 2006 to 2013 period, Ausgrid's raw efficiency score is 44.7 percent.
Step 3 – Choose the comparison point
For the purposes of determining our alternative estimate of base opex, we did not base our estimate on the efficient opex estimated by the model.
The comparison point we used was the lowest performing service provider in the top quartile of possible scores, AusNet Services. According to this model AusNet Services' opex is 76.8 percent efficient based on its performance over the 2006 to 2013 period. Therefore to determine our substitute base we have assumed a prudent and efficient Ausgrid would be operating at an equivalent level of efficiency to AusNet Services.
Step 3 – Adjust Ausgrid's raw efficiency score for operating environment factors
The economic benchmarking model does not capture all operating environment factors likely to affect opex incurred by a prudent and efficient Ausgrid.
We have estimated the effect of these factors and made a further reduction to our estimate where required. We have determined an 11.7 percent reduction to Ausgrid's comparison point based on our assessment of these factors.
A material operating environment factor we considered was not accounted for in the model is the different subtransmission configurations in NSW.
= 0.768 /
(1 + 0.117)
Step 4 – Calculate the percentage reduction in opex
We then calculate the opex reduction by comparing Ausgrid's efficiency score with the adjusted comparison point score.
= 1 – (0.447 /
Step 5 – Calculate the midpoint efficient opex
We estimate efficient opex at the midpoint of the 2006 to 2013 period by applying the percentage reduction in opex to Ausgrid's average opex over the period.
This represents our estimate of efficient opex at the midpoint of the 2006 to 2013 period.
= (1 – 0.350)*
Step 6 – Trend midpoint efficient opex forward to 2012–13
Our forecasting approach is to use a 2012–13 base year. We have trended the midpoint efficient opex forward to a 2012–13 base year based on Economic Insights’ opex partial factor productivity growth model. It estimates the growth in efficient opex based on growth in customer numbers, line length, ratcheted maximum demand and share of undergrounding.
It estimated the growth in efficient opex based on Ausgrid’s growth in these inputs in this period to be 8.48 percent.
= 330.9 x (1 =
Step 7 – Adjust our estimate of 2012–13 base year opex for CPI
The output in step 6 is in real 2013 dollars. We need to convert it to real 2013–14 dollars for the purposes of forming our substitute estimate of base opex. This reflects one and a half years of inflation. This is our estimate of base opex.
= 359.0 x (1 +
Source: AER analysis
212 As may be seen from Table 7.4 reproduced above, the AER started with Ausgrid’s average opex over the period 2006 to 2013, namely, $509.3m.
213 It then, in Step 2, used the EI model to calculate a raw efficiency score of 44.7 percent.
214 The third step involved the choice of a comparison point of 76.8 percent, based on AusNet, which is at the bottom of the top quartile of possible scores using the EI model.
215 The fourth step (described as Step 3 in Table 7.4) involved an adjustment to the raw efficiency score for OEFs – the adjustment being made to the comparison point on the assumption that the comparator (AusNet) would have to contend with the same environmental factors as Ausgrid. This adjustment, made by the AER once the model had been run, lowers the comparison point to 68.7 percent.
216 The mathematical consequence of this “after-the-event” adjustment, is that the efficiency score of Ausgrid becomes 44.7 percent and, therefore, in what is described as Step 4 in Table 7.4, it is calculated that Ausgrid is, in effect, 35 percent below the comparison point.
217 Then in what is described as Step 5 in Table 7.4, the AER calculated a mid-point efficient opex (ie the average efficient opex over the 2006-13 period based on the EI model). In Step 6 adjusted that mid-point forwards to 2012-13 following that adjustment in Step 7 with an adjustment of its estimate of Ausgrid’s 2012-13 base year opex for consumer price index (CPI) to arrive at a figure of $374.2m ($2013-14).
218 A table similar to the Table 7.4: Arriving at our alternative estimate of base opex which appeared in Attachment 7 to the Ausgrid Final Decision as reproduced above also appeared in each of the Attachments 7 to the AER’s Essential and ActewAGL Final Decisions. Those tables also included two steps designated as Step 3. The table in the Essential Final Decision also included an additional Step 8 as reproduced below:
Step 8- Convert to
final year estimate
The guideline specifies that we will convert our estimate of base year opex into a final year estimate.
We used the formula in the guideline to determine our unadjusted estimate of opex for 2013–14.
We used the 2012–13 efficient opex value in step 7 to determine what the efficiency adjustment would be for 2012–13 (–26.3%), taking into account changes to Essential Energy's service classification.
To arrive at our adjusted final year estimate, we applied the efficiency adjustment to our unadjusted estimate of 2013–14 opex.
See AER opex model
for Essential Energy
Source: AER analysis
219 It may be seen from Table 7.4 in Attachment 7 to the Essential Final Decision that:
(a) the AER started with Essential’s average opex over the period 2006 to 2013, namely, $352.5m ($2013);
(b) it then, in Step 2, used the EI model to calculate a raw efficiency score of 54.9 percent;
(c) the third step involved the choice of a comparison point of 76.8 percent, based on AusNet which is at the bottom of the top quartile of possible scores using the EI model;
(d) the fourth step (described as Step 3 in the table) involved an adjustment to the raw efficiency score for OEFs – the adjustment being made to the comparison point on the assumption that the comparator (AusNet) would have to contend with the same environmental factors as Essential. This adjustment, made by the AER once the model had been run, lowers the comparison point to 69.4 percent;
(e) the mathematical consequence of this “after-the-event” adjustment, is that the efficiency score of Essential is calculated in what is described as Step 4 in the table at 20.9 percent below the comparison point;
(f) then in what is described as Step 5 in the table, the AER calculated a mid-point efficient opex ie, the average efficient opex over the 2006-13 period based on the EI model. It then in Step 6 adjusted that mid-point forwards to 2012-13 following that adjustment in Step 7 with an adjustment of its estimate of Essential’s 2012-2013 base year opex for CPI to arrive at a figure of $308.2m ($2013-2014); and
(g) finally, in what is described as Step 8 in the table, the AER converts its estimate of Essential’s base year opex into a final year estimate of $311.9m ($2013-14).
220 As observed above, because the AER was not satisfied that a forecast based on Endeavour’s actual opex could be regarded as materially inefficient, the AER did not undertake the steps taken outlined in Table7.4 to each of the Ausgrid and Essential Final Decisions to arrive at their base year opex to arrive at a base year opex for Endeavour.
221 The AER did, however, undertake those steps to arrive at ActewAGL’s base year opex. It may be seen by reference to Table 7.4 in Attachment 7 to the ActewAGL Final Decision that:
(a) the AER started with ActewAGL’s average opex over the period 2006 to 2013, namely, $59.9m ($2013);
(b) it then, in Step 2, used the EI model to calculate a raw efficiency score of 39.9 percent;
(c) the third step involved the choice of a comparison point of 76.8 percent, based on AusNet which is at the bottom of the top quartile of possible scores using the EI model;
(d) the fourth step (described as Step 3 in the table) involved an adjustment to the raw efficiency score for OEFs – the adjustment being made to the comparison point on the assumption that the comparator (AusNet) would have to contend with the same environmental factors as ActewAGL. This adjustment, made by the AER once the model had been run, lowers the comparison point to 62.4 percent;
(e) the mathematical consequence of this “after-the-event” adjustment, is that the efficiency score of ActewAGL is calculated in what is described as Step 4 in the table at 36.2 percent below the comparison point; and
(f) then in what is described as Step 5 in the table, the AER calculated a mid-point efficient opex (ie, the average efficient opex over the 2006-2013 period based on the EI model). It then in Step 6 adjusted that mid-point forwards to 2012-2013 following that adjustment in Step 7 with an adjustment of its estimate of ActewAGL’s 2012-2013 base year opex for CPI to arrive at a figure of $45.1m ($2013-14).
222 Table A.1: Final determination estimates of efficient base year opex ($million 2013–14) in Attachment 7 to the Ausgrid Final Decision sets out the AER’S final determination estimates of base year opex for each Networks NSW DNSP as follows.
Table A.1 Final determination estimates of efficient base year opex ($million 2013-14)
Revealed base opex (adjusted)a
AER base opex
Percentage base opex reduction
Note: (a) This number is the revealed 2012–13 opex, so it differs from the starting number in Table 7.4, which is average opex over 2006–13. We have adjusted the service providers’ revealed opex for debt raising costs, new CAM [cost allocation method] (if applicable) and new service classifications.
(b) Our estimate of base opex for Endeavour is slightly higher than Endeavour's because the reduced benchmark comparison point means its revealed costs are lower than the efficiency target.
Source: AER analysis.
223 In a paragraph following Table A.1 in the Ausgrid Final Decision, the AER observed at pp 7-52 to 7-53 that:
Our reduction to Endeavour Energy's revealed opex was lower than that for Essential Energy and Ausgrid. Our analysis showed that Endeavour Energy had implemented efficiency programs earlier and to a greater extent than its two peers. However, we considered that as at 2012–13 (the base year), Endeavour Energy had further efficiency improvements to realise.
224 A like table in the ActewAGL Final Decision is reproduced below:
Table A.1 Final determination estimates of efficient base year opex ($million 2013-14)
Revealed base opex (adjusted)a
AER base opex
Percentage base opex reduction
Note: (a) This number is the revealed 2012–13 opex, so it differs from the starting number in Table 7.4, which is average opex over 2006–13. We have adjusted ActewAGL’s revealed opex for debt raising costs, new CAM (if applicable) and new service classifications.
Source: AER analysis.
225 Table 7.7: Our consideration of opex factors which appears in Attachment 7 to the Ausgrid Final Decision, reproduced below provides, a convenient summary of the AER’s application of the benchmarking opex factor in r 6.5.6(e)(4) and the opex factors in rr 6.5.6(e)(5), (5A), (6), (7), (8), (9A) and (10):
Table 7.7 Our consideration of opex factors
The most recent annual benchmarking report that has been published under rule 6.27 and the benchmark operating expenditure that would be incurred by an efficient Distribution Network Service Provider over the relevant regulatory control period.
There are two elements to this factor. First, we must have regard to the most recent annual benchmarking report. Second, we must have regard to the benchmark operating expenditure that would be incurred by an efficient distribution network service provider over the period. The annual benchmarking report is intended to provide an annual snapshot of the relative efficiency of each service provider.
The second element, that is, the benchmark operating expenditure that would be incurred an efficient provider during the forecast period, necessarily provides a different focus. This is because this second element requires us to construct the benchmark opex that would be incurred by a hypothetically efficient provider for that particular network over the relevant period.
We have used several assessment techniques that enable us to estimate the benchmark opex that an efficient service provider would require over the forecast period. These techniques include economic benchmarking, opex cost function modelling, category analysis and a detailed review of Ausgrid's labour and workforce practices. We have used our judgment based on the results from all of these techniques to holistically form a view on the efficiency of Ausgrid's proposed total forecast opex compared to the benchmark efficient opex that would be incurred over the relevant regulatory control period.
The actual and expected operating expenditure of the distribution network service provider during any proceeding regulatory control periods.
Our forecasting approach uses the service provider's actual opex as the starting point. We have compared several years of Ausgrid's actual past opex with that of other service providers to form a view about whether or not its revealed expenditure is sufficiently efficient to rely on it as the basis for forecasting required opex in the forthcoming period.
The extent to which the operating expenditure forecast includes expenditure to address the concerns of electricity consumers as identified by the distribution network service provider in the course of its engagement with electricity consumers.
We understand the intention of this particular factor is to require us to have regard to the extent to which service providers have engaged with consumers in preparing their regulatory proposals, such that they factor in the needs of consumers.103 We have considered the concerns of electricity consumers as identified by Ausgrid in assessing its proposal – particularly those expressed in the consumer-focussed overview provided as an attachment to its regulatory proposal. For example, a clear theme present in this document is that customers consider electricity prices are too high.104
The relative prices of capital and operating inputs
We have considered capex/opex trade-offs in considering step changes for Ausgrid's head office building and for demand management expenditure. We considered the relative expense of capex and opex solutions in considering these step changes.
We have had regard to multilateral total factor productivity benchmarking when deciding whether or not forecast opex reflects the opex criteria. Our multilateral total factor productivity analysis considers the overall efficiency of networks in the use of both capital and operating inputs with respect to the prices of capital and operating inputs.
The substitution possibilities between operating and capital expenditure.
As noted above we considered capex/opex trade-offs in considering step changes for Ausgrid's head office building and for demand management expenditure. We considered the substitution possibilities in considering these step changes.
Some of our assessment techniques examine opex in isolation – either at the total level or by category. Other techniques consider service providers' overall efficiency, including their capital efficiency. We have relied on several metrics when assessing efficiency to ensure we appropriately capture capex and opex substitutability.
In developing our benchmarking models we have had regard to the relationship between capital, opex and outputs.
We also had regard to multilateral total factor productivity benchmarking when deciding whether or not forecast opex reflects the opex criteria. Our multilateral total factor productivity analysis considers the overall efficiency of networks with in the use of both capital and operating inputs.
Further, we considered the different capitalisation policies of the service providers and how this may affect opex performance under benchmarking.
Whether the operating expenditure forecast is consistent with any incentive scheme or schemes that apply to the distribution network service provider under clauses 6.5.8 or 6.6.2 to 6.6.4.
The incentive scheme that applied to Ausgrid's opex in the 2009–14 regulatory control period, the EBSS, was intended to work in conjunction with a revealed cost forecasting approach.
In this instance, we have forecast efficient opex based on benchmark efficient service provider. We have considered this in deciding how the EBSS should apply to Ausgrid in the 2009–14 regulatory control period and the 2014–19 period.
The extent the operating expenditure forecast is referable to arrangements with a person other than the distribution network service provider that, in our opinion, do not reflect arm's length terms.
Some of our techniques assess the total expenditure efficiency of service providers and some assess the total opex efficiency.
Given this, we are not necessarily concerned whether arrangements do or do not reflect arm's length terms. A service provider which uses related party providers could be efficient or it could be inefficient. Likewise, for a service provider who does not use related party providers. If a service provider is inefficient, we adjust their total forecast opex proposal, regardless of their arrangements with related providers.
Whether the operating expenditure forecast includes an amount relating to a project that should more appropriately be included as a contingent project under clause 6.6A.1(b).
This factor is only relevant in the context of assessing proposed step changes (which may be explicit projects or programs). We did not identify any contingent projects in reaching our final decision.
The extent the distribution network service provider has considered, and made provision for, efficient and prudent non-network alternatives.
We have not found this factor to be significant in reaching our final decision.
Source: AER analysis
226 Table 7.8: Other factors we have had regard to (reproduced below) in the Ausgrid Final Decision summarises other factors that the AER considered relevant and, pursuant to r 6.5.6(e)(12), notified each DNSP of prior to the DNSP submitting its revised regulatory proposal.
Table 7.8 Other factors we have had regard to
Our benchmarking data sets, including, but not necessarily limited to:
1. data contained in any economic benchmarking RIN, category analysis RIN, reset RIN or annual reporting RIN
2. any relevant data from international sources
3. data sets that support econometric modelling and other assessment techniques consistent with the approach set out in the Guideline
as updated from time to time.
This information may potentially fall within opex factor (4). However, for absolute clarity, we are using data we gather from NEM service providers, and data from service providers in other countries to provide insight into the benchmark operating expenditure that would be incurred by an efficient and prudent distribution network service provider over the relevant regulatory period.
Economic benchmarking techniques for assessing benchmark efficient expenditure including stochastic frontier analysis and regressions utilising functional forms such as Cobb Douglas and Translog.
This information may potentially fall within opex factor (4). For clarity, and consistent with our approach to assessment set out in the Guideline, we are have regard to a range of assessment techniques to provide insight into the benchmark operating expenditure that an efficient and prudent service provider would incur over the relevant regulatory control period..
Source: AER Analysis
227 Networks NSW and ActewAGL’s main submissions addressing the principal issue whether the AER’s application of the EI model discharged its obligations under rr 6.5.6 and 6.12.1(4) are addressed below under the following headings:
(a) inadequacies in the EI model’s data set and comparability issues;
(b) the AER’s lowering of the EI model’s comparison point;
(c) the AER’s OEF adjustments;
(d) the efficiency of the DNSPs’ vegetation management costs; and
(e) the AER’s use of the EI model as the sole determinative of opex;
228 PIAC also addressed the principal issue, contending in particular that the AER had erred in its application of the EI model by its lowering of the EI model’s comparison point and in its OEF adjustments. The DNSP interveners broadly speaking supported the position taken by Networks NSW and ActewAGL, but it is noted that Ergon made a more refined criticism of the AER’s use of historic costs.
229 Other opex issues identified in the AER’s written opex submissions, addressed to the extent necessary having regard to the Tribunal’s conclusion on the principle issue, are:
(a) did the AER fail to corroborate the EI model’s results;
(b) did the AER’s have proper regard to endogenous circumstances;
(c) did the AER have proper regard to the consequences;
(d) alleged errors with respect to partial performance indicators (PPIs);
(e) alleged errors with respect to labour costs; and
(f) alleged errors relating to average or current efficiency.
230 The DNSPs’ submissions under this heading address what they see as a twofold weakness in the EI model’s data set. First, reliance on the Australian RIN benchmarking data. Secondly, the augmentation of the Australian RIN data with overseas data.
231 Networks NSW contends that the first weakness stems from the AER’s failure to collect data pursuant to its RINs in a manner sufficiently rigorous to be useful for benchmarking. It submits that the RINs were unclear and the requests for eight years of data resulted in:
(a) different DNSPs recording data differently; and
(b) some DNSPs estimating and backcasting some data.
232 Responding to the submission of a lack of clarity in the RINs resulting in DNSPs recording data differently, the AER:
(a) points to the extensive consultation between it and the DNSPs to ensure that the DNSPs understood what was required of them; and
(b) contends that the process of consulting with stakeholders and making necessary clarifications was part of the AER’s planned procedure for ensuring that the RINs were clear and the instructions and definitions were sufficient.
233 Support for the AER’s contention is to be found in Table A.2 Full process of the development of benchmarking data set which appears at pp 7-101 to 7-102 of Attachment 7 to the Ausgrid Final Decision. Table A.2 summarises the development of the AER’s benchmarking methodology, including the RINs, but stops short of the circulation of the EI model. The origins of the EI model does, however, appear as the final item in a three page chronology submitted by Mr O’Bryan on behalf of the AER.
234 The chronology shows that the AER’s development of its benchmarking methodology commenced in November 2011 with a joint AER / Australian Competition and Consumer Commission examination of benchmarking opex and capex in energy networks. The chronology then outlines some 41 steps (including seven workshops with DNSPs, consultants and consumer groups between December 2012 and May 2013). Those steps included:
(a) papers developed by EI and issued by the AER;
(b) workshops and public forums hosted by the AER;
(c) input from the applicant DNSPs;
(d) papers reviewing different benchmarking methods and an examination of the benchmarking practices of overseas regulators;
(e) consultations between the AER and stakeholders on the development of the RIN pursuant to s 28D of the NEL for the purposes of obtaining information from each DNSP relevant to the AER’s obligation under r 6.27 to prepare and publish an annual benchmarking report (the first of which was required to be published on 30 September 2013 (r 6.27(d));
(f) the AER publishing its economic benchmarking RINs (Economic benchmarking RIN-For distribution network service providers-Instructions and Definitions -NSP Name (ACN xxx xxx xxx), November 2013) together with an explanatory statement and instructions and definition documents (Better regulation - Explanatory statement - Regulatory information notices to collect information for economic benchmarking, November 2013);
(g) the DNSPs submitting to the AER audited and certified responses to the economic benchmarking RINs on 30 April 2014;
(h) the AER publishing its first annual benchmarking report AER, Electricity distribution network service providers–annual benchmarking report, November 2014, (six weeks after the date specified in r 6.27(d)); and
(i) the circulation on 27 November 2014 of the First EI Report.
235 The chronology also shows that the DNSPs were given four opportunities to comment on the RIN templates before submitting their unaudited responses and then their audited RIN responses.
236 Networks NSW provides the following examples of the disparate reporting of opex in response to the RINs:
(a) some Victorian DNSPs reporting nil vegetation management;
(b) DNSPs adopting different allocations of their respective regulatory asset bases (RABs);
(c) DNSPs applying different policies in classifying their opex and capex; and
(d) DNSPs applying different policies in classifying expenditure as opex or provisions.
237 Networks NSW’s submissions attribute the disparity in vegetation management reporting to an unclear definition of “vegetation management activities” in the RINs. A report by its consultant PricewaterhouseCoopers (PwC), Ausgrid, Essential Energy and Endeavour Energy – Appropriateness of RIN data for benchmarking, 9 January 2015 (PWC January 2015 Report), states at p 38 that:
Most businesses found that the definitions of ‘vegetation management activities’ provided by the AER were unclear, deeming them unworkable.
238 The AER dismisses Networks NSW’s submission on the disparity in vegetation management reporting by noting that vegetation management expenditure data was not used separately in its econometric benchmarking – the relevant opex measure being the aggregated network services opex figure.
239 It also dismisses Networks NSW’s submission relating to the disparity in classification of RAB values as irrelevant because such values were not included in its econometric models and, insofar as RAB data was used to weight the volume of inputs and outputs in the MTFP model, because MTFP is an index-based benchmarking method, the outcomes of that model are less sensitive to the weighting of inputs than they will be to the quantum of the inputs. Therefore any comparability issues in the RAB data have only a minimal impact on the MTFP results.
240 The AER also summarily dismisses Networks NSW’s concerns about the DNSPs classifying their opex and capex differently by observing that was adjusted, where necessary, through an OEF adjustment. This summary dismissal does not address the concern expressed by Networks NSW’s consultant PwC in its January 2015 Report that:
The capex / opex split between the businesses differs, ranging from 62% capex / 38% opex at SA Power Networks compared to 74% capex / 26% opex at CitiPower. This could be due to a range of factors including the relative age of the networks, capitalisation policies and cost allocation approaches. If there is more capitalisation, the operating expenditure reported by the business will be lower. Cost allocation methodologies and capitalisation policies affect the data provided by the DNSPs in the RIN, in particular the allocation of labour costs and overheads. This affected the AER’s calculation of the opex efficiency score and the level of reductions to opex for each of the three NSW DNSPs.
241 Nor does it address the conclusion in the PwC January 2015 Report at p 37 that the AER did not meet the AEMC’s necessary preconditions for benchmarking because (with emphasis as in the original):
• the benchmarking data is not long term reliable information as it was not provided on a like-for-like basis due to differences in capitalisation policies and approaches;
• the benchmarking data is not high quality due to the different cost allocation approaches undertaken by the DNSPs which impact the cost structures and expenditure incurred;
• the benchmarking data is not consistent time series data due to the differences in allocation of indirect costs over the last decade;
• the benchmarking data is not based on consistent definitions for the purpose of benchmarking.
242 The AER notes that Networks NSW’s final example of the disparate reporting of opex in response to the RINs (differing policies in classifying expenditure as opex or provisions) was incorrect in implying that opex and provisions are separate cost categories. It says provisions simply affect when a DNSP records a cost (ie costs which have been incurred but will be paid in the future) and each of the examples in Networks NSW’s submissions is an opex provision meaning that the cost of the provision is reflected in network services opex.
243 Expanding on its submission to the effect that weaknesses in the RINs resulted in some DNSPs estimating and backcasting some data, Networks NSW describe backcasting as involving the DNSPs creating estimates based on a set of assumptions for data points related to the past where actual results were not classified or categorised in the way the RIN required the information.
244 Expanding on its description of backcasting, Networks NSW notes that:
(a) the RINs were issued in November 2013 and required the DNSPs to supply data on the values and quantities of outputs, inputs and OEFs for an eight year period (from 2005-06 to 2012-13) within a short space of time (three months for an initial draft, and two further months for the final); and
(b) the data requested was not all kept in the ordinary course of the DNSPs’ business. Accordingly, the DNSPs were required, to “backcast” the data required.
245 Networks NSW provided the following examples of backcasted data:
(a) Ausgrid and Endeavour were required to backcast data for all current opex categories for FY2006 to FY2010, as a result of a material change in the Annual Reporting Requirements by the AER in 2011; and
(b) Endeavour was required to backcast data for 2009 to 2012 for certain OEFs, including the total number of spans and the average urban and rural maintenance span cycles.
246 Networks NSW also submitted that in a significant number of instances the DNSPs did not have accurate records of the required data and had to provide an estimate in lieu thereof – in support of its submission it provided details of Ausgrid, Endeavour and AusNet estimating route line length, CitiPower estimating its overhead conductors and underground cables and referred to a number of DNSPs estimating the number of urban/rural maintenance spans.
247 Networks NSW’s consultant, Frontier Economics Pty Ltd (Frontier), recognised some backcasting is, perhaps, inevitable when compiling a new dataset such as that for the purposes of the AER’s annual benchmarking report. It did, however, make the point that compiling several years of RIN data at once “carries major risks.” In its report Review of AER’s econometric models and their application in draft determinations for Networks NSW, January 2015 (the Frontier Report), it illustrated (at p 78) the risks as follows:
… if a network misinterprets how it ought to report certain data (which is very possible for the first time it reports RIN data), that mistake may be propagated through the full eight years of information reported. That, in turn, would distort comparisons with other networks not just for a single year but for all years that the data are reported. Such misreporting over the entire period would impact directly on the measures of ‘inefficiency’ derived by EI’s modelling.
248 Having identified how RIN data errors and inconsistencies may arise, the Frontier Report noted (at p 82) that:
Given the very real scope for data errors …, owing to the newness of the RIN data collection process and the lack of opportunity for learning and refinement, it is surprising to us that EI and the AER apparently have such confidence in the reliability of the modelling results.
249 Responding to Networks NSW’s submissions that the AER failed to collect the RIN data in a manner sufficiently rigorous to be useful for benchmarking, the AER submits that those submissions “create a misleading picture” for three reasons:
(a) first, many of the illustrations of estimated data referred to concern data that was not used in the benchmarking models, eg the backcast data for rural spans as described by Networks NSW was not included in any of the AER models on which its decision was based;
(b) secondly, contrary to Networks NSW’s submission, estimates were not used in “a significant number of cases”, most of the Australian data in the AER benchmarking models is actual data, not estimates; and
(c) thirdly, the general assertion that estimated or backcast data used in the AER benchmarking models is unreliable is not supported by evidence.
250 The AER sought to support its second reason by reference to the results of an examination of each of the six categories of data used in its models, namely, network services opex, customer numbers, ratcheted maximum demand, circuit line length, proportion of undergrounding and system average interruption duration index (SAIDI) reliability data (the SAIDI data being used only in the MTFP and MPFP models).
251 As presented by the AER the examination shows:
(a) network services opex: 8 of the 13 DNSPs provided actual data;
(b) customer numbers: since 2006 the Australian Energy Market Operator has required the DNSPs to maintain a unique National Metering Identifier (NMI) for each customer and information for reporting data in this category was available to each DNSP – estimates only having to be made where the NMI could not identify a “de-energised” customer (ie a property not currently receiving electricity) or in respect of a very small portion of customers who are unmetered;
(c) ratcheted maximum demand: 10 of the 13 DNSPs provided actual data – the remainder estimated data in some years and actual data in the others;
(d) circuit length: 7 of the 13 DNSPs provided actual data – only one, Essential, provided an estimate for all years and the other five provided an estimate for some years and actual data in others;
(e) proportion of undergrounding: 7 of the 13 DNSPs provided actual data in every year – only one, Essential, provided an estimate for all years and the other five provided an estimate for some years and actual data in others; and
(f) SAIDI reliability data: 11 of the 13 DNSPs provided actual SAIDI data in every year – only one DNSP estimated SAIDI data in all years and the other estimated data in some years and actual data in the others.
252 While, as presented by the AER, the results of its examination of the six categories of data does show that the majority of the 13 DNSPs may have responded to the RINs with actual data, the following alternative reading of the results shows:
(a) 5 of the 13 DNSPs estimated network services opex data;
(b) 3 of the 13 DNSPs estimated ratcheted maximum demand data in some years and provided actual data in the others;
(c) 6 of the 13 DNSPs estimated circuit length – one, Essential, provided an estimate for all years and the other five provided an estimate for some years and actual data in others;
(d) 6 of the 13 DNSPs estimated their proportion of undergrounding – one, Essential, provided an estimate for all years and the other five provided an estimate for some years and actual data in others; and
(e) 2 of the 13 DNSPs estimated SAIDI reliability data – one in all years and the other in some years with actual data being provided in the others.
253 This alternative reading of the results supports Networks NSW’s submission that a significant number of DNSPs estimated three categories of data used by the AER in its models. It also shows that estimates were used in another two categories (by three DNSPs in one category and by two in another).
254 Such an alternative reading should have put the AER on notice that it may, at this point in the evolution of the RIN data, have to treat the RIN data with greater caution than it did and not rely on it to the extent that it did, particularly as observed by its consultant, EI in the Second EI Report at pp x and 25:
… this is the first time economic benchmarking is being used as the primary basis for an Australian regulatory decision.”
… it is important to recognise that the characteristics of the Australian RIN data make any econometric model estimated using only the RIN data insufficiently robust to support regulatory decisions.
255 Also, as ActewAGL submits, it is in light of the second of the quotations from the Second EI Report, “somewhat surprising” that EI ran a model using only RIN data to corroborate its EI model.
256 Expanding on the third reason for Networks NSW submissions on the robustness of the RIN data capture creating a misleading impression (because it is not supported by evidence), the AER submits that:
(a) where DNSPs provided estimates, they were not generally required to create a completely new series of data;
(b) in most cases, the DNSPs produced estimates by drawing together actual information from a number of different sources in their business records, eg although Essential estimated data for network services opex, circuit length and proportion of undergrounding, the estimates were based directly on information that Essential had been collecting for the whole benchmarking period; and
(c) in each case where data was estimated by a DNSP, “… the DNSP’s CEO signed a statutory declaration attesting to the robustness of the data”.
257 As Networks NSW points out, the statement that “… the DNSP’s CEO signed a statutory declaration attesting to the robustness of the data” (which appears at p 7-128 of Attachment 7 of the Ausgrid Final Decision) is “wrong, and quite misleading.” According to Networks NSW, the form of statutory declaration which CEOs were required to sign did not attest to the “robustness” of the data. Rather, it submits, each CEO attested that the actual information provided was true and accurate and, where it was not possible to provide actual information, that a best estimate had been provided along with the basis of the estimate. In fact, Networks NSW submits, many businesses pointed out that the data was not robust and could not be relied upon for the purposes for which the AER apparently sought it.
258 Likewise, ActewAGL submits in reply “[A]t no point did ActewAGL’s CEO attest to the “robustness” of the data or as to its fitness for purpose in benchmarking. Rather, he attested to the fact that the data was ActewAGL's ‘best estimate’.”
259 The DNSPs’ submissions to the effect that the AER’s reliance on the RIN data point to a weakness in its benchmarking are persuasive. Support for those submissions is found in the above quoted passages from the DNSPs’ consultants’ reports. That support is reinforced by the passages from their consultants’ reports quoted below under the heading “The AER’s use of the EI model as the sole determinative of opex”.
260 Having regard to the above paragraphs under the heading “RIN data” and to paragraphs appearing below following the heading “The AER’s use of the EI model as the sole determinative of opex”, it is the view of the Tribunal that at this point in its evolution the RIN data is not data upon which the AER might rely on to the extent that it did:
(a) in its application of r 6.5.6(e)(4) to determine the benchmark opex that would be incurred by an efficient DNSP; or
(b) to corroborate the AER’s use of the EI model.
261 Turning now to outline what the DNSPs perceive as the second weakness in the EI model’s data, namely, the augmentation of the Australian RIN data with overseas data.
262 As may be observed from the following extracts from the First EI Report at p 29, ActewAGL rightly submits that it appears that EI chose to incorporate data from New Zealand and Ontario, not because the DNSPs in those countries were comparable, but because data was available that appeared to be in a similar form to the data in the RIN database.
Given that the New Zealand database has been constructed in a largely similar fashion to the AER’s economic benchmarking RIN database in terms of variable coverage, it is a prime candidate for use in supplementing the number of observations available from the RIN database.
… … …
The other jurisdiction that has a relatively long and consistent history of electricity DNSP productivity measurement is Ontario. Pacific Economic Group Research (PERG 2013) recently undertook … benchmarking work for the Ontario Energy Board (OEB) using a similar output specification to that used in section 3 above [ie ratcheted maximum demand, customer numbers and circuit length]. The OEB has put the database used in the public domain. While the Ontario database has similar coverage of outputs (other than reliability) to that used above and has good detail on opex, it is much more limited with regard to capital input and operating environment factor variables. Asset values are based on historic cost, for example, and, while there is data on the number of transformers, there is no data on transformer capacity. While Ontario’s climate is somewhat different to Australia’s, a significant attraction of the OEB database is the number of observations it offers with data for 73 DNSPs over 11 years from 2002 to 2012.
263 ActewAGL also highlights the significant disparities in size of the Australian DNSPs, on the one hand, and the overseas DNSPs, on the other, by reference to the following table which appeared in the Frontier Report at p 26:
Australian value as multiple of Ontarian value
Australian value as multiple of New Zealand value
Maximum Demand (MW)
Ratcheted Maximum Demand (MW)
Circuit Length (kms)
264 It may be concluded from the Frontier table reproduced above that the Australian DNSPs are, on average:
(a) four times larger than the companies in Ontario when compared using energy delivered and demand, six times larger when compared using customer numbers and eleven times larger when compared using circuit length; and
(b) eight times larger than the DNSPs in New Zealand when compared against all these measures of scale.
265 Also, as may be observed from the following extracts from the First EI Report at p 11, ActewAGL rightly submits that it appears that EI’s choice of output variables in the EI model (ratcheted maximum demand, customer numbers and circuit length) have been largely determined, in a somewhat circular manner, by the data that was available between all three countries.
… we agree with PEGR [PEG] that the four output specification covering energy throughput, ratcheted maximum demand, customer numbers and circuit length represents a useful way forward as it captures the key elements of DNSP functional output in a linear fashion and introduces an important demand side element to the measurement of system capacity outputs. Because we have reliable data on all four output variables, all four are included in our analysis.
266 ActewAGL rightly points out that:
(a) in essence both the construction of the EI models, and the data used within those models, were dictated by what was available to EI, not what was likely to give the most reliable indicator of efficient costs; and
(b) this ought to cause the Tribunal significant concern about the reliability of the output of EI’s models.
267 Each of the following issues that the AER identifies as the parties raising in relation to the EI model’s use of overseas data are addressed seriatim:
(a) whether the New Zealand and Ontario regulators used the overseas data for benchmarking opex;
(b) whether the parties had an adequate opportunity to verify the accuracy of the overseas data;
(c) whether the overseas data contains errors;
(d) whether there are errors in vegetation management costs which are not reported separately in New Zealand or Ontario;
(e) whether the overseas entities adopted different definitions for the purpose of collecting data – particularly in relation to maximum demand;
(f) whether there was a failure by either the AER or EI to conducted sensitivity testing in relation to potential errors in the overseas data, or inconsistencies between the overseas data and the Australian data, in order to quantify their potential impact; and
(g) whether the AER and EI were correct to ignore other potential sources of overseas data which yielded different results – particularly US data which is preferable to the Ontario data.
268 Addressing the first of those overseas data issues identified by the AER (whether the New Zealand and Ontario regulators used the overseas data for benchmarking opex) both ActewAGL and Networks NSW point to s 53P of the Commerce Act 1986 (NZ). Part 4 of that Act establishes a regulatory regime with a number of similarities to the NEL and s 53P provides, in effect, that before the end of a regulatory period the NZ Commerce Commission must set out the starting prices, rates of change and quality standards that are to apply for the following regulatory period. However, s 53P(10) precludes the Commission from using:
… comparative benchmarking on efficiency in order to set starting prices, rates of change, quality standards, or incentives to improve quality of supply.
269 Thus, ActewAGL submits, it cannot be inferred, that the Commission has made any attempt to ensure that the data has been recorded in a manner that makes it fit for the purpose of benchmarking.
270 As to the Ontario data, Networks NSW submit that:
(a) the regulator (the Ontario Energy Board (OEB)) has not adopted a similar model to the EI model and, perhaps, in recognition of the difficulties in classification of expenditure between opex and capex and of the legitimate trade-offs which electricity distribution entities must make, benchmarks on total expenditure (totex, ie opex plus capex); and
(b) there is therefore no particular reason to assume that the overseas data is suitable for benchmarking opex.
271 The AER rejects the implications that might be drawn from the ActewAGL and Networks NSW submissions addressing the issue whether the New Zealand and Ontario regulators use the overseas data for benchmarking opex.
272 In relation to the New Zealand data, relying on an earlier October 2014 EI report, the AER observes that the Commerce Commission:
(a) uses a total factor productivity measurement (a form of economic benchmarking), in forming a view about long-term productivity growth for New Zealand DNSPs to set their X factors in their default price paths; and
(b) measures opex productivity growth to form a view about the opex productivity growth rate to include in its DNSP opex forecasts which are used to set starting prices in its DNSP default price paths and that these productivity growth rates are industry average rates formed from the same database used in the EI’s model.
273 As to the claim that the OEB does not use the Ontario data for benchmarking opex, the AER makes the points that:
(a) it does use the data as part of benchmarking total cost performance (ie opex, plus the return of and on capital);
(b) the fact that the data is combined with other data to estimate total cost is irrelevant; and
(c) the OEB is conducting benchmarking using the same opex data used in the AER econometric models.
274 While the Final Decisions do not fully address issues relating to New Zealand and Ontario regulators’ respective use of the data and application of benchmarking, the AER’s submissions are persuasive in answering the DNSPs’ submissions on those particular issues. That is not to take the step of saying that the benchmarking by the AER is itself free of error of the character under consideration by the Tribunal.
275 The Tribunal now turns to consider the second of the issues identified by the AER as being raised against the EI model’s use of overseas data, namely, whether the parties had an adequate opportunity to verify the accuracy of the data.
276 While ActewAGL does not raise what it describes as “the relative speed” by which the AER developed its benchmarking methodology as a ground of review, it rightly submits that the circumstances in which the AER developed its methodology and implemented it is:
… an important contextual matter, and one which informs the extent to which the Tribunal can have confidence in the robustness of the AER’s approach.
277 The steps taken by the AER in the development of its benchmarking methodology are outlined above.
278 The thrust of r 6.5.6(e)(4) is that the AER must have regard to the most recent annual benchmarking report that has been published under r 6.27 and the benchmark opex that would be incurred by an efficient DNSP.
279 Rule 6.27 enlivens r 8.74 in relation to the steps that the AER must take before preparing and publishing an annual benchmarking report. Those steps include consultation with the DNSP and a 30 day period to make submissions before the report is published including an opportunity to comment on material of a factual nature to be included in the report.
280 It appears from the above mentioned chronology submitted on behalf of the AER and from ActewAGL’s leave application and its submissions that:
(a) on 5 August 2014, the AER provided the DNSPs with a copy of its Draft Annual Benchmarking Report;
(b) on 18 November 2014 (nine days before the Draft Decisions were published on 27 November 2014) the AER provided the DNSPs with:
(i) a copy its Annual Benchmarking Report; and
(ii) a copy of the First EI Report;
(c) the Annual Benchmarking Report was directed towards the use of a MTFP benchmarking model; and
(d) the DNSPs’ submissions in the course of the consultation process leading to the publication of the Draft and Final Annual Benchmarking Report were directed to the use of a MTFP model, not the econometric EI model which was (with adjustments) adopted by the AER in its Draft Decisions and, with further adjustments, in its Final Decisions when it did not accept the DNSPs’ opex forecast and made its own estimates of their required opex.
281 It is ActewAGL’s view that the manner in which the AER prepared its Draft Decisions in reliance on the EI model deprived the AER of a robust exchange of views with the DNSPs and their experts. Such an exchange, it rightly submits, was:
… particularly important in a context where the AER proposed to rely on an ambitious modelling technique, for the first time, as the sole basis on which to reach an assessment of the quantum of the DNSPs’ opex allowances.
282 ActewAGL supported its submission with the following extracts from pp 104 and 105 the Frontier report:
A key flaw of the analysis undertaken by the AER is the application of very ambitious modelling techniques, such as … [the EI model]…, to very imperfect data. Indeed, it appears that the main reason the AER has felt the need to employ overseas data, without appropriate checks for robustness and consistency, is its desire to employ sophisticated techniques such as … [the EI model].
We recognise that the AER is obliged to undertake benchmarking under the National Electricity Rules (NER). However, the NER also provide the AER with considerable flexibility to choose the most appropriate benchmarking techniques and methodologies. The AER should not feel constrained to restrict itself to benchmarking using formal statistical techniques alone.
Given the limitations of the Australian RIN data, and the lack of time for learning and iterative improvement of the data, we recommend that the AER rely on much simpler benchmarking techniques. We reiterate that regulators in Europe, who have had considerably more experience, and time to compile consistent data, than has the AER, typically use much simpler, and more pragmatic benchmarking techniques.
The AER has applied a very narrow interpretation of benchmarking. Its Expenditure Forecast Assessment Guideline sets out a very long list of potential benchmarking techniques, all of which would be recognised in Europe and many of which are used by regulators overseas. Whilst it canvassed in its Guideline the potential use of many alternative techniques, its assessment of relative efficiency seems to drive off only one technique, … [the EI model]…, and that too in a very mechanistic fashion. Given the sensitivity of such techniques to the quality of the data, and the fact that the RIN data are very new and relatively untested, the AER should not have, in our view, placed so much reliance on statistical techniques such as … [the EI model]. Rather, in our view, the AER should have initially tried much simpler, less ambitious techniques and then aimed to build up to more complex techniques once it, and networks and customers, have greater confidence in the data and in the AER’s approach to benchmarking.
283 ActewAGL is correct in its submission that having regard to the limited time between the date of publication of the Draft Decisions and the date by which the Final Decisions had to be made, the AER had little opportunity to develop an alternative methodology to estimate the required opex in accordance with r 6.12.1(4)(ii). A fortiori, it says, as the AER would need to discharge its obligations to accord procedural fairness under administrative law and s 16(1)(b) of the NEL in respect of any new methodology.
284 This, ActewAGL submits, is not to suggest that the AER did not conscientiously examine submissions received after the Draft Decisions. It is:
… simply to recognise that the AER’s decision making approach could hardly have been more unfortunate, in that it did not ensure that its methodology was exposed to an unhurried and carefully considered dialogue with all interested stakeholders, without the imminent deadline of the date by which the Final Determination had to be made.
285 ActewAGL’s submission is reinforced when regard is had to the numerous DNSPs’ consultants’ reports critical of the AER’s application of the EI model, many of which disparaged the model’s use of overseas data.
286 Networks NSW submitted six such reports. Each raised issues which, in circumstances where the AER is employing a sophisticated and complex benchmarking model for the first time, would have benefited from a wider critical exposure and response through the AER’s consultation procedures than just the responses by the AER and EI. The consultants (each highly qualified and well recognised experts) their reports and the issues they raised are:
(a) The Frontier Report. As fairly summarised by Networks NSW, the Frontier report is critical of the inadequacy of the RIN data, comparability and other issues associated with the use of overseas data, the failure of the EI model to consider alternative explanations of heterogeneity and the AER’s deterministic application of the EI model’s results.
(b) Huegin Consulting, Huegin’s response to Draft Determination on behalf of NNSW and ActewAGL - Technical response to the application of benchmarking by the AER, 16 January 2015 (the Huegin Report). Huegin highlights the inadequacies of the AER’s approach by comparing it to best regulatory practice and identifying what it perceives as errors associated with the cost drivers chosen by EI, inadequate consideration of environmental variables and the use of average results over an eight year period. Huegin also criticises EI’s sensitivity analysis and questions the corroborative value of the alternative methods presented by EI.
(c) Cambridge Economic Policy Associates Ltd (CEPA), Networks NSW - AER draft determination, 16 January 2015 (the CEPA Report). The CEPA report observes that the author’s:
“… investigations, initially of ActewAGL, but applying more generally to all DNSPs, … cast doubt on the claim that the AER has correctly carried out the opex benchmarking, and at least has not given sufficient consideration to the limits of its opex benchmarking.
(d) The Second PEG Report. This Report provides a detailed introductory background to benchmarking, the salient considerations in the benchmarking of network services opex, reviews the practice of international regulators and critiques EI’s work on behalf of the AER. It particularly identifies issues with their use of the RIN and overseas data, and the use of the results by the AER.
(e) Advisian Pty Ltd (Advisian), Review of AER Benchmarking Networks NSW, 16 January 2015. Advisian found that there were significant differences between the DNSPs used for benchmarking purposes, with limited meaningful consideration by the AER or EI to ensure that the benchmark data has appropriately been normalised. Advisian also identified issues with, amongst other things, the AER’s benchmarking approach relating to:
(i) comparability of the DNSPs used for benchmarking purposes;
(ii) the failure to appropriately consider the effect of spatial density (customers/km2) in addition to linear density (customers/km) on efficient opex;
(iii) the need for each DNSP to operate and maintain, in a safe and reliable manner, the assets it actually has, rather than the assets it might have had;
(iv) a failure to account for exogenous factors (such as the nature of the assets and the development history of the network) that have influenced the development of the existing asset base; and
(v) the AER’s assessment of Essential’s vegetation management expenditure to support its conclusion that the NSW DNSPs are inefficient.
(f) The PwC January 2015 Report reviewed the RIN data for the NSW DNSPs and five other DNSPs (CitiPower, Powercor, AusNet, United Energy and SA Power). The author identified seven issues with the RIN data where he considered a correction should be made and considered when assessing the efficiency of the DNSPs. In the author’s opinion the issues are a central part of the AER’s MTFP and PPI analysis and directly impact the AER’s benchmarking results.
287 ActewAGL also submitted reports by CEPA, Huegin and Advisian raising similar issues to those raised in their reports on behalf of Networks NSW. In addition to those reports, ActewAGL submitted the report by AECOM Australia Pty Ltd (The impact of the AER’s Draft Decision on ActewAGL’s Service and Safety performance, 15 January 2015 which, like its other reports and those submitted by Networks NSW, would have benefited from a wider critical exposure and response through the AER’s consultation procedures than just the responses by the AER and EI. The report concluded, amongst other things:
(a) the benchmarks used by the AER have little or no relevance to a DNSP, and judgments made based on them are of limited value;
(b) that a forced reduction in ActewAGL’s replacement / renewal expenditure (repex) and opex of the scale suggested by the AER would have a significant impact on the level of service it is able to provide including a potential impact on safety levels associated with its assets; and
(c) ActewAGL has experienced engineers who use all the information available to them and sophisticated analysis tools to optimise total cost of ownership for critical assets and therefore determine the optimal timing for replacement / renewal (and therefore for repex) whereas the AER relied on generic econometric models and distantly related data from other sources to over-rule ActewAGL’s experienced projections, in many cases using ‘average’ asset lives that are almost double ActewAGL’s experience-based estimates.
288 The limited opportunity afforded the DNSPs and the lack of opportunity afforded to PIAC and other interested parties to comment on the AER’s application of a benchmarking methodology reliant on overseas data does not of itself give rise to a relevant ground of review. However, as the DNSPs submitted, it tends to tell strongly against the acceptance of that methodology and the resulting estimates of the DNSPs’ required opex that the AER derived from it.
289 In response to the third issue identified by the AER as being raised against the EI model’s use of overseas data (whether it contains errors) the AER notes that there is no evidence that suggests the data collected by the NZ Commerce Commission and the OEB is unreliable and that, contrary to the DNSPs’ submissions, the Commission and the OEB do rely on the data – data, which the AER asserts, is collected and verified in a manner similar to the AER’s collection and verification process.
290 In responding to Networks NSW’s concerns that sudden, unexpected and unexplained changes in the overseas data appear to be the product of human error, the AER relies on the Second EI Report’s conclusion that, with one exception, they are unlikely to be errors, but simply movements attributable to the small size of some of the DNSPs.
291 That response, however, gives credence to a submission by Networks NSW that a finding that the relationship between cost drivers and opex is not consistent across Australia, New Zealand and Ontario DNSPs and that it makes no sense to compare small New Zealand and Canadian DNSPs, which may experience volatile changes, to large DNSPs in Australia which do not.
292 In relation to the fourth issue on overseas data (vegetation management costs) it is Networks NSW’s contention that because such costs are not reported separately in New Zealand or Ontario for the period of the RIN data, it is impossible to verify whether there are errors in relation to this category of opex. The AER dismisses this contention as irrelevant to the AER models because the input measure used in its models is total opex and vegetation management costs are not a separate input measure. Thus, in the AER’s view, as long as total opex is consistent within each country, it is not necessary to review the individual components of that figure across countries.
293 Networks NSW relies on the Second PEG Report to develop its submissions on the fifth overseas data issue, whether the overseas entities adopted different definitions for the purpose of collecting data – particularly in relation to the definitions of:
(a) opex – opex in Ontario includes the costs of customer care services such as metering and billing but excludes costs of maintaining substations with incoming voltage exceeding 50kV; and
(b) ratcheted maximum demand – Australian and New Zealand DNSPs report “coincident ratcheted maximum demand” representing the peak level of demand across a DNSP’s entire network at any one time whereas Ontario DNSPs report ‘non-coincident ratcheted maximum demand’, representing the aggregate of the peak levels of demand for the individual constituent parts of a DNSP’s network (the constituent parts usually being aggregated at either the subtransmission substation level or the zone substation level).
294 It is Networks NSW’s submission that non-coincident ratcheted maximum demand is generally higher than coincident ratcheted demand, because different parts of a DNSP’s network may experience peak levels of demand at different times and the difference can be substantial – for the 13 Australian DNSPs during the RIN data period, in 5 cases the difference was over 10 percent and in one case over 25 percent.
295 The AER’s answer to Networks NSW’s overseas definition issues is that the differences do not have any material effect on the results of the EI model because the country dummy variables are explicitly designed to correct for systematic reporting differences. Networks NSW’s reply to the AER’s answer is a blunt, on target, “It will not.”
296 The Tribunal is not of the view that the country dummy variables, in the present circumstances, correct for systematic reporting differences. As Networks NSW rightly submits, the type of country dummy used by EI assumes the relevant relationships between cost drivers and opex is the same across the three jurisdictions and cannot control for the situation where one of the relevant cost drivers has been defined differently in one jurisdiction, thereby altering the relationship between it and opex for that jurisdiction. Thus, for the same reason, the country dummy variables do not “correct” for the differences in the examples provided by the AER relying on the Second EI Report, as outlined above.
297 As observed by the following Networks NSW consultants:
(a) the CEPA Report at p 17:
Including a dummy variable in the model specification does not necessarily control for these within and across country differences. A dummy variable only controls for level differences between datasets not cost relationship differences.
(b) the Frontier Report at p ix:
The consequences of the significant differences in operating environment across the sample is that the business models applied by the businesses are likely to be very different – for example, an Ontarian business operating in a harsh wintry environment will have a completely different business model to achieve a given level of security of supply than a rural Australian network operating over an enormous service region. In turn, this will mean that the relationship between costs and cost drivers is quite different across the two jurisdictions, and is not amenable to being captured by a relatively small number of high level explanatory factors combined with country dummy variables (as per EI’s approach).
298 Also, relying on the following extract from the CEPA Report at p 16, ActewAGL submitted that a dummy variable is entirely inapposite to adjust for differing cost relationships between jurisdictions:
The introduction of the dummy variable takes a fixed amount of Country A’s opex per network length to bring its average in line with Country B's. However, the slope of the line (the relationship between opex and network length) is not impacted by the introduction of the country dummy variable. A proper econometric analysis is more complex than this and should take account of country-specific slopes, which will require more variables to take this into account. For example, if the relative prices of labour and capital differ, then one would expect a different relationship between cost and customer numbers (e.g. higher labour costs should lead to more capex and lower maintenance costs, but higher costs of dealing with customers).
299 As seen by ActewAGL, the issue is of “significant concern”. That is because most of the data used within EI’s models, including the EI model, comes from overseas. Australia only accounts for 19 percent of the data points used. Accordingly, even with the use of a “dummy variable”, the slopes (coefficients) estimated by the regression models will closely follow the overseas DNSPs, rather than the Australian DNSPs, because of the sheer volume of data that comes from overseas. That is, the model will reflect cost relationships between opex and drivers of opex that exist in the overseas DNSPs, rather than modelling relationships that exist in Australia.
300 The AER’s response to CEPA, Frontier and PEG’s concerns about the use of country dummy variables appears in Attachment 7 to the Ausgrid Final Decision at p 7-121. Noting that:
(a) while CEPA agrees that the dummy variables control for level differences between databases but considers they do not account for cost relationship differences; and
(b) similarly, Frontier and PEG submit that each service provider's costs are influenced by factors not captured by the explanatory variables in the EI model,
the AER observes that:
In response to this, Economic Insights considers for such differences to have a material impact on the model results, significant differences in the technology to distribute electricity would need to exist. Economic Insights notes the international service providers deliver the same services using poles, wires and transformers so it does not agree that such a fundamental difference exists.
301 It is the view of the Tribunal that that observation:
(a) glosses over the multitude of differences that no doubt exist between the poles, wires and transformers of the Australian DNSPs and those of their overseas counterparts – differences due to such matters as topography, climate and regulations relating to the standards of the poles, wires and transformers; and
(b) fails to address the fact that the EI model’s use of country dummy variables impugns the robustness claimed for the model by the AER.
302 Indeed, the EI model’s use of country dummy variables reveals a significant weakness in the model.
303 In respect of the sixth issue, raised by Networks NSW, relating to the use of overseas data (an alleged failure by either the AER or EI to conduct sensitivity testing in relation to potential errors in the overseas data, or inconsistencies between the overseas data and the Australian data, in order to quantify their potential impact) the AER says that the allegation is not correct because EI undertook extensive sensitivity testing by creating four different economic models and PPI analysis, some of which incorporated the overseas data and some of which did not. Moreover, EI calculated each of the AER econometric models using different combinations of overseas data (the “full”, “large”, “medium” and “small” datasets), as well as different combinations of included outputs, different functional forms and different estimation methods. The AER submits that:
(a) as a result EI found that the modelling results were relatively insensitive to the dataset used and to small changes in specification; and
(b) EI’s finding indicates the results of the AER models are robust and insensitive to possible errors of the type alleged.
304 In reply, Networks NSW asserts that the AER’s submission should be rejected because the “sensitivity testing” referred to by the AER involved the following:
(a) the econometric models estimated originally by EI which contained five variables (energy throughput, customer numbers, circuit length, ratcheted maximum demand and share of undergrounding);
(b) EI subsequently dropping the energy throughput variable as it was correlated with ratcheted maximum demand causing the variables to lack statistical significance;
(c) EI running four econometric models (an LSE model with Cobb-Douglas and translog functional forms and an SFA model with Cobb-Douglas and translog functional forms) however, EI did not present the results of the translog SFA model;
(d) concerns by EI that the inclusion of very small DNSPs in the Ontario and New Zealand data sets might exert undue influence on the results, causing EI to run each of the models on different data sets with a different cut off of customer numbers. These were the Full (all DNSPs), Large (DNSPs > 10,000 customers), Medium (DNSPs > 20,000 customers) and Small (DNSPs > 50,000 customers) data sets. The preferred data set was the medium data set, containing 68 DNSPs, of which 37 were from Ontario and 18 from New Zealand and, as submitted by Networks NSW, this data set still includes a large number of DNSPs that are very small compared to the Australian businesses;
(e) EI has not presented the results from any data set aside from the medium data set.
305 Thus Networks NSW claims that the “sensitivity testing” done by EI was not sensitivity testing at all because all econometric models presented and relied upon use the same data and the same variables. Networks NSW went on to submit that the following options for genuine sensitivity testing might have been adopted by EI:
(a) using any of the data from the United States which PEG had prepared at the request of the AER (even as a cross-check);
(b) seeing what would be the effect of “normalising” the data prior to modelling by removing DNSP specific unusual expenditure (which was done by its consultant CEPA, producing substantially different results);
(c) using data envelopment analysis (DEA) – a methodology that the AER had said in its EFA Guidelines it would use;
(d) using fixed effects or random effects SFA models, which allow for heterogeneity at the DNSP level (as opposed to EI’s SFA model which essentially assumes no latent heterogeneity);
(e) inclusion of any additional variables (as some of Networks NSW’s consultants have done).
306 There is merit in Networks NSW’s suggested options for sensitivity testing to identify potential errors in the overseas data, or inconsistencies between the overseas data and the Australian data, in order to quantify their potential impact. While it might be understandable that time constraint pressures on the AER precluded it pursuing all the options, that they were not pursued does not increase confidence in the AER’s reliance on overseas data to arrive at its estimate of a DNSP’s required opex.
307 The seventh and final issue relating to the use of overseas data advanced by Networks NSW (whether the AER and EI were correct to ignore other potential sources of overseas data – particularly US data compiled by PEG for the AER which is preferable to the Ontario data) is rejected by the AER because:
(a) the data was assembled from a disparate range of sources, most of which were for the purpose of financial reporting, rather than economic benchmarking, meaning that the data mainly focused on financial variables;
(b) much of the US data was derived from vertically integrated businesses, which resulted in numerous cost-allocation issues; and
(c) many quantity measures are either not reported at all or not reported consistently in the US, including fundamental variables such as line length, maximum demand and reliability.
The AER notes that:
(a) having regard to the limitations PEG was only able to assemble data for 15 United States DNSPs, consisting of only 170 observations;
(b) the sample was widely unbalanced, eg three DNSPs had 19 years of data and three DNSPs had only two years of data;
(c) the data was not generally comparable with the New Zealand and Ontario data in terms of coverage and definition, nor did it support EI’s preferred specification using Australian data;
(d) the database did not have consistent coverage of the variables used in the AER econometric models and the AER’s productivity models; and
(e) for these reasons, EI concluded that the US database was not fit for the purpose of being included in the AER’s economic benchmarking.
308 While the AER addresses Networks NSW’s submission on the seventh and final issue relating to the use of overseas data, there is significant substance to other overseas data issues raised by the DNSPs – in particular the second of those issues, namely, limitations on the opportunity to enhance the AER’s reliance on overseas data by subjecting the data to greater consultation through the mandated consultation processes had it been flagged earlier in the chronology leading to the Draft Decisions. As the Tribunal has noted, the fact of a lack of opportunity to consult more fully is not itself a ground of review relied upon by the DNSPs.
The AER’s lowering of the EI model’s comparison point
309 As observed above, the AER lowered the efficiency target comparison point from the 0.86 in its Draft Decisions to 0.77 in its Final Decisions.
310 Attachment 7 to the Ausgrid Final Decision details (at p 7-276) the AER’s reasons for lowering the comparison point as follows (without footnotes):
We have decided, on balance, for this decision, that the appropriate benchmark comparison point is the lowest of the efficiency scores in the top quartile of possible scores rather than the average approach we used in our draft decision. This is equivalent to the efficiency score for the business at the bottom of the upper third (top 33 percent) of companies in the benchmark sample (represented by AusNet Services). Our revised comparison point is appropriate for the following reasons.
First, our draft decision averaging approach produced an unusual result for service providers ranked in the top quartile of efficiency scores, but below the average of that top quartile. These service providers would require an efficiency adjustment to reach the average benchmark comparison point (because their scores are below the average) despite being efficient enough to be ranked in the top quartile and, hence, included in the average.
Second, given it is our first application of benchmarking, it is appropriate to adopt a cautious approach. We have decided to increase the margin for error for modelling and data issues provided for in the draft decision (which reduced the benchmark comparison point from 0.95 to 0.86).
Third, we consider this approach better achieves the NEO and RPPs. In particular we have considered:
the principle that we should provide service providers with an opportunity to recover at least their efficient costs
we wish to create a high-powered efficiency incentive (which supports making an adjustment when it is clear there is material inefficiency in revealed costs) but we are mindful of providing sufficient stability to promote efficient investment
our decision should allow a return that is commensurate with both regulatory and commercial risks.
A number of service providers, representing more than a third of the NEM [national electricity market], and operating in varied environments, are able to perform at or above our benchmark comparison point. We are confident that a firm that performs below this level is, therefore, spending in a manner that does not reasonably reflect the opex criteria. An adjustment back to an appropriate threshold is sufficient to remove the material over-expenditure in the revealed costs while still incorporating an appropriately wide margin for potential modelling and data errors and other uncertainties. Economic Insights agrees that this approach is appropriate.
Our approach of using benchmarking as a basis for making adjustments to opex is also consistent with Ofgem's approach.
311 There is merit in the following submissions that PIAC advances against the AER’s above quoted reasons for lowering the comparison point:
(a) The AER’s first reason is irrelevant to its decision to reduce the benchmark comparison point. The AER’s “unusual result” is simply the consequence of deriving the benchmark comparison point as the average of any grouping of networks, rather than the score achieved by a single network. It would arise if the average were taken of the top five networks as it would if the average were taken of the top eight or nine networks.
(b) The AER’s second reason is a repetition of its original explanation of why it set the benchmark comparison point below the efficiency frontier, merely reciting that it has decided to increase that downward allowance, in order to take a “cautious” approach. No explanation is provided why the AER has concluded “modelling and the data issues” justify an initial adjustment more than twice the magnitude of the reduction in the Draft Decisions.
(c) The three points comprising the AER’s third reason amount to no more than unreasoned box-ticking of some, but not all, of the RPP in s 7A of the NEL.
(i) Invoking the principle of providing service providers with an opportunity to recover “at least their efficient costs” is circular: the purpose of the benchmarking exercise is to ascertain what level of performance may be said to “reasonably reflect” the DNSPs’ efficient costs.
(ii) No explanation is given of what “regulatory and commercial risks” the doubling of the initial downward adjustment is supposed to reflect.
(iii) The AER’s reference to “providing sufficient stability to promote efficient investment” is wholly out of place in its consideration of whether a DNSP’s proposed opex requirements reflect the efficient operating costs of a prudent network operator.
(iv) If the AER is suggesting that it had decided to reduce the benchmark comparison point to insulate the DNSPs’ shareholders from the effect of an immediate reduction in their opex allowance, that suggestion is at odds with the AER’s observation in Attachment 7 to the Ausgrid Final Decision that:
If a transition is a “premium” above the efficient costs that a prudent operator would require, we cannot include that premium in our estimate of total forecast opex that we are satisfied reasonably reflects these opex criteria. Conversely, if a transition is included as part of a forecast that does reasonably reflect the opex criteria, no further premium is required or possible.
(d) Insofar as the AER relies on EI’s advice to justify its further reduction, EI’s own reasoning is shown to be informed by nothing more than the adoption of an even more conservative benchmark than the adjustment it described as “conservative” and “generous” at the Draft Decision stage – see the Second EI Report at p 65:
We have previously noted that it is prudent to adopt a conservative approach to choosing an appropriate benchmark for efficiency comparisons. Adopting a conservative approach allows for general limitations of the models with respect to the specification of outputs and inputs, data imperfections and other uncertainties. While we have not found any of the criticisms made of our 2014 economic benchmarking study warrant changes to be made to our underlying approach, we are of the view there may be a case for setting a more conservative target than that used in Economic Insights (2014). This is particularly the case given that this is the first time economic benchmarking is being used as the primary basis for an Australian regulatory decision.
We are of the view that instead of using the customer weighted average of efficiency scores in the top quartile of possible scores, a more conservative approach of using the lowest of the efficiency scores in the top quartile of possible scores is appropriate. This would make the average efficiency score of 0.77 achieved by AusNet Distribution the appropriate opex efficiency target, before allowance for additional operating environment factors not included in the econometric modelling. Being a predominantly rural DNSP also makes choosing AusNet Distribution’s score a relatively conservative choice for the efficiency target. This change represents a 9 percentage point reduction in the opex efficiency target (from 0.86 to 0.77) and so is a generous additional allowance for any remaining modelling limitations, data imperfections and other uncertainties. It also represents a lower target of around the bottom of the top third of DNSPs compared to Ofgem’s target of the 75th percentile DNSP. Allowance for the additional operating environment factors not included in the econometric modelling then further reduces this target again (to between 0.62 and 0.69, depending on the DNSP).
(e) Finally, PIAC contends that the AER’s assertion of the approach of consistency with the UK regulator, the Office of Gas and Electricity Markets’ (Ofgem) approach is, for the following reasons advanced by PIAC, inapt:
(i) Ofgem’s totex benchmarking uses the upper quartile level of performance, under a weighted average of three benchmarking models. That is explained by Ofgem in the following extract from its RIIO-ED1: [Revenue using Incentives to deliver Innovation and Outputs-Electricity Determination 1] Final determinations for the slow-track electricity distribution companies - overview, 28 November 2014 set out in PIAC’s opex submissions at :
4.12. We benchmark the efficient level of totex for each DNO [distribution network operator] using the upper quartile (UQ) of the combined outputs from the three models. This addresses the risk that the combination of three separate UQ benchmarks might result in a benchmark that is tougher than any of the DNO forecasts. We use UQ rather than the frontier to allow for other factors that may influence the DNOs’ costs. The UQ level of efficiency (lower quartile level of costs) is the 25th percentile in the distribution of efficiency scores.
(ii) Properly understood, the AER’s approach is not a “top quartile” approach. Rather, the AER applied “the lowest of the efficiency scores in the top quartile of possible scores (represented by AusNet Services)” ie the DNSP whose EI model efficiency score was the lowest above 0.75 or the fifth most efficient of the 13 DNSPs (Attachment 7 to the Ausgrid Final Decision at p 7-721).
(iii) The AER described this as “equivalent to the efficiency score for the business at the bottom of the upper third (top 33 percent) of companies in the benchmark sample” (Attachment 7 to the Ausgrid Final Decision at p 7-721).
(iv) Yet even that understates how far the benchmark had been lowered at this initial step as AusNet lay outside the upper third of the sample. In distributional terms, AusNet sat at the 62nd percentile, which most closely approximates the cut-off for the top three octiles, ie implying a 50 percent larger reduction than if the AER had adhered to Ofgem’s “upper quartile” approach.
(v) Moreover, an important facet of Ofgem’s totex benchmarking method is that, having set the benchmark at that upper quartile level, it makes only very limited allowances for what may be described as OEFs. In particular, Ofgem granted “company specific factors” allowances to only three out of 14 of its DNOs – see Ofgem’s RIIO-ED1: Final determinations for the slow-track electricity distribution companies Business plan expenditure assessment, 28 November 2015, at [4.1] ff which explains Ofgem’s approach as follows:
4.1. We consider whether DNO submitted data require adjustments prior to carrying out our comparative benchmarking. This is to ensure that the comparisons are on a like-for-like basis. Where we decide adjustments are appropriate, we adjust the DNO submitted costs before our totex and disaggregated assessments. These adjustments fall into four broad categories:
1. Regional labour costs. These adjustments are made as operating in certain parts of the country attracts significantly higher labour costs. These apply to the two totex models and the disaggregated model in the same way.
2. Company specific factors. These are additional costs associated with operating a particular DNO network. The size of the adjustments differs in the disaggregated model compared to the two totex models. For some activities the disaggregated analysis already factors in the special case and to apply these adjustments again would be a double count. For example, if the special case is based on the need to do more volumes of work and our disaggregated model allows all the submitted volumes, we would not make a further company specific adjustment.
3. Exclusions from totex models. These are costs that are inappropriate for comparative benchmarking because they are not adequately explained by cost drivers that are being used in the totex models or because there is a substantial change in the nature of the activity between DPCR5 and RIIO-ED1. These exclusions only apply to the totex models. This does not apply to the disaggregated analysis. At the disaggregated level each cost activity is assessed by a bespoke model which uses the most intuitive cost driver and accounts for any changes in historical and/or forecast costs.
4. Other adjustments. Three other adjustments we make are to remove costs outside the price control, to remove non-controllable costs and to account for indirect cost allocation. These apply to the two totex models and the disaggregated model in the same way.
4.2. Once we estimate the modelled costs for each activity and for totex, we reverse the regional labour adjustments and company specific adjustments and add back an efficient view of those cost items excluded from our benchmarking analysis.
(vi) Thus, it is PIAC’s submission, that properly understood, Ofgem’s “upper quartile” method involved both a smaller lowering of the benchmark comparison point from the efficiency frontier, and a considerably more selective approach to making any additional ad hoc, network-specific adjustments that would further lower the efficiency benchmark comparison point. Unlike the AER’s method, Ofgem’s “upper quartile” benchmarking method encompasses allowances for all but the most material variances in operating environment.
312 There is also merit in ActewAGL’s submission that the AER’s decision to lower the EI model’s comparison point was arbitrary and that it indicates that the AER was not confident in the results of the EI model – a submission made notwithstanding that ActewAGL benefits from the lowering of the comparison point.
313 ActewAGL questions why, if EI and the AER regard the criticisms of its benchmarking methodology to be without foundation, was the shift made? What was the underlying rationale for the decision to make the shift? If the AER was confident in the EI model and its OEF adjustments, why not set ActewAGL’s opex by reference to the most efficient Australian firm, as assessed by the EI model? Those questions are, ActewAGL submits, incapable of definitive answer, because the decision about where the frontier should sit does not have an analytical premise.
314 That decision, it submits, is based on nothing more than unease about the reliability of the EI model and the OEF adjustments. Setting ActewAGL’s opex allowance by reference to the bottom firm of the top third of efficient firms (as per the Final Decisions) is no more defensible than setting it by reference to the average of the top third of DNSPs (as per the Draft Decisions). ActewAGL concludes its submissions on the AER’s lowering of the comparison point by observing the fact that such an arbitrary choice is apparently necessary, because of deficiencies in the EI model, suggests that no reliance should be placed on the results of that model.
315 Networks NSW’s submissions in relation to the AER’s lowering of the comparison point are mainly directed at rebutting PIAC’s submissions on the issue. To that end, Networks NSW makes two submissions:
(a) first, PIAC’s contentions rest on the assumption that the EI model and the AER’s application of it is valid and if the Tribunal were to uphold Networks NSW’s submissions to the contrary, PIAC’s submissions would fall away; and
(b) second, PIAC has put on no evidence, nor suggested any reason, why the average of the upper quartile is a more appropriate benchmark than the bottom of the upper quartile in support of its suggestion that the AER return to the average of the upper quartile.
316 The above quoted reasons advanced by the AER for lowering the comparison point make:
(a) a significant acknowledgment of general limitations in its models with respect to the specification of outputs and inputs, data imperfections and other uncertainties; and
(b) do nothing to assuage concerns about the use of such a sophisticated econometric model as the EI model where economic benchmarking is being used for the first time in Australia.
317 The submissions put by PIAC against those reasons are cogent. So too are the unanswered questions raised by ActewAGL.
318 Each of Networks NSW, ActewAGL, Ergon and PIAC submit that the AER erred in making its post modelling OEF adjustments described above.
319 As observed above, recognising that the EI model does not account for many OEFs relevant to estimating opex, the AER made ex post adjustments for them. For ease of reference the adjustment process detailed above may be summarised as follows:
(a) first, the AER decided whether the OEF was material or immaterial;
(b) if in the AER’s opinion the OEF was material, the AER assessed its quantum vis á vis the DNSP under consideration;
(c) if, however, the AER was of the opinion that the OEF was individually immaterial and:
(i) it was likely to provide a cost disadvantage, it made a uniform +0.5 percent adjustment;
(ii) was, in PIAC’s terms “directionally ambiguous”, ie there was doubt whether it was likely to provide a cost advantage or disadvantage, it made a uniform +0.5 percent adjustment; and
(iii) it was likely to provide a cost advantage, it made a uniform –0.5 percent adjustment.
320 The results of each adjustment for the Networks NSW DNSPs and for ActewAGL appear in two tables in Attachment 7 to the relevant Final Decisions (Table A.6 Summary of final decision on OEF adjustments; and Table A.8 Summary of individually immaterial OEF adjustments, respectively) which are reproduced above.
321 PIAC notes that that the seemly inconsequential 0.5 percent adjustments in Table 6.A have a very substantial overall effect of the Networks NSW DNSPs’ aggregate opex allowances over the five-year regulatory period. PIAC estimates that each individual +0.5 percent adjustment results in an increase in the base-year opex of approximately $1.7m for Ausgrid, $1.1m for Endeavour, and $1.4m for Essential. Applied across the regulatory period, PIAC estimates that each individual +0.5 percent adjustment in turn is equivalent to an increase in the nominal aggregate opex allowance of $9.8m, for Ausgrid, $6.1m for Endeavour and $8.1m for Essential. When the effect of a single +0.5 percent adjustment is multiplied by the number of directionally ambiguous OEF allowances, PIAC estimates that the overall impact of the directionally ambiguous OEF allowances is approximately $108m for Ausgrid, $73m for Endeavour and $89m for Essential. There was no submission directly controverting those arithmetical estimates.
322 ActewAGL and Networks NSW submit that the AER’s OEF adjustment process was subjective and arbitrary.
323 Networks NSW also submits that the AER’s OEF adjustments underscore the novelty of the AER’s approach to econometric benchmarking and that the approach changed in the course of its application by that approach. Moreover, the adjustments were, in Networks NSW’s submissions:
(a) an acknowledgement by the AER that the four EI model parameters do not adequately account for differences between the DNSPs; and
(b) an “attempted bandaid on a flawed process”.
324 Further submissions by Networks NSW are to the effect that the AER did not:
(a) have regard to data supplied by Essential relating to extreme weather events or corrosive environments and by Networks NSW relating to the difficulties and cost-inefficiencies associated with trying to enforce vegetation management on landowners; and
(b) make an adjustment for Networks NSW DNSPs’ proportion of wooden poles.
325 Both PIAC and Networks NSW complain that they were not given the opportunity to be heard on the AER’s post-draft decisions to make the OEF adjustments. As earlier noted, that is not a complaint which they translate on these applications to a complaint of failing to accord procedural fairness in a way which directly enlivens a ground of review.
326 PIAC also submits that:
(a) it was erroneous and inconsistent with the selection of a lower comparison point for the AER to ascertain each OEF’s effect on opex by comparing its effect on the opex of each of the Networks NSW DNSPs to the weighted average of the effect it had on the opex of the DNSPs whose efficiency scores were equal to or above the benchmark comparison point; and
(b) the AER erred by allowing a +0.5 percent adjustment for directionally ambiguous (and immaterial) OEFs;
327 Ergon submits that the AER’s OEF adjustments do not overcome the underlying error that the AER made in considering the EI model as sufficiently robust and credible to produce a forecast opex allowance for the applicant DNSPs and otherwise adopts the submission made by Networks NSW.
328 The AER responds to the assertions that it was arbitrary in its selection of a materiality threshold for its OEF adjustments by way of a submission stating that it exercised its discretion guided by EI’s and its own expertise and that the decision to use 0.5 percent was reasonable for the following reasons:
(a) the 0.5 percent threshold is substantially lower than the level (5 percent) required by accounting principles and the selection of a low materiality threshold increased the detail in which it examined the effects of OEFs, which ultimately resulted in a more accurate determination; and
(b) the same threshold was chosen for the RIN process in order to engender a high degree of confidence in the cost data applied in the RINs (DNSPs calculated the difference that their previous and current cost allocation methods would produce in reported opex and, if the figure was above 0.5 percent, they were required to backcast their costs using the current method).
329 The AER’s response is more an explanation of why it choose the figure of 0.5 percent rather than a justification of its reasonableness. As Networks NSW rightly submits:
No reasonable regulator could accept, as the AER does, that its benchmark model does not properly account for many OEFs without then seeking to take those into account.
330 ActewAGL makes a similar pertinent submission – the 23 percent OEF adjustments that the AER made to ActewAGL’s efficiency score generated by the EI model demonstrates that even the AER considers there are serious comparability issues that cannot be addressed by the specification of the EI model. In that context, it submits, it is difficult to understand how the AER considers it to be reasonable to place reliance on the outcome of that model as the sole determinant of ActewAGL’s opex allowance.
331 In support of its submission that the OEF adjustments are arbitrary, ActewAGL cites the following passage from, the Huegin Report (at pp 51 and 52):
[t]here is ... no detailed analysis or explanation of the justification for deeming other variables insignificant in the draft decision to support the AER’s claim that only a few of the factors have a material effect on total opex
… … …
For all the consideration of environmental variables individually in the draft decision, the adjustment amount allowed for the collective influence of these variables is merely a subjective estimate.
332 ActewAGL rightly submits that Huegin’s analysis applies equally to the approach taken by the AER in the Final Decisions.
333 Moreover, even if the OEF adjustments were properly quantified, the manner in which they have been applied is flawed. As ActewAGL submits, adjustments, where required, should be made before modelling, by normalising the data set, rather than ex post modelling. In support of its submission it cites the following passage from the CEPA Report Benchmarking and setting efficiency targets for the Australian DNSPs: ActewAGL Distribution, January 2015 (CEPA ActewAGL Report), (at pp 10-11) which concludes that it would be more appropriate to make the adjustments before the modelling as inconsistent data may be affecting the modelling:
Economic Insights has taken account of these adjustments and proposed that the frontier for AAD [ActewAGL] could be adjusted by 30% as a result. While I do not disagree that adjustments should be made where data are inconsistent, given the magnitude of the adjustments proposed by Economic Insights I consider that it would be more appropriate to make these adjustments before modelling … as the inconsistent data are likely to affect the modelling.
334 The conclusion is apt. If adopted by the AER it would bring its approach more in to line with Ofgem’s as outlined in the above quoted passage from its RIIO-ED1: Final determinations for the slow-track electricity distribution companies Business plan expenditure assessment, 28 November 2015.
335 As submitted by ActewAGL, the difficulty with the AER’s approach is that despite making post modelling OEF adjustments, the efficiency scores of the EI model have been affected by the inclusion of non-comparable data. Post modelling adjustments do not address the fact that the costs relationships within the model, including those DNSPs for which no OEF adjustments have been made, have been affected by the non-comparable data. Thus, those cost relationships are skewed by heterogeneous differences between the DNSPs. The output of the model is therefore skewed by flawed data. That skewed cost relationship cannot be corrected by post modelling OEF adjustments made to some only of the DNSPs (ie the three Networks NSW DNSPs and ActewAGL).
336 ActewAGL concluded its submission on this point noting that while EI has recognised the importance of making adjustments to data before modelling to create a comparable data set, EI failed to do so in its work for the AER leading to the Final Decisions. In ActewAGL’s view, the reason that EI did not do so appears to be because it was not possible to normalise the data used within the EI model. There is, therefore, in ActewAGL’s opinion, a recurring theme in EI’s approach: its methodology for assessing the DNSPs opex was driven by data (and in particular, the use of international data), rather than an a priori decision about what would provide the most robust means of assessing each DNSP’s opex. While ActewAGL acknowledges that is understandable at a practical level, it submits that it does not alter the fact that those practical limitations have diminished the probative value of the model to the point of non-existence.
337 Expanding on its submission that the AER erred by allowing a +0.5 percent adjustment for directionally ambiguous adjustments, PIAC submits that the only approach reasonably open to the AER in relation to such OEFs was to make no adjustment at all.
338 It is PIAC’s submission that:
(a) the possibility that a DNSP’s base-year opex might be affected indiscernibly – in one direction or the other, if at all – by such an OEF is something for which the AER’s lowering of the comparison point provided more than an adequate allowance; and
(b) the AER exercised its discretion incorrectly and made an unreasonable decision which does not contribute to the NEO.
339 In support of its submission PIAC cites the following extract from the transcript of the consultations held pursuant to s 71R of the NEL at p 36 which records the view of the Energy Users Association of Australia’s representative, Mr Hugh Grant as follows:
In essence, the AER’s approach to those adjustments … in my view … are arbitrary, unprincipled and in many cases illogical. And it’s very important to note that the NSW distributors are also challenging the logic of those adjustments, labelling them as fundamentally flawed, unreliable, unreasonable and arbitrary.
340 The AER response to PIAC’s criticism of its allowance of a +0.5 percent for directionally ambiguous OEFs is that it was an appropriate conservative approach, consistent with the RPP in s 7A of the NEL, that would ensure that the DNSPs would have a reasonable opportunity to recoup at least the efficient costs incurred as a result of those OEFs. That response is drawn from Attachment 7 to the Ausgrid Final Decision at p 7-182 where the AER observes:
In future, as our information set improves we may reconsider our approach to immaterial OEFs.
341 There is merit in PIAC’s reply to the AER’s submission that the AER has not endeavoured to defend the adjustment for directionally ambiguous OEFs as an adjustment that was justified or necessary in order to ensure that it would make the decision that it was satisfied would advance the NEO to the greatest degree.
342 It is also to be noted that the AER’s response is an acknowledgement by it of the immaturity of its data. An acknowledgement that should have alerted it to the vagaries of relying on the data to the extent that it did.
343 While, perhaps, the AER’s citation of the Second EI Report’s observation that “… more detailed and improved estimates are now incorporated for factors with small impacts” provides a partial answer to PIAC’s submission that the AER’s approach on directionally ambiguous OEFs was not based on a recommendation by EI, it does not address PIAC’s submissions that EI’s advice in the Second EI Report at p 98 and in Table 5.1 was that:
(a) any adjustment should be “positive for factors that disadvantage the DNSP being reviewed and negative for those that advantage it”; and
(b) in EI’s tabulation of the OEF adjustments, the directionally ambiguous OEFs were swept up into a category titled “accumulated other factors”, about which EI offered no further explanation.
344 PIAC’s submission that the directionally ambiguous OEFs have an “unobservably small” impact elicits a response from the AER that that: “… the direction of the advantage of these OEFs could not be ascertained not necessarily because their effects were unobservably small, but because sufficient data about their effects had not been provided to the AER” – a response that confirms the applicants’ concerns about the adequacy of the data upon which the AER founded its benchmarking. Indeed, another acknowledgement by the AER of the immaturity of its data that should have put it on notice of the consequences of relying on such data.
345 That confirmation of the applicants’ concerns about the adequacy of the data is reinforced by the AER’s response to PIAC’s submission that the AER erred by applying an adjustment where detailed cost data was unavailable – again, the response advances “deficiencies in the information provided by DNSPs” to explain why exact quantification of the OEFs was not always possible.
346 A further flaw advanced by PIAC in its challenge to the AER’s adjustments for immaterial OEFs is that:
(a) it defines the materiality threshold so as to categorise an OEF as immaterial if it produces a cost advantage / disadvantage of –0.5 percent;
(b) but in making a quantitative adjustment for each immaterial OEF, the AER applies an adjustment of ±0.5 percent of base year opex, ie at the outer bounds of the range of immaterial opex impacts
347 Thus, as PIAC correctly submits, by the AER’s own definition, it is to be expected that:
(a) insofar as their direction can be ascertained, the immaterial OEFs will, on average, have an absolute opex impact of less than ±0.5 percent;
(b) the directionally ambiguous OEFs will, on average, have a zero opex impact;
(c) but the AER gives no explanation why it applied adjustments at the outer bounds of the immaterial range for immaterial OEFs that would, by definition, have absolute magnitudes only less than (or equal to) those outer bounds.
348 Citing the following passage from submissions to the consultations held pursuant to s 71R of the NEL by the Major Energy User’s representative, Mr David Headberry at transcript p 116, PIAC submits that it was reasonable for the AER to make some minor adjustment of the comparison point below the efficiency frontier, in order to make allowance for inherent limitations in the benchmarking model.
The AER initially used an average of the highest quintile [sic] of the benchmark outcomes to set that efficient allowance … and, again, when you look at the highest quintile the firms that are in that highest quintile include South Australia Power Networks, Powercor, CitiPower, United Energy; all of which have got similar characteristics to those in the NSW networks. So if you accept that there’s a problem that the NSW networks’ opex is too high and they’re not efficient, how do you go about setting the right number? And my view is that the approach that was used initially by the AER of using the average of the highest quintile is probably about right because it has a mix of the CBD, urban, regional cities and also rural and remoter areas. And when you look at the mix of the networks that are included in that top quintile they are not too bad a surrogate for what we actually see in NSW.
349 But, PIAC went on to submit, having made an adjustment to the comparison point, the AER was wrong to compare the effect the OEF had on the opex of a DNSP under consideration to the weighted average of the effect it had on the opex of the DNSPs whose efficiency scores were equal to or above the benchmark. This it submitted is inconsistent with the AER’s decision to use the single efficiency score of AusNet, the DNSP at the bottom of the upper third of DNSPs as the comparison point.
350 PIAC expanded on its submissions under this heading by outlining what it saw as the following effects of lowering the comparison point:
(a) the initial reduction from the efficiency frontier (CitiPower, at 0.950) to the benchmark comparison point was more than doubled from the Draft Decision to the Final Decision:
(b) the new OEF unadjusted benchmark comparison point is lower than the post-OEF-adjustment comparison point used for each of the Networks NSW DNSPs in the Draft Decision, namely, 0.786.
(c) thus, when the AER made the final decision OEF adjustments, it started from a lower comparison point than when making the OEF adjustments in the Draft Decisions.
351 Responding to PIAC’s submission under this heading the AER drew on Attachment 7 to the Ausgrid Final Decision (at p 7-183) to explain that:
(a) it used the average of the top five DNSPs for the OEF process because in that process it is necessary to estimate an OEF’s effect on the opex of an efficient DNSP and such effects may vary between DNSPs – it cannot be assumed that the effect of an OEF on the opex of the DNSP at the benchmark comparison point would be replicated in other efficient DNSPs;
(b) if it were to be so assumed, it could lead to OEF adjustments that unfairly advantage or disadvantage DNSPs; and
(c) thus, an average is more accurate than a comparison to a single firm.
352 Rather than challenging the premise of PIAC’s submission that the AER’s averaging approach artificially inflates the impact of the OEFs and improves Networks NSW DNSPs’ apparent relative efficiency, the AER submitted that there is no evidence that its approach affects their efficiency scores positively rather than negatively.
353 Finally, in response to PIAC’s submission that the lowering of the benchmark comparison point was sufficient to allow for any potential modelling and data uncertainties, including the immaterial OEFs, the AER repeated that it adopted its approach to immaterial OEFs in order to provide DNSPs with the opportunity to recoup at least their efficient costs, consistently with the RPP – a submission which, for reasons outlined above, was rightly exposed as tenuous by PIAC.
354 Having regard to the conclusions that may be drawn from the above considerations of the parties’ submissions (particularly those of PIAC) challenging the AER’s approach to determining the OEFs, it is not necessary consider the following challenges to particular OEFs , other than to note them:
(a) PIAC’s claim that the AER’s reduction of its estimate of the quantum of the advantage that Networks NSW DNSPs enjoyed vis á vis the Victorian DNSPs (from 2.4 percent in the Draft Decisions to 0.5 percent in the Final Decisions;
(b) PIAC’s claim that the AER erred in applying information provided by Essential in relation to fungal decay in wooden poles to Ausgrid and Essential;
(c) claims by PIAC and by Networks NSW that they were not given an opportunity to be heard on the AER’s OEF adjustments; and
(d) Networks NSW’s claim that the AER erred by not making an adjustment for a DNSP’s portion of wooden poles.
355 Each of Endeavour, Essential, and ActewAGL challenge the AER’s assessment that the vegetation management costs in their respective 2012-13 base year opex were not efficient.
356 The gravamen of Endeavour’s challenge is premised on the AER’s refusal to provide an allowance for increased vegetation management occasioned by its retendering of outsourced contracts for that purpose. Endeavour claims that the AER’s refusal results in a reduction of some $240.7m in Endeavour’s forecast of its vegetation management opex.
357 The essence of Essential’s challenge is premised on a failure by the AER to examine whether the forecast reasonably reflects the opex criteria because it took the view that Essential’s forecast of a reduction in 2014-19 vegetation management costs was an admission of past vegetation management inefficiencies. In its regulatory proposal, Essential forecast a $151m reduction based on it implementing efficiencies identified in a report by Select Solutions, Review of Essential Energy Vegetation Management Strategy, 22 March 2013 (the Select Solutions Report). While Essential maintained the reduction in its revised regulatory proposal, it increased its opex forecast by $67m for rectification of non-compliant clearance levels and other deficiencies identified as a result of the introduction of new technology
358 In summary, ActewAGL’s challenge is directed at the AER’s conclusion that a report by Energy Market Consulting associates (EMCa), Review of ActewAGL Distribution’s Labour Resourcing and Vegetation Management Practices at 2012/13, April 2015, (the EMCa Report) confirmed a finding in the ActewAGL Draft Decision that its labour and vegetation management costs are likely drivers of its poor benchmarking performance.
359 It is Endeavour’s contention that its vegetation management opex forecast is:
(a) required to enable it to comply with a NSW industry standard, Industry Safety Steering Committee 3 Standard – Guidelines for Managing Vegetation Near Powerlines, December 2005 (ISSC 3); and
(b) directed at meeting the opex objectives in r 6.5.6(a)(2), (3) and (4).
360 Endeavour claims that it has been seeking to improve its compliance with ISSC 3 since 2009 but that by 2011-12 its compliance level was only 76 percent because contrary to their contracts, some contractors were trimming vegetation only to the minimum clearances and were not making any allowance for regrowth as required by ISSC 3. Thus, commencing in 2011-12, it sought new contracts by way of competitive tender. It is Networks NSW’s contention that the new contracts resulting from a tender process resulted in an increase of contractor opex from $21.3m for 2012-13 to $37.4m in 2013-14. The new contracts (together with increased opex costs attributable to internal vegetation management work, internal processes and overheads associated with those contracts) resulted in the compliance level of 92 percent in 2013-14.
361 Endeavour considers the increase in opex required to comply with the safety and reliability standards in ISSC 3 is within r 6.5.6(c)(3), ie a realistic expectation of the demand forecast and cost inputs required to achieve the opex objectives.
362 Relevantly, the opex objectives speak of the forecast opex which the relevant DNSP considers is required to achieve each of the opex objectives including:
(1) comply with all applicable regulatory obligations or requirements associated with the provision of standard control services;
(2) to the extent that there is no applicable regulatory obligation or requirement in relation to:
(i) the quality, reliability or security of supply of standard control services; or
(ii) the reliability or security of the distribution system through the supply of standard control services,
to the relevant extent:
(iii) maintain the quality, reliability and security of supply of standard control services; and
(iv) maintain the reliability and security of the distribution system through the supply of standard control services;
363 In support of its vegetation management opex forecasts, Networks NSW:
(a) relied on its competitive tendering process to demonstrate efficiency in contractor costs; and
(b) provided to the AER a copy of an internal analysis (an attachment to a Networks NSW Executive Leadership Group Meeting, dated 17 October 2013) which, Networks NSW submits, confirms that there had been an increase in Endeavour’s vegetation management costs due to a need to manage contractors more closely and enforce clearance standards.
364 In Attachment 7 to the Endeavour Final Decision, the AER dismissed the competitive tendering process at p 7-297 as follows:
competitive process only helps in demonstrating the efficiency of the contracts that were tendered. Whether this amount reasonably reflects the prudent and efficient costs of complying with regulatory obligations will depend on whether the scope of works listed in the contracts reflects a prudent scope of works.
365 Attachment 7 to the Endeavour Final Decision at p 7-297 also dismissed the Networks NSW analysis as suggesting that Endeavour’s:
… vegetation management practices are too risk averse, … [and] … are contributing to relatively high vegetation management costs.
366 As observed above, the AER was not prepared to classify Endeavour’s vegetation management opex forecast as a “step change” in terms of its EFA Guideline – see Attachment 7 to the Endeavour Final Decision at p 7-288ff where it sets out its reasons why it was not prepared to accept that opex forecast as a step change.
367 The 2013 EFA Guideline at pp 11 and 27 describes step changes and its approach to them in two places as follows:
Our approach is to separately assess the prudence and efficiency of forecast cost increases or decreases associated with new regulatory obligations and capex/opex trade-offs. For capex/opex trade-off step changes, we will assess whether it is prudent and efficient to substitute capex for opex or vice versa.
For step changes arising from new regulatory obligations, we will assess (among other things):
• whether there is a binding (that is, uncontrollable) change in regulatory obligations that affects their efficient forecast expenditure
• when this change event occurs and when it is efficient to incur expenditure to comply with the changed obligation
• what options were considered to meet the change in regulatory obligations
• whether the option selected was an efficient option––that is, whether the DNSP took appropriate steps to minimise its expected cost of compliance from the time there was sufficient certainty that the obligation would become binding
• when the DNSP can be expected to make the changes to meet the changed regulatory obligations, including whether it can be completed over the regulatory period
• the efficient costs associated with making the step change
• whether the costs can be met from existing regulatory allowances or from other elements of the expenditure forecasts.
We will assess changes in regulatory obligations in the context of the core category they affect, which will ensure consistency across DNSPs. Accordingly, DNSPs must allocate step changes arising from regulatory obligations to our expenditure categories (for example, augmentation, replacement, vegetation management).
We will not allow step changes for any short-term cost to the DNSP of implementing efficiency improvements in expectation of being rewarded through expenditure incentive mechanisms such as the EBSS. We expect DNSPs to bear such costs and thereby make efficient trade-offs between bearing these costs and achieving future efficiencies.
4.3 Step changes
Step changes may be added (or subtracted) for any other costs not captured in base opex or the rate of change that are required for forecast opex to meet the opex criteria.
We will assess step changes in accordance with section 2.2 above. Step changes should not double count costs included in other elements of the opex forecast:
• Step changes should not double count the costs of increased volume or scale compensated through the output measure in the rate of change.
• Step changes should not double count the cost of increased regulatory burden over time, which forecast productivity growth may already account for. We will only approve step changes in costs if they demonstrably do not reflect the historic 'average' change in costs associated with regulatory obligations. We will consider what might constitute a compensable step change at resets, but our starting position is that only exceptional events are likely to require explicit compensation as step changes. Similarly, forecast productivity growth may also account for the cost increases associated with good industry practice.
• Step changes should not double count the costs of discretionary changes in inputs. Efficient discretionary changes in inputs (not required to increase output) should normally have a net negative impact on expenditure.
If it is efficient to substitute capex with opex, a step change may be included for these costs (capex/opex trade-offs).
368 In rejecting the increase the AER observed in the Endeavour Final Decision – Attachment 7 at p 7-288 that:
(a) Endeavour had stated that it did not face any change to the minimum risk standards with which it must comply;
(b) without persuasive evidence that a DNSP’s total historical opex was too low to achieve the opex objectives, it did not consider increased contract costs to be a reason to increase the total opex forecast to meet what are unchanged regulatory obligations;
(c) Endeavour does not benchmark well when compared to other DNSPs in the national electricity market; and
(d) it would be inconsistent with the application of Endeavour’s EBSS to include the vegetation management opex in the opex forecast (Endeavour proposed that it retain gains from its EBSS rather than share them with its customers).
369 It is Networks NSW’s submission that the AER was wrong to analyse the proposed vegetation management opex in terms of it being a “step change” because it lead it to an ex ante view that that opex might only be accepted if it were the result of an opex / capex trade-off or new regulatory obligation whereas the correct approach was to first consider whether that opex reasonably reflected the r 6.5.6(c) opex criteria.
370 The AER’s response to that submission notes that its approach to assessing step changes, as set out in the Endeavour Final Decision, adopts the approach in the 2013 EFA Guideline and cites the following passage from Attachment 7 to the Endeavour Final Decision at p 7-291:
We only include a step change in our alternative opex forecast if we are satisfied a prudent and efficient service provider would need an increase in its opex to reasonably reflect the opex criteria.
That question, the AER submits, is precisely directed to the AER’s task under r 6.5.6(c).
371 While that question may be so directed, a reading of Attachment 7 to the Endeavour Final Decision reveals that what may be described as a step change analysis was very much in the forefront of the AER’s considerations of Endeavour’s vegetation management opex forecast in applying its benchmarking methodology.
372 Thus, the AER did not consider the fact that outsourcing its vegetation management contracts by way of competitive tender was sufficient to demonstrate that Endeavour’s vegetation management opex forecast reasonably reflects the r 6.5.6(c) opex criteria or the opex objectives. Thus too it dismissed the abovementioned internal analysis.
373 In circumstances where Endeavour:
(a) provided its abovementioned analysis to the AER; and
(b) is committed to paying contractors retained through a competitive tender process so that it may comply with an applicable regulatory obligation,
the AER’s reasons for rejecting Endeavour’s vegetation management opex forecast are tenuous. A fortiori, having regard to the significant adverse consequences that may flow from a failure to comply with regulatory vegetation management requirements as demonstrated by the Victorian bushfires.
374 The AER also submitted that its assessment was influenced in part by a Deloitte Report NSW Distribution Network Service Providers Labour Analysis, Final Addendum to 2014 Report, 28 April 2015 (the 2015 Deloitte Labour Report). The AER previously engaged Deloitte to conduct an analysis of the Networks NSW DNSPs’ labour costs in the 2009-2014 regulatory period. Its report, NSW Distribution Network Service Providers Labour Analysis, 17 November 2014 (the 2014 Deloitte Labour Report) informed the AER’s assessment of the DNSPs’ 2015-2019 capex and opex forecasts, and was referenced in the Draft Decisions for each of the DNSPs.
375 The 2015 Deloitte Labour Report concluded:
Our view remains that the NSW DNSPs have higher labour costs than their peers (driven by the number of employees rather than costs per employee) due to in [sic] restrictive EBA provisions, a high degree of unionisation and inefficient labour practices, which means that their base year opex was not efficient.
376 Insofar as the 2014 Deloitte Labour Report may have informed the AER’s Draft Decision Networks NSW had an opportunity to respond to it. It is, however, Networks NSW’s unchallenged submission that it was not given the opportunity to be heard in relation to the 2015 Deloitte Labour Report.
377 As noted elsewhere, of itself s 71C of the NEL does not provide as a ground of review that procedural fairness was not accorded. It is understandable that, with the range of reviewable regulatory decisions being made by the AER and the extensive refined obligations imposed on it by the 2012 Rule Amendments, the regulatory period 2015-19 presented significant administrative challenges. The timeframe in which the DNSPs and the AER must operate in relation to the commencement of the relevant regulatory control period is set out in rr 6.8 to 6.11 of the NER. Recognising those time constraints, r 6.11.1(c) imposes a “best endeavour” obligation on the AER to publish any post Draft Decision analysis for the purposes of its Final Decision.
378 There is presently no basis for thinking that the AER did not properly comply with r 6.11.1(c). It was not the focus of submissions. The DNSPs focus was to rely on the timing and quality of the analysis upon which they did not have an opportunity to comment as part of the picture on which a ground or grounds of review under s 71C(1) of the NEL are made out. Certainly, it is a matter of common sense that a report such as the 2015 Deloitte Labour Report might carry greater weight if it had been the subject of any response from Endeavour, depending of course on the terms of that response.
379 In rejecting Essential’s forecast of its vegetation management opex, Attachment 7 to the Essential Final Decision states (at pp 7-160 and 7-162):
… we consider the Select Solutions report (and Essential Energy documentation discussing it) submitted by Essential Energy with its regulatory proposal demonstrates there are inefficiencies in its vegetation management practices in the 2012–13 base year. While Essential may have since improved its practices, the evidence suggests it had not done so in 2012–13. Therefore, the costs in 2012–13 are overstated.
… … …
We placed most weight on the findings of the Select Solutions review, noting that Essential Energy had proposed a step down in its vegetation management opex in the forecast period (rather than in the base year). Therefore, we maintain our draft decision view that Essential Energy's performance on our economic benchmarking techniques is likely to be partly driven by its vegetation management opex.
380 It is Essential’s submission that the AER:
(a) contrary to r 6.5.6, performed no analysis, and gave no consideration to, whether Essential’s forecast vegetation management over the 2014-19 period, including the cost reductions proposed, was efficient or prudent;
(b) made no attempt to quantify any inefficiencies in terms of the effect on opex of the vegetation management issues identified, thus its analysis provided no probative corroboration of inefficiencies of the scale identified by the AER in reliance upon the EI Model; and
(c) did not ascribe a figure quantifying inefficiencies in Essential’s vegetation management.
381 Thus, in Essential’s submission:
(a) the AER’s findings regarding vegetation management did not, and could not, justify a quantitative assessment of efficient opex; and
(b) the AER has conducted an abstract and superficial assessment of Essential’s vegetation management practices, focused on only one year, without actually undertaking the task of identifying the amount of inefficient expenditure.
382 It is Essential’s contention that any inefficiency that may be implied from Essential’s forecast reductions does not amount to the inefficiency identified by the AER’s benchmarking exercise. In support of its contention, Essential claims that:
(a) its proposed reduction in vegetation management costs in its regulatory proposal was approximately 16.5 percent, which drops to 9.2 percent with the increases in its revised regulatory proposal; and
(b) by comparison, the EI model identified Essential as having an efficiency score of 54.9 percent compared to a frontier (after making the OEF adjustments) of 69.4 percent, implying a 26.4 percent reduction to opex.
383 It is also Essential’s contention that the AER ignored evidence before it that Essential compares favourably to most other Australian DNSPs in terms of vegetation management opex per vegetation management span (per kilometre of overhead circuit which passes through an area requiring vegetation management) because a “service provider’s estimation assumptions seem to influence the data on maintenance spans” (Attachment 7 to the Essential Draft Decision at page 7-84). Essential notes in this regard that its consultant Advisian observed that: “it is not logical to simply ignore it in a detailed assessment of vegetation management”.
384 Essential’s challenge to the AER’s vegetation management findings concludes with a submission that what data is available on vegetation management spending per vegetation management span is wholly at odds with the conclusions the AER sought to draw from the Select Solutions Report and Essential’s proposed reductions.
385 Insofar as Essential’s challenge is premised on a supposition that the NER requires the AER to conduct a line-by-line, bottom up review of each category of forecast expenditure, it is rejected by the AER. It is the AER’s contention that:
(a) rule 6.5.6(c) requires the AER to assess whether the total of the DNSP’s forecast opex reasonably reflects the opex criteria;
(b) as described in its 2013 EFA Guideline, the AER undertakes that assessment using the base-step-trend approach, the first step being an assessment whether the DNSP’s base year opex reflects the opex criteria;
(c) the AER assessed Essential’s base year opex using a number of assessment methods, including a review of vegetation management costs as a key category of expenditure; and
(d) the purpose of the review was to investigate whether vegetation management in the base year indicated inefficiency and supported and explained Essential’s poor performance on other assessment techniques – quantification of the inefficiency was not the purpose of the review, and was not required by the AER’s assessment process.
386 It appears from Attachment 7 to the Essential Final Decision that the AER’s focus was on the Select Solutions Report and its 16 recommendations to improve Essential’s vegetation management. At p 7-162 of the Attachment, the AER observes that Essential had noted that recommendations from the report were in the process of being implemented and concluded that:
… they could not have been implemented in the 2012–13 base year. This year is the relevant year for determining the appropriateness of Essential Energy's revealed costs as the starting point for determining an estimate of efficient and prudent total forecast opex.
387 The AER dismisses Essential’s contention based on the inconsistency between the extent of the reduction to vegetation management costs proposed by Essential (9.2 percent) and the AER’s comparison to the efficient frontier (26.4 percent) as a non-sequitur.
388 The AER also dismisses Essential’s submissions that it ignored evidence that Essential performs comparatively well in terms of vegetation management opex per vegetation management span by submitting it did not ignore the evidence, it simply did not find it persuasive for the reason stated by it in Attachment 7 to the Essential Draft Decision.
389 While, as submitted by the AER, neither the NEL nor the NER mandate a line-by-line, bottom-up review of each category of forecast opex, in circumstances where benchmarking in Australia is in its infancy, sensible administration dictates that the AER should not have cast aside its previous practice of conducting bottom-up reviews in favour of the emphasis it placed on benchmarking. A fortiori, in circumstances where its preferred EI model’s reliance on overseas data and the AER’s final OEF adjustments could not have the benefit of full exposure to the consultation processes mandated by the NEL and the NER.
390 Viewed in that context, the AER’s apparently untested conclusion that the recommendations of the Select Solutions Report could not have been implemented in the 2012–13 base year and its preference for its assessment of Essential’s overall opex based on the EI model are unconvincing. Likewise, the Tribunal is not convinced by its dismissal of Essential’s submission that the AER should have quantified its vegetation management inefficiencies and should not have ignored Essential’s comparative vegetation management opex per vegetation management performance.
391 Attachment 7 to the ActewAGL Final Decision notes (at p 7-54) that:
(a) the AER’s analysis of ActewAGL’s opex categories showed it had “very high” costs on labour and vegetation management metrics compared to most of its peers;
(b) because those categories account for a significant proportion of ActewAGL's opex (labour is approximately 80 percent) the AER conducted detailed reviews of labour and vegetation management opex; and
(c) the “detailed review”, referred to in these reasons as the EMCa Review, found significant issues in those categories of ActewAGL’s opex, which the AER considered evidence of base year inefficiency, supporting its benchmarking results.
392 ActewAGL submits that the “detailed review” does not support the AER’s benchmarking, is “fundamentally flawed”, and “as a ‘qualitative assessment’ has no utility in determining the accuracy and reliability of a quantitative assessment of ActewAGL’s opex … .”
393 Expanding on its submission that the EMCa review does not support the AER’s benchmarking, ActewAGL challenges the following findings by the AER in Attachment 7 to the ActewAGL Final Decision (at p 7-146 and p 7-153):
(a) ActewAGL’s labour costs are “… driven by having too many employees rather than by cost per employee”; and
(b) “ActewAGL could potentially achieve efficiencies by outsourcing more”
on the basis that:
(c) the AER did not provide evidence to establish that a higher level of outsourcing would deliver more efficient expenditure; and
(d) the evidence of ActewAGL’s expert, Advisian, which was submitted to the AER establishes that: “… the question of whether opex or capex tasks are carried out by internal or external labour is largely irrelevant to the efficiency of the outcome.” (Advisian, Opex cost drivers: ActewAGL, January 2015, at p 96).
394 ActewAGL responds to the AER’s finding in Attachment 7 to the ActewAGL Final Decision (at p 7-153) that due to restrictions on outsourcing in ActewAGL’s Enterprise Bargaining Agreement, it does not “appear to be adopting the lowest cost option” and its conclusion (at p 7-156 based on the EMCa review) that ActewAGL’s “lack of outsourcing is a key reason why its labour costs in 2012-13 are not reflective of those of a prudent and efficient service provider”. It does so by citing the findings in a report by Australian Business Lawyers & Advisors Pty Ltd Review and comparison of ActewAGL’s enterprise agreement provisions against other electricity network service providers, 13 January 2015) that:
(a) many of the criticisms made about the operation of the ActewAGL EA [Enterprise Bargaining Agreement] are unfounded; and
(b) contrary to the conclusions reached in the AER Draft Decision, the ActewAGL EA is equivalent to and in many respects demonstrably more flexible than the norm in the electricity sector by comparison to other major electricity providers’ enterprise agreements.
395 ActewAGL challenges EMCa’s claim (at p 11 of EMCa Review) that two reports generated by ActewAGL’s consultants, support evidence of systemic issues in ActewAGL’s work practices, processes and systems that existed in 2012-2013 that have translated into material operational cost inefficiency.
396 Attachment 7 to the ActewAGL Final Decision records (at p 7-156) EMCa’s findings based on the first of those reports (Marchment Hill Consulting (MHC) report Organisation Review-ActewAGL Energy Networks, February 2011 (the MHC Report) as follows (without footnotes):
EMCa’s review included an examination of how ActewAGL runs its business. EMCa examined the MHC report, which ActewAGL commissioned in 2011. MHC found that problems exist in all areas of ActewAGL’s operations, from the way ActewAGL plans its work, through to delivery, and how it monitors and controls its performance operationally and strategically. ActewAGL advised that it had ‘rolled-out’ the majority of its 34 initiatives in response to the 26 issues identified by MHC in 2011.
EMCa disagrees with ActewAGL's view that it could have implemented the recommendations from the MHC report, which ActewAGL consider are implicit in its forecast productivity growth. In EMCa's opinion, a service provider would require 3 to 5 years to extract the full net benefits from the recommendations of the MHC report. However, ActewAGL has indicated the time period to implement these recommendations was only 6-9 months.
EMCa accept that some of the initiatives could be implemented in twelve months or less but the substantial net benefits are typically achieved over a longer time period, particularly given MHC observed that improvements were needed to all elements of ActewAGL’s Operating Model – changing the organisational structure alone would not address all of the issues sustainably.
EMCa considers that in the absence of compelling evidence – ActewAGL has not provided evidence of quantified efficiency gains – ActewAGL has not made significant efficiency gains quickly enough to offset the implementation costs by 2012–13.
397 It is, however, ActewAGL’s submission that the primary objective of the MHC Report was to understand and address performance issues identified by ActewAGL’s management over the longer term and, except for small direct salary savings, does not quantify any cost savings that would flow from an implementation of its recommendations. In support of that submission it points to a letter, dated 4 March 2015, from MHC which states (at p 3):
A close examination of the 2011 MHC report finds no explicit references to inefficiency or poor productivity associated with that Reviews’ [sic] findings.
Given that the scope of MHC’s review was intentionally wide ranging, if such concerns had been identified, they would have been noted.
MHC considered all efficiency opportunities which might flow from the recommendations made in our earlier 2011 report. We did not state an explicit efficiency benefit as these were not directly apparent from our work.
398 Based on the MHC letter, ActewAGL rightly submits that because EMCa’s analysis of the MHC review proceeds on an incorrect basis (ie that the MHC review relates to opex efficiency) there is no basis for EMCa’s conclusions with respect to the MHC review or for EMCa’s conclusion that ActewAGL’s 2012-13 labour costs were inefficient.
399 EMCa relied on the following sentence in the in the second of the reports (Sinclair Knight Merz (SKM) report Resource Planning to deliver ActewAGL’s Program of Works for the FY 2012/13, Final Report, 27 March 2012 at p 5 (the SKM Report)) to point to “a systemic issue with projects and program delivery”:
However, the business does not have a consolidated works management system making resource scheduling and forecasting on an ongoing basis, difficult.
400 That sentence, however, as ActewAGL submits (also rightly), appears in a context referring to its capex, not its opex: see the SKM Report at p 5. And, as ActewAGL goes on to point out, SKM’s conclusion with respect to ActewAGL’s opex is to the contrary, as the following extract from the SKM Report at p 7 shows:
Finally a very high level review of the operational expenditure under the AMSP [Asset Management Strategy Plan] was conducted. This revealed that most programs involving inspection and maintenance work were being achieved within reasonable tolerances for quantity and budget.
401 Thus, ActewAGL submits (again rightly) that no significant probative value should be attached to the EMCa Report and that it does not provide any support for the AER’s conclusion of inefficiencies in ActewAGL’s labour practices.
402 Nor, in its submission, does the EMCa Report support the AER’s finding at p 7-146 of Attachment 7 to the ActewAGL Final Decision that:
ActewAGL’s labour costs are driven by having too many employees rather than by cost per employee.
403 In support of that submission ActewAGL points to a statement in the EMCa Report that:
staffing levels should be determined as part of comprehensive resourcing analysis
and notes that neither the AER nor the EMCa conducted such an analysis. In ActewAGL’s submission, absent such an analysis, the AER erred in its conclusions regarding ActewAGL’s staffing levels.
404 Noting that:
(a) in the ActewAGL Draft Decision the AER relied on analysis in the 2014 Deloitte Labour Report which, despite its requests, was not provided to ActewAGL; and
(b) while the AER excised any reference to that report in the ActewAGL Final Decision, the overall conclusions reached in the Final Decision are largely identical to those expressed in the Draft Decision from which it may be inferred that the AER continued to place reliance, or at least took into account, the contents of the report in the Final Decision,
ActewAGL submits that because it was deprived of an opportunity to review and make submissions in relation to the 2014 Deloitte Labour Report, ActewAGL was denied procedural fairness and the conclusions reached by the AER about ActewAGL’s workforce practices cannot be considered to be reliable. As to procedural fairness, the Tribunal refers to its observations above. Having regard to the following paragraphs, it is not necessary for the Tribunal to decide whether the second part of that submission has merit.
405 In response to criticism from ActewAGL that the Draft Decision analysis of its vegetation management expenditure did not corroborate the AER’s benchmarking results because it did not identify at least 40 percent of ActewAGL’s vegetation management expenditure as inefficient, Attachment 7 to the ActewAGL Final Decision states (at p 7-158) that the AER was not applying the detailed review in the manner suggested by ActewAGL and that:
The evidence we present in the detailed review will not necessarily explain the entire performance gap quantified in the economic benchmarking because our intention is not to examine all of opex. Economic benchmarking techniques, on the other hand, do assess opex in totality. The detailed review helps us to identify if the benchmarking results are consistent with our more detailed examinations of ActewAGL’s opex.
406 The AER’s submissions fail in their endeavour to defend the EMCa Report by:
(a) addressing some detail of ActewAGL’s criticisms of the review; and
(b) stating that its purpose of the review was not to quantify inefficiency.
407 The EMCa Report is, in its own words at p i, no more than a “limited scope review”. To put it another way, it is but a “desk top” qualitative review which, as rightly submitted by ActewAGL, relied on, but misconstrued, earlier reports commissioned by ActewAGL.
408 Where, as here, the application of a new untested benchmarking model is applied to arrive at a total opex figure, sensible administration suggests that the regulator responsible for its application would apply some form of quantitative “reasonableness check” bottom-up analysis to at least some, if not all, of the opex components. That is, however, not the case here.
409 Networks NSW challenge the AER’s findings that inefficiencies in the NSW DNSPs’ labour management practices are, in part, responsible for the gap between them and the frontier DNSPs identified in the AER’s economic benchmarking analysis.
410 As may be seen by reference to Table 7.3: Assessment of Ausgrid’s base opex reproduced above and the paragraphs that follow that table, the findings in respect of the Networks NSW DNSPs are based on the 2014 and 2015 Deloitte Labour Reports – being the reports referenced above in canvassing the DNSPs’ challenges to the AER’s finding on their vegetation management practices.
411 The 2015 Deloitte Labour Report concluded (at p 20):
… the NSW DNSPs have a relatively high number of employees compared to private DNSPs in the NEM [national electricity market]. Considering that labour costs represent the vast majority of opex and given that their unit labour costs do not appear to be greater than their peers’, the higher number of employees in Ausgrid, Endeavour and Essential is likely the primary factor driving high opex costs per customer in NSW. Although the number of employees in NSW DNSPs is high due to historical workforce decisions, the current high number of employees is likely being sustained by restrictive EBA provisions relating to no forced redundancies and a relatively high proportion of employees employed under EBAs.
412 A footnote to the above quoted passage from the 2015 Deloitte Labour Report noted that while the “no forced redundancies” provisions is not unique to the Networks NSW DNSPs, the fact that they currently have large workforces makes the provisions more relevant as they impede any large reduction in workforce size.
413 It is Networks NSW’s submission that because the vast majority of the Networks NSW DNSPs workforce are engaged pursuant to Enterprise Bargaining Agreements (EBAs) and compliance with the EBA is a “regulatory obligation or requirement” within the meaning of r 6.5.6(a)(2), the AER’s analysis of Networks NSW’s labour costs do not establish that the DNSPs’ revised regulatory proposals include any labour costs which do not reasonably reflect the operating expenditure criteria.
414 Rule 6.5.6(a)(2) provides:
(a) A building block proposal must include the total forecast operating expenditure for the relevant regulatory control period which the Distribution Network Service Provider considers is required in order to achieve each of the following (the operating expenditure objectives):
… … …
(2) comply with all applicable regulatory obligations or requirements associated with the provision of standard control services;
415 The phrase regulatory obligations or requirements is relevantly defined in s 2D(b)(v) of the NEL as follows:
an Act of a participating jurisdiction, or any instrument made or issued under or for the purposes of that Act … that materially affects the provision, by a regulated network service provider, of electricity network services that are the subject of a distribution determination or transmission determination.
416 Networks NSW submits that the NSW DNSPs’ obligations to comply with the EBAs constitute a regulatory obligation or requirement in terms of s 2D(b)(v) of the NEL as the EBAs are made pursuant to Part 2 - 4 of the Fair Work Act 2009 (Cth) – that Act being an Act of a “participating jurisdiction” which obliges the Networks NSW DNSPs to comply with the EBAs. Networks NSW advances its submission by citing Toyota Motor Corporation Australia Limited v Marmara  222 FCR 152 at  and Teys Australia Beenleigh Pty Ltd v Australasian Meat Industry Employees Union  317 ALR 636 at  in support of its contention that the EBAs are not mere contractual agreements; they are specific instruments made under a detailed regime and enforceable only as provided by the Fair Work Act 2009 (Cth).
417 Attachment 7 to the Ausgrid Final Decision shows that the thrust of the AER’s decision to reject the EBAs as a regulatory obligation or requirement is that:
(a) they are a creature of Commonwealth law; and
(b) the Commonwealth is not a “participating jurisdiction”.
418 Noting that in its Final Decisions the AER maintained the view that the Commonwealth is not a “participating jurisdiction”, Networks NSW draws on the Minister’s intervention in these proceedings to submit that the AER now acknowledges that status. It is, however, unnecessary to delve further into whether the Minister’s intervention amounts to a concession on the part of the AER. That is because the EBAs may be reasonably regarded as:
(a) otherwise required to achieve an opex objective, namely, the r 6.5.6(a)(4) objective to: “maintain the safety of the distribution system through the supply of standard control services”; and
(b) reasonably reflecting the opex criteria in r 6.5.6(c)(3): “a realistic expectation of the demand forecast and cost inputs required to achieve the operating expenditure objectives.”
419 That the EBAs may be so regarded may be seen in the following paragraphs (without footnotes) from Attachment 7 to the Ausgrid Final Decision (at p 7-86) in which the AER, while rejecting the EBAs as a 6.5.6(a)(2) “regulatory obligation or requirement’, recognised that the EBAs may affect the Networks NSW DNSPs’ provision of standard control services:
We also disagree with the service providers’ submissions that compliance with the terms of their own EBAs is a ‘regulatory obligation or requirement’. For example, service providers have referred to redundancy costs ‘required to be paid as a regulatory obligation’.
… of the six possible (and exhaustive) categories of obligations or requirements … , EBAs could conceivably only fall with an Act or instrument made or issued that ‘materially affects a service provider's provision of electricity network services’. This is because the terms of an EBA could plausibly materially affect a service provider’s provision of standard control services. However, that Act or instrument must be made by a ‘participating jurisdiction’. Given a participating jurisdiction must have passed a version of the NEL, an EBA made under the Commonwealth’s Fair Work Act 2009 appears to be imposed by a law other than of a participating jurisdiction. Further, the terms of an EBA itself are not contained in the Fair Work Act 2009.
420 Consistent with the above quoted extract from Attachment 7 to the Ausgrid Final Decision, the AER’s submissions do not contend that the EBAs, and labour costs more generally, are irrelevant to its assessment of required forecast opex. Indeed, the AER submits that the opex criteria include a realistic expectation of cost inputs and labour costs are one such input.
421 The AER’s recognition that the terms of an EBA “… could plausibly materially affect a service provider’s provision of standard control services” enlivens Networks NSW’s submission that its opex allowances must be such as to permit them to comply with their obligations under the EBAs. This, Networks NSW submits, includes (but is not limited to) making sufficient allowance for redundancy payments that Networks NSW will be required to pay under the EBAs in relation to the forecast reductions in employee numbers expected over the 2014-19 period.
422 However, as the following extract (without footnotes) from pp 7-41ff of Attachment 7 to the Ausgrid Final Decision illustrates, the AER rejects such submissions:
Consistent with our approach in our draft decision, we do not agree with these submissions. We are not denying the service providers the ability to transform their businesses and pay staff their entitlements. Recruitment and removal of staff are both ‘legitimate costs’ that the service providers would need to incur. However, we do not ‘fund’ the service providers for these (or any specific) activities. We assess a service provider's revealed opex in order to form a view on whether it reasonably reflects the opex a prudent and efficient (objective) service provider would require in the future to comply with its obligations. Service providers have broad discretion about all contractual arrangements and the manner in which they carry out those obligations.
423 Notwithstanding the AER’s statement that: “Recruitment and removal of staff are both ‘legitimate costs’ that the service providers would need to incur.” and its submission that labour costs are relevant to its assessment of required opex, its focus on benchmarking (in particular the EI model and its total opex outcomes) have lead it to treat the EBAs as endogenous (rather than exogenous) – an endogenous factor to be ignored in the AER’s estimate of the total required opex made pursuant to r 6.12.1(4)(ii).
424 The AER’s approach to endogenous factors is illustrated at p 7-184ff of Attachment 7 to the Ausgrid Final Decision where the AER observed (without footnote):
Differences in work practices and operating techniques are endogenous. The AEMC provides guidance on what it considers to be an endogenous factor that should not be taken into account when benchmarking. It stated:
Endogenous factors not to be taken into account may include:
• the nature of ownership of the NSP;
• quality of management; and
• financial decisions.
Differences in opex due to work practices and operating techniques are a direct outcome of management decisions. Therefore we do not provide an OEF adjustment for them. In general we consider that any OEFs that are a result of the quality of management do not meet the exogeneity OEF criterion. [Emphasis added]
425 While the extract from p 113 of the 2012 Rule Amendments determination quoted by the AER in the above extract provides some support for the AER’s reasoning, the AER’s transformation of the AEMC’s “may” to “should not” somewhat overextends the AEMC’s guidance. As a preceding paragraph to the above extract from p 113 of the the 2012 Rule Amendments shows, the AEMC’s view on when endogenous factors may or may not be taken into account is not an inflexible rule:
The final rule gives the AER discretion as to how and when it undertakes benchmarking in its decision-making. However, when undertaking a benchmarking exercise, circumstances exogenous to a NSP should generally be taken into account, and endogenous circumstances should generally not be considered. In respect of each NSP, the AER must exercise its judgement as to the circumstances which should or should not be included. [emphasis added]
426 While the AER’s submissions recognise that it is not an absolute rule, that is not how it was applied vis á vis the EBAs.
427 Thus, although the EBAs may lack either the NEL’s s 2D jurisdictional foundation or the genus of a safety or reliability standard etc of a r 6.5.6(a)(3) “regulatory requirement or obligation”, the Networks NSW DNSPs are bound by their EBAs as a matter of law. Unlike a contract, which according to its terms may be terminated, an EBA continues in force until its nominal expiry date after which it may, with the approval of the Fair Work Commission, be terminated by agreement between an employer and the employees it covers (ss 219-224 of the Fair Work Act 2009 (Cth)). Absent agreement, an application must be made to the Fair Work Commission to terminate an EBA. Termination may only occur if the Commission is satisfied that to do so is not contrary to the public interest and is appropriate in all the circumstances (ss 225-227 of the Fair Work Act 2009 (Cth)).
428 After reviewing the DNSPs’ revised regulatory proposals, the 2015 Deloitte Labour Report found (at p 16) that the primary driver of the NSW DNSPs’ labour costs being higher than their peers is the number of employees rather than cost per employee.
429 It appears from the 2015 Deloitte Labour Report that the higher number of employees is attributable to changes to Ministerial licence conditions in 2005 and in 2007 which placed considerable pressure on the NSW DNSPs during the 2009-14 regulatory period. In particular, clause 14.2 of the 1 December 2007 Design, Reliability and Performance Licence Conditions For Distribution Network Service Providers required that the NSW DNSPs be:
…as compliant as reasonably practicable with the applicable design planning criteria in Schedule 1 in relation to all network elements by 1 July 2014; and fully compliant with the applicable design planning criteria in Schedule 1 in relation to all network elements by 1 July 2019.
430 The 2015 Deloitte Labour Report noted at p 2ff that:
The 2014 Report set out our view that, given the licence requirement to be ‘as compliant as reasonably practicable’, the DNSPs acted in a manner consistent with a prudent and efficient DNSP by aiming to be largely compliant by 2014. Had they not strived to do so, and particularly had a major network incident occurred that could have been avoided had compliance with the new standards been achieved, the DNSPs would rightly have been criticised.
431 The report also agreed with the Networks NSW DNSPs’ view that the EBAs as a whole are no more generous in terms of base level wages and other employee conditions than those of their peers – that agreement being qualified by a note that the EBAs contain a range of generous terms and a citation of the following passage from Essential’s Revised Regulatory Proposal:
In general we agree with the observations made in the  Deloitte Report that high levels of unionisation in the electricity supply sector can result in more restrictive work practices which are difficult to remove once negotiated in enterprise agreements. This can lead to relatively inflexible, high cost and unproductive work practices once labour costs become entrenched in EBAs.
432 The 2015 Deloitte Labour Report also found (p 18) that the majority of distributors are not allowed to carry out forced redundancies as a result of provisions in their respective EBAs and that this is an important impediment to any program of reductions in workforce size, outside of natural attrition.
433 Networks NSW rightly submit that insofar as the 2014 and 2015 Deloitte Reports suggested inefficiencies in the NSW DNSPs’ labour practices the reports do not quantify those inefficiencies and provide no corroboration of inefficiencies of the scale identified by the AER in reliance on the EI model.
434 As Networks NSW submit, Ausgrid, Essential and Endeavour are bound by the EBAs and remain bound by them and they should not be viewed as an endogenous managerial choice. At least not in circumstances where the AER has quite radically shifted from an itemised bottom-up approach to assessing opex to benchmarking total opex per se – particularly where that benchmarking has not been exposed to the rigors of the consultation the NEL and NER envisage for such a radical change.
435 The AER having flagged its approach to EBAs may be better placed to defend such an approach to an EBA when its approach to benchmarking is on a firmer footing and where there is hard information to support a finding that a DNSP’s labour practices are inefficient vis á vis its peers. But, having regard to the 2015 Deloitte Labour Report, that is not the case here. Here the Networks NSW DNSPs are shackled with EBAs that effectively restrict their ability to efficiently reduce their workforce in the regulatory period – that restriction being attributable to an exogenous factor, namely, the Fair Work Act 2009 (Cth).
436 It may be said that, in the view of the Tribunal, it is the policy of the legislative arm of government that, to the extent that the EBA’s are (if they are) an inefficient imposition on the DNSPs, nevertheless they are a cost to be borne by the consumers of electricity. The AER may, of course, assess the extent of inefficiency reflected by the number of employees. It may review the terms upon which the number of employees may be reduced under the EBAs. It may consider the timing for the expiration of the EBAs. But, having regard to the regulatory prescriptions, the Tribunal does not accept that it may, by the use of the EI model, simply select the measurement of efficiency which it did in this respect without regard to the obligations under the EBAs as they presently exist. Over time, and probably during the new current regulatory period, any such inefficiencies as the AER considers to exist may progressively be reduced by the reduction in employee numbers to what the AER considers to be the efficient number, and any allowances under the EBAs (as they expire) which the AER considers to be inefficient may also by the same elapse of time be reduced to an efficient level.
437 It is not necessary to canvass Networks NSW’s other grounds for challenging the AER’s labour costs decisions as set out in  of its submissions, namely:
(a) there is no proper basis for Deloitte’s conclusions that:
(i) the Networks NSW DNSPs employ too many staff; or
(ii) that their past practice of hiring permanent labour left them with too many staff;
(c) the AER failed to take into account the efficiency programs implemented by the Networks NSW DNSPs; and
(d) in relation to Essential, the AER incorrectly weighed Deloitte’s view that there is a possibility Essential could realise significant cost savings by using a Local Service Agent model.
438 In that regard, however, it is proper to note that the Networks NSW’s submissions assert approximately $3 billion in efficiencies in capex and opex over the 2009-14 regulatory control period and that within these savings were reductions in the number of employees.
439 In Attachment 7 to the Ausgrid Final Decision at p 7-158, and correspondingly in Attachment 7 to the Endeavour Energy and Essential Energy Final Decisions, the AER concluded that the service providers had managed to achieve significant reductions in labour costs through reducing the number of staff, and were forecasting further savings, but that most of the reductions took place after the 2012/13 base year. The AER adopted its position based on the 2015 Deloitte Labour Report which stated that the scale and speed of the reductions in staff suggested there were still cost efficiencies to be realised.
440 The AER at p 7-286 of Attachment 7 to the Ausgrid Final Decision stated that the efficiency programs represented “catch up” in productivity, and efficient distributors would not be implementing the same productivity improvements.
441 The AER at p 7-52 of Attachment 7 to its Ausgrid Final Decision noted that Endeavour’s revealed opex was lower than that for Essential and Ausgrid because it had implemented its efficiency programs earlier and to a greater extent than its two peers. It still considered that there were efficiency gains to be realised in the 2012-13 base year.
442 As the 2015 Deloitte Labour Report contended that the NSW DNSPs did not have an efficient workforce in the base year and compared employee numbers across the regulatory control period with other DNSPs, the AER will have to consider how the efficiency programs implemented by the NSW DNSPs into the 2014-19 regulatory control period have been effective.
443 It is Networks NSW’s submission that the AER has used an experimental model as the sole determinant of opex, contrary to sensible regulatory practice including significant experience of modelling in other jurisdictions.
444 Networks NSW’s submission is cogent. There are lessons to be learnt from overseas regulators, particularly the UK regulator, Ofgem which the AER cites in support of its approach – see for example Attachment 7 to the Ausgrid Final Decision, at p 7-60, where, noting that that Ofgem assesses totex rather than capex and opex separately, the AER states:
Our approach of using benchmarking as a basis for making adjustments to opex is consistent with Ofgem’s approach.
445 Ofgem is a regulator with over a decade’s experience in benchmarking and, because of that long history, is the primary point of reference when it comes to assessing the soundness of another regulator’s approach to benchmarking and its benchmarking models. As the Huegin Report observes, at p 20:
The Productivity Commission report and the AER’s Guideline and the associated documents that fed into both rely heavily on the experiences of regulators such as OFGEM.
446 It is, however, Networks NSW’s submission that the AER’s approach is nothing like Ofgem’s. That submission is supported by the following observation by CEPA’s Chairman (Professor David Newbery, who has led numerous CEPA assignments for Ofgem) (see CEPA Report at p 30):
Using a top-down model to assess opex (or totex) is consistent with best practice in the UK as it does not enforce choices on the companies as to which activities to undertake, however, using only a single model with few explanatory variables and no bottom-up assessment is not best practice. For instance, Ofgem in its RIIO-ED1 decision stated:
Our use of three models [two top-down and one bottom-up] acknowledges that there is no definitive answer for assessing comparative efficiency and we expect the models to give different results. There are advantages and disadvantages to each approach. Totex models internalize operational expenditure (opex) and capital expenditure (capex) trade-offs and are relatively immune to cost categorisation issues. They give an aggregate view of efficiency. The bottom-up, activity-level analysis has activity drivers that can more closely match the costs being considered.
His reference is to Ofgem (2014), RIIO-ED1: Final determinations for the slow-track electricity distribution companies: Business plan expenditure assessment.
447 In contrast to the AER’s post modelling OEF adjustments, Ofgem adjusts to the data supplied by the Distribution Network Operators (DNOs) (the UK equivalent of DNSPs) it regulates prior to undertaking its modelling – see Ofgem’s, RIIO-ED1 Final determinations for the slow-track electricity distribution companies, 28 November 2014, at p 41:
We consider whether DNO submitted data require adjustments prior to carrying out our comparative benchmarking. This is to ensure the comparisons are on a like for like basis. Where we decide adjustments are appropriate, we adjust the DNO submitted costs before our totex and disaggregated assessments. These adjustments fall into four broad categories: regional labour costs; company specific factors; exclusions from totex models; and other adjustments.
448 As the Huegin Report notes at p 23:
The OFGEM approach is … based on many years of regulatory reporting to a consistent format and common reporting timeframes which are more favourable conditions for data accuracy … [than the Australian staggered reporting and/or regulatory determination cycle]. Yet OFGEM still recognise the need to normalise the data prior to modelling. Regional and company specific factor adjustments recognise that particular locations and particular networks incur costs beyond the control of the operating business and these costs should not be included in efficiency models.
449 Frontier also comments favourably on Ofgem’s approach which results in its final allowances being comprised of 25 percent of the DNOs’ submitted costs and 75 percent of its benchmarking models and notes in the Frontier Report (at p 96) that this is despite the fact that:
• Ofgem uses a ‘toolkit’ of approaches to determine its benchmarking target, including top-down econometric models, bottom-up unit cost analysis, bottom-up engineering assessments, assessments of historic costs and assessments of forecast costs, in order to provide the scope to cross check and sense check the efficiency estimates derived by any single approach.
• The quality of data available to Ofgem is significantly better than the data available to the AER, owing to the prodigious effort that has been invested in improving the underlying data, in particular the cost data.
• There has been a significant amount of engagement with the …[DNOs] … to develop the Ofgem models in the first place, allowing them to comment on Ofgem’s technique, cost driver choice, the quality of their own and other’s data, cost drivers that are not adequately captured by the models, differences in business model that may be picked up as inefficiency and any circumstances otherwise unique to the company that should be adjusted for or at least understood when interpreting the results.
450 Noting that that Ofgem has undertaken a decade or more of development work in respect of its data collection, Frontier also observes at p 103, that the AER should anticipate the need to undertake a similar programme of work and that:
We recognise that the AER has gone through a process to develop RIN templates but, set against Ofgem’s experience, it would be naïve for the AER to think that the RIN data obtained to date is sufficiently free from errors and inconsistencies as to warrant the degree of confidence the AER has placed in its modelling.
451 Based on its review of the Australian data and its experience of applying benchmarking techniques across Europe, Frontier further observes that the AER is regulating a sector with an unprecedented degree of heterogeneity. It notes in that regard that the one of the largest DNSPs, Essential, serves an area significantly greater than France and another, Ergon, an area significantly greater than France, the UK and Spain combined. It is Frontier’s opinion at p 104 that:
These statistics alone ought give the AER pause to consider whether it is sensible to treat networks of such scale the same as networks that serve much smaller geographies. Yet, the AER appears to have given no particular consideration to the unique circumstances faced by these networks. Instead, the AER has relied on very crude modelling tools to capture the effects of extreme scale, rurality, and sparsity. As a result, the AER’s modelling identifies these two networks as among the least efficient DNSPs in Australia. This is very surprising to us because European regulators, such as Ofgem, engage closely with networks with much less extreme characteristics than Essential Energy and Ergon Energy to understand any important factors that their modelling may have failed to capture.
452 Again drawing on its experience of practice in Europe, Frontier observes that:
… it is common for regulators to seek to triangulate “top down” benchmarking, of the kind produced by EI, with other sources of information, e.g. review by expert engineering consultants of unit costs, volumes of work, policies and practices in order to gain a more holistic view of network performance.
453 While ActewAGL accepts that r 6.5.6(e)(4) requires the AER to have regard to the benchmark opex that would be incurred by an efficient DNSP, it submits that the NER do not require the AER to give benchmarked opex any particular weight or mandate that the AER must give benchmarking a weight that is disproportionate to its probative value. In ActewAGL’s submission, the AER’s benchmarking methodology is incapable of providing much, if any, guidance whether ActewAGL’s forecast opex reasonably reflects the opex criteria and it is unreasonable to use it as the principal basis for the AER’s decisions.
454 It is also ActewAGL’s submission that:
(a) although its Final Decision contains many lengthy descriptions and diagrammatical representations of the decision making processes adopted by the AER (eg Step 1 in Table 7.4 Arriving at our alternative estimate of base opex, Attachment 7 to the ActewAGL Final Decision at p 7-26), the sole basis on which the AER estimated ActewAGL’s opex was the EI model and the post modelling OEF adjustments; and
(b) because that estimate was lower than ActewAGL’s forecast, and the AER was not satisfied by ActewAGL’s explanation of the difference, the AER adopted its own estimate based on the EI model.
455 That the AER’s decision making process is as submitted by ActewAGL may be seen by reference to p 16 of the Overview to the ActewAGL Final Decision where it is stated:
In this final decision we used our preferred benchmarking model [the EI model] as the starting point to arrive at an alternative estimate of opex that reasonably reflects an efficient base level.
456 This confirmed by reference to pp 15-16 of the Overview to the Ausgrid’s Final Decision:
In its revised proposal, Ausgrid based its opex forecast on its historical costs. … we are not satisfied that those forecasts are the appropriate starting point for forecasting its opex for 2015–19.
Instead, we have used our benchmarking analysis as the starting point for assessing Ausgrid's base level of opex. We are satisfied that our resulting opex forecast reasonably reflects the opex criteria.
… … …
In this final decision we used our preferred benchmarking model [the EI model] as the starting point to arrive at an alternative estimate of opex that reasonably reflects an efficient base level.
457 Thus, ActewAGL rightly submits that considerable caution ought be exercised about the manner in which the Final Decisions describe the AER’s decision-making processes, in particular, when claiming that the DNSPs’ opex as the starting point for its estimate of the required opex pursuant to rule 6.12.1(4)(ii). It is noted that ActewAGL’s submission goes further by saying that the AER’s statement that it “started” with ActewAGL’s forecast is “window dressing” and “meaningless” when viewed in light of the above quoted passages from the Overviews to the Final Decisions.
458 Attachment 7 to each of the Final Decisions in issue makes it plain that the AER arrived at its own estimate of opex based on the EI model that was lower than the DNSP’s and, as it was not satisfied that there was an explanation for the difference, rejected the DNSP’s forecast and deemed its own as the appropriate estimate. The DNSP’s forecast did not otherwise play a role in the AER's decision-making process, whether as a “starting point” or otherwise.
459 In response to the submissions to the effect that it is in error in not using the DNSPs’ forecasts as a starting point, the AER submits that:
(a) the NER do not stipulate that it must undertake a bottom-up engineering approach;
(b) it exposed the DNSPs’ forecasts to a multitude of assessment techniques using the information contained in the DNSPs’ regulatory proposals, eg
(i) category analysis disaggregated the costs in the opex forecasts and compared those costs to the DNSPs’ peers; and
(ii) the reviews undertaken by the AER’s consultants, Deloitte and EMCa relied upon the information contained in the regulatory proposals.
460 Responding to the applicants’ submissions that it applied the EI model in a deterministic manner, the AER submits that:
(a) it considered a range of analytical methods (not limited to the EI model) before concluding that it did not accept the DNSPs’ opex forecasts; and
(b) once it decided not to accept the DNSPs’ opex forecasts, its task was to make its own estimate of forecast opex;
(c) it checked the results of the EI model against two other econometric models and the index based opex MPFP technique (noting that the MPFP last technique does not use overseas data and uses a different output specification); and
(d) all of the techniques produced similar results for each DNSP.
461 Viewed in light of the following acute observation by Networks NSW’s expert, Frontier, in the Frontier Report at p 105, the AER’s submissions are tenuous:
… the AER appears to have put undue faith in the ability of it, and its advisers, to develop a single benchmarking model (or suite of very closely related models, all derived from the same data and missing the same wider review of factors and sense checks) that can capture very well relative inefficiency.
462 Having regard to the conclusions that may be drawn from the above considerations of the applicants’ submissions challenging the AER’s approach to determining the DNSPs’ opex, it is not necessary consider the following challenges to the AER’s benchmarking methodology, other than to note them:
(a) the AER’s failure to corroborate the results of the EI model;
(b) the AER’s failure to have proper regard to the DNSPs’ endogenous circumstances; and
(c) whether the AER had proper regard to the consequences of its estimates of the DNSPs’ opex.
463 Conceptually, the parties submissions address the principal issue (whether the AER’s application of the EI model discharged its obligations under rr 6.5.6 and 6.12.1(4)) at two levels. The first, involving the effect of the 2012 Rule Amendments, particularly the changes to r 6.5.6 and other rules relevant to its interpretation and application. The second, contingent on a DNSP establishing that the AER’s application of the EI model failed to discharge its obligations, involving the effect of the 2013 Legislative Amendments, particularly, the introduction of s 71P(2a) and (2b).
464 While the DNSPs’ submissions in support of their opex forecasts tend to an interpretation and application of the 2012 Rule Amendments favouring the RPP, the AER and PIAC’s submissions in support of lower opex allowances tend to an interpretation and application of the 2012 Rule Amendments favouring the NEO.
465 Insofar as there is such a tendency in a party’s submission it is rejected. The 2012 Rule Amendments simply do not contemplate that the NEO and the RPP are at cross purposes, or that their meaning has changed. They do not lead to a fresh policy subsidy to consumers by way of an artificially low opex figure or a bonus to a DNSP by way of an artificially high figure. Indeed, the AER (and the Tribunal on review) has a delicate task. Both must be conscious of the interests of consumers and the AER is bound to carefully scrutinise the information provided to it in support of a DNSP’s opex allowance. It must also have regard to the legitimate business interests of a DNSP and should not put itself in an adversarial position in relation to the DNSP so that it may be perceived as a champion of consumers – cf: Re East Australian Pipeline Limited  ACompT 8 at  and .
466 The 2012 Rule Amendments together with the 2013 Legislative Amendments give rise to a multifaceted regulatory regime calling for a balance between the interests of consumers on the one hand and the interests of DNSPs on the other. The observations of the High Court (per Gleeson CJ, Heydon and Crennan JJ) in East Australian Pipeline at  in relation to an earlier gas access regulatory regime are most apt in the Tribunal’s consideration of the regime now before it:
Stripped to essentials, such a regime is at least intended to allow efficient costs recovery to a service provider and at the same time ensure pricing arrangements for the consuming public which reflect the benefits of competition, despite the provision of such services by monopolies. The balancing of those objectives properly has a natural flow-on effect for future investment in infrastructure in Australia.
467 As noted above, there are a number of issues with the EI model and the AER’s application of it:
Inadequacies in the EI model’s data set and comparability issues
o the RIN data;
o the overseas data;
o the country dummy variables;
the lowering of the EI model’s comparison point
the OEF adjustments;
in circumstances where economic benchmarking is in its infancy in Australia, the reliance on qualitative analysis rather than bottom-up quantitative assessment to test issues such as those raised by the DNSPS re their vegetation management opex; and
the AER’s use of the EI model as the sole or principal determinative of opex.
468 As a first step in its consideration, the AER was required to decide whether it was satisfied that the total of the forecast opex in the Revised Regulatory Proposals of each of the DNSPs reasonably reflected each of the operating expenditure criteria set out in r 6.5.6(c). The AER’s analysis of the Networks NSW and ActewAGL Revised Regulatory Proposals led to it expressing concerns about a number of components or elements of those proposals. The Tribunal is not persuaded, having regard to those concerns, that the AER’s lack of satisfaction on that question exposes a ground of review. There was material upon which it could have reached that conclusion. There is no demonstrated ground of review made out in that step, even though (not surprisingly) there is considerable debate in the submissions about a number of the matters considered by the AER.
469 Consequently, the Tribunal does not consider that the step taken by the AER under r 6.5.6(d) involved error on its part so as to enliven any grounds of review under s 71C of the NEL.
470 Rule 16.12.1(4)(ii) then obliges the AER, on making its Final Decisions in relation to each of the DNSPs, to include with its reasons for the lack of satisfaction under r 6.5.6(d) an estimate of the forecast opex for each of the DNSPs for the 2015-19 regulatory control period that the AER:
… is satisfied reasonably reflects the operating expenditure criteria, taking into account the operating expenditure factors (see rule 6.5.6(e)).
471 As is apparent from the above, there are a number of respects in which, or reasons for, the Tribunal on these applications being of the view that one or more of the grounds of review under s 71C(1) are made out. At a general level, that is because the AER placed too much weight on the outcome of the EI model. That, in the Tribunal’s view represents an exercise of the AER’s discretion about the use to which the EI model should have been put which was incorrect.
472 Underlying that view are a series of concerns about the inputs to the EI model, and the OEF adjustments (including those of concern to PIAC), and including the AER’s treatment of the vegetation management costs of Essential, Endeavour and ActewAGL, and further including the AER’s treatment of the labour costs of the Networks NSW DNSPs. Those concerns can generally be described as errors of fact by the AER in its findings of fact, as discussed in detail above. Those errors do not simply reflect the AER’s choice of competing expert views. There are underlying elements to the EI model which mean that the AER at this point (accepting that the available Australian data is not sufficiently extensive for appropriate modelling) should not have placed the weight it did on the output of the EI model. As the earlier Introduction to these reasons discuss, there may be room for debate about whether a particular step shows an error of fact in a finding of fact, or is an incorrect exercise of a discretion. It would be possible, in a number of the specific instances (in particular in relation to the OEFs) to use either description by the use of different semantics. The line between the two is often hard to draw. The Tribunal, having regard to its conclusion in the preceding paragraph, does not think it is helpful to embark on that exercise.
473 It is desirable to add some further comments on this topic. On the one hand, the AER and PIAC perceive the 2012 Rule Amendments as shifting the emphasis in r 6.5.6 away from the individual actual circumstances of the DNSP whose opex forecast is subject to assessment to that of each DNSP’s opex forecast being assessed against a benchmark entity.
474 On the other hand, the DNSPs emphasise that:
(a) each of the r 6.5.6 opex objectives, the opex criteria and the opex factors; and
(b) in particular, each of the r 6.5.6(c) three opex criteria,
should not be conflated to arrive at a one size fits all benchmark assessment of a DNSP’s opex.
475 Also, the DNSPs perceive that the AER’s focus on the new r 6.5.6(e)(4) (benchmark opex that would be incurred by an efficient DNSP) lead it to ignore, or give insufficient weight to, other opex factors, in particular:
(a) the actual and expected opex of the DNSP during any preceding regulatory control periods (r 6.5.6(e)(5));
(b) the substitution possibilities between opex and capex (r 6.5.6(e)(7)); and
(c) whether the DNSP’s opex forecast is consistent with the r 6.5.8 EBSS or the r 6.6.2 STPIS (r 6.5.6(e)(8));
476 Furthermore, to the extent that the DNSPs acknowledge that the 2012 Rule Amendments gave some greater standing to benchmarking, the DNSPs are critical of the AER’s benchmarking methodology inputs and outcomes.
477 The parties seek to advance their respective perceptions of the impact of the 2012 Rule Amendments by reference to extrinsic material within the ambit of Schedule 2 of the NEL, namely, the AEMC’s Final Position Paper. Reliance was also placed on a review related material Productivity Commission’s report: Productivity Commission, Electricity network regulatory frameworks, inquiry report, 9 April 2013.
478 As often is the outcome when one a party or another goes from the words of a section or a rule to the words of another document to seek to bolster their interpretation of the section or rule, apt words of comfort for the position of either party may be found in the extrinsic and review related material.
479 However, having regard to the Tribunal’s conclusion (as to which see below) that the AER was, (because of inherent weaknesses in the EI model and the ex post adjustments to its outcomes) wrong to rely on the EI model to estimate the DNSPs’ required opex, it is not necessary to explore and rule on:
(a) the way that the parties advance the extrinsic and the review related material in support of their interpretation of r 6.5.6; or
(b) the minutiae of the parties’ lengthy submissions on how r 6.5.6 should be interpreted.
480 Suffice to say at this point that in a context where it is applying benchmarking for the first time, the AER’s application of the EI model gave a discordant weight to r 6.5.6(e)(4) (benchmark opex that would be incurred by an efficient DNSP) vis á vis the other r 6.5.6(e) opex factors.
481 It is nevertheless appropriate to record that, as the Tribunal observed above, it does not consider that the 2013 Legislative Amendments changed the meaning of the NEO or the RPP, or their relationship. Clearly, the introduction of a “materially preferable NEO decision” and “preferable reviewable regulatory decision” and ss 16(1)(d) and 71P(2a) and (2b) in the NEL (and the complementary changes in the NGL) refined the focus of the respective decisions of the AER and the Tribunal, but if there were an intention to change ss 7 and 7A of the NEL (or ss 23 and 24 of the NGL), that would have been clearly expressed. The same general comment may be made about the 2012 Rule Amendments themselves, at least in relation to r 6.5.6 of the NER. The changes clearly refine the focus of both the AER and on review the Tribunal. But the concepts of “operating expenditure criteria” and “operating expenditure factors” pre-existed the 2012 Rule Amendments. Apart from the particular changes, some of which are clearly to accommodate the new and more sophisticated processes expected of the AER (eg in r 6.8 of the NER), and noting the amendments made to r 6.5.6(c) and (e), there is no reason to think that one or other of the opex criteria or the opex factors is intended to have a pre-eminence over that of the others.
Section 71P(2a) and (2b) of the NEL.
482 The Tribunal, having been satisfied that there are grounds of review made out in relation to the AER’s opex allowance for each of the Networks NSW DNSPs, and for ActewAGL, has the power under s 71P(2) available to it. They are expressly subject to s 71P(2a) as explained in s 71P(2b). In the case of the Networks NSW DNSPs, the appropriateness of granting particular relief or indeed of granting any relief at all is a more complex one because of the findings about PIAC’s application and in particular how the OEFs were addressed.
483 It is premature to deal with those questions at this point, because of the need to consider the inter-relationship of the constituent elements of the relevant Final Decisions. Where there are other elements of the four Final Decisions which are also the subject of challenge, it is appropriate at first to address those issues.
484 However, as is almost self-evident, if the Tribunal were to be satisfied in terms of s 71P(2a)(c), the Tribunal would not decide to vary the Final Decisions under s 71P(2)(b) because it is not satisfied that to do so would not require the Tribunal to undertake an assessment of such complexity that the preferable course is to set aside the Final Decisions, and remit the matters to the AER, having regard to these reasons. In short, the Tribunal is left with the options of affirming or setting aside the four Final Decisions and remitting them to the AER under s 71(2)(c).
485 That really follows from a reading of the submissions addressing the principal issue under this heading that for every competing argument there is a supporting expert or experts and given that context, the use of the phrase “materially preferable” requires the Tribunal to look through the inevitable conflict and difference of views between experts, all advocating positions which they regard as being preferable, and to determine whether an advocated materially preferable NEO decision is, indeed, materially preferable: ie a decision which, notwithstanding that divergence of views, is sufficiently compelling to be seen by the Tribunal as being “materially preferable” than that advocated by the AER: cf: Wellington International Airport Limited & Ors v Commerce Commission  NZHC 3289 at .
486 One further matter should be mentioned. It was raised by each of Networks NSW, ActewAGL and Ergon.
487 They contended that, if the AER was correct in its conclusion that the distributors’ opex forecasts did not reflect efficient costs, the AER’s decision involved an incorrect exercise of discretion, or an unreasonable decision, in failing to provide a transition path for the DNSPs to reduce their opex to efficient levels (as decided by the AER) or any allowance for the costs involved in transitioning. Ergon also alleges that the AER made an error of fact in finding that a transition path was not required under r 6.5.5(c)(3).
488 Similar grounds are also raised by Networks NSW, ActewAGL and Ergon in relation to the X factor. The Tribunal’s reasons in relation to those grounds are outlined in the X factor section of these reasons.
489 It is submitted that, if there is a step decrease in opex between two regulatory periods, the AER should provide an allowance over and above the AER’s opex allowance, because of the time the particular DNSP would have to take, and the costs that it would have to incur, to transition its business to one that can operate at the AER’s proposed opex allowance.
490 The justification for such an allowance was summarised by Networks NSW as necessary for the following reasons. First, paying redundancy expenses constitutes the efficient, prudent and realistic costs of compliance with a “regulatory obligation or requirement” within the meaning of r 6.5.6(a)(2) of the NER – namely the provisions for redundancy pay contained in the EBAs which bind the Networks NSW businesses. The reasons why Networks NSW says compliance with EBA provisions constitutes a “regulatory obligation or requirement” are set out in relation to opex above.
491 Secondly, the AER’s view that the Networks NSW businesses should bear redundancy costs is predicated to a substantial extent on the proposition that the Networks NSW businesses acted inefficiently or imprudently in hiring a permanent workforce on EBAs in order to meet increased licence conditions. It is said that this is not correct and that, if those costs were prudent and efficient, the costs of reducing that workforce now that the need for labour is reduced should similarly be seen as prudent and efficient.
492 Thirdly, it is submitted that an immediate transition to a materially lower level of opex is neither prudent nor realistic, and that a smoother transition path with respect to any required opex reductions is in the long term interests of consumers. That is because, amongst other things, it would provide the relevant DNSP with a reasonable opportunity to recover at least their efficient costs and would provide sufficient incentive for an entity to invest in a manner that will best achieve the NEO: s 7A(2) of the NEL.
493 Ergon also notes that these considerations are equally important in the context of regulatory decisions which the AER is required to make in relation to capex, as, similarly capex forecasts must reasonably reflect a DNSP’s efficient costs, the costs that would be incurred by a prudent service provider and a realistic expectation of the DNSP’s demand forecasts and cost inputs.
494 Due to the Tribunal’s findings on opex, the Tribunal does not, in the circumstances, need to determine whether these contentions by Networks NSW, ActewAGL and Ergon are correct. When the AER revisits and redetermines the opex allowance, it will have to consider the costs involved in transitioning. It will do so at a time, and in relation to revenue streams, which will require it to make a fresh decision. The Tribunal is anxious not to inhibit the AER at this point in exercising its discretion in that regard.
495 Having regard to the DNSPs and PIAC’s submissions as a whole the Tribunal concludes that the AER’S reliance on the EI model failed to discharge its obligations under rr 6.5.6 and 6.12.1(4).
496 In reaching that conclusion, the Tribunal has the following matters in mind.
(a) The AER’s undue reliance on the EI model as a determinative factor in the AER’s estimation of each DNSP’s required opex pursuant to r 6.12.1(4)(ii). That reliance being placed on the model notwithstanding that it recognised it had limitations with respect to the specification of outputs and inputs, data imperfections, and other uncertainties in a context where economic benchmarking is being used for the first time to set opex allowances – see eg: Attachment 7 to the Endeavour Final Decision, at pp 7-268 and 7-269, Attachment 7 to the Ausgrid Final Decision, at p 7-64 and Attachment 7 to the ActewAGL Final decision at p 7-250.
(b) The restricted opportunity afforded to the parties (and denied to third parties as the AER’s obligation to consult was past) to test the veracity of the EI model. That is not to cast an adverse reflection on the AER. Nor is it to suggest that the AER did not conscientiously examine submissions it received after its draft decisions. It is simply to recognise that the AER has a large and most difficult task to perform within a limited timeframe – a timeframe that did not permit it to conduct the consultation required to:
(i) adequately test the data in the EI model and the other models to which it had regard; and
(ii) expose the DNSPs’ consultants reports to the rigours of examination that the AER’s consultation obligations are designed to foster.
497 The Tribunal will address later in these reasons, under the heading “The Tribunal’s Determination” the application of s 71P(2a) and (2b).
498 This issue is confined to the Networks NSW DNSPs. For reasons which will shortly be apparent, it is an element of the Final Decisions of the AER in relation to the Networks NSW DNSPs which will need to be revisited by the AER. In short, it will follow from any adjustment to be made to the opex allowances of the Networks NSW DNSPs that the X-factor will have to be re-applied.
499 It is not a matter upon which PIAC made submissions, either as an applicant in relation to those Final Decisions or as an intervener. Nor did any other intervener address submissions in relation to it.
500 It is nevertheless appropriate for the Tribunal to address it as a separate issue, because Networks NSW assert that grounds of review may be made out in relation to its application in any event.
501 One of the constituent decisions of the relevant Final Decisions concerns the control mechanism for standard control services (r 6.12.1(11)). One part for these mechanisms is the “X factor”. The X factor determined for a particular year represents the real rate of change in revenues for that year that have been approved by the AER (before any annual adjustments). In effect, it operates as a “smoothing factor” for revenue over consecutive years.
502 The X factor to be used in the control mechanism is to be determined by reference to the requirements set out in r 6.5.9(b) of the NER. Rule 6.5.9(b) describes the X factor relevantly as follows:
(a) must be set by the AER with regard to the DNSP’s total revenue requirement for the regulatory control period (6.5.9(b)(1));
(b) must be such as to minimise, as far as reasonably possible, variance between expected revenue for the last regulatory year of that regulatory control period and the annual revenue requirement for that last regulatory year (6.5.9(b)(2)); and
(c) must conform with (relevantly for standard control services) the X factor must be designed to equalise (in terms of net present value) the revenue to be earned by the DNSP from the provision of standard control services over the regulatory control period with the provider’s total revenue requirement for the regulatory control period (6.5.9(b)(3)(i)).
503 The X factor is generally calculated so as to allow for smoothing of revenues subject to the requirement that both smoothed and unsmoothed revenues are equal in net present value (NPV) terms over a 5 year regulatory control period. Due to the transitional rules the regulatory years 2014-19 were split over two regulatory control periods, with a “transitional regulatory control period” for 2014-15 and a “subsequent regulatory control period: for 2015-19: see r 11.55.1 of the NER. For the transitional regulatory control period, the AER determined a “placeholder revenue” separately from the annual revenue requirement for 2014-15 established through the determination process.
504 In relation to each of the Networks NSW DNSPs, the “placeholder Revenue” amount was higher than the annual revenue requirement which was finally determined.
505 In effect, after making a determination about the annual revenue requirement for each year in the regulatory period and the total revenue requirement for the regulatory control period, the AER is required to adjust and “smooth” the revenue in accordance with the NER.
506 The dispute in relation to the control mechanism for standard control services concerns the AER’s relative allocation of revenue for each year of the regulatory period 2014-19 and how the difference between the “placeholder revenue” amount and the annual revenue requirement, and the differences between annual revenue for each year, should be minimised or “smoothed”.
507 The following description is taken from the AER’s submissions, and to a degree incorporates its explanations for the steps it took.
508 In 2014, the AER determined placeholder revenue allowances for the transitional regulatory control period. In the Draft and the Final Decisions for the years 2015-19, the AER made a full regulatory determination for each year, and accounted for any adjustment amount related to the transitional regulatory control period. As part of this process, it was required to determine annual revenue requirements for each year of the five year period (2014-19) and use a NPV neutral true-up mechanism to account for any difference between:
(a) the placeholder revenue for the transitional regulatory control period; and
(b) the annual revenue requirement for 2014-15 established through the full determination process.
509 To give effect to the true-up, the AER set each of the Networks NSW DNSP’s first year expected revenue in the post-tax revenue model equal to the AER approved placeholder revenue for 2014-15: see eg Ausgrid Final Decision, Attachment 1 at p 1-14. The AER considered that this was the only practical option, as distribution and transmission prices were set for 2014-15 based on these approved placeholder amounts. This meant that the difference in revenues for 2014-15 between the transitional and Final Decisions needed to be accounted for in the 2015-19 regulatory control period. That is, the placeholder revenue for 2014-15 from the transitional determination provided a base from which the expected revenues (smoothed) for the remaining four years of the 2014-19 period were calculated, giving effect to the true-up and returning the difference to customers over the 2015-19 period.
510 In determining the X factors for the remaining four years of the regulatory control period, the AER was constrained in smoothing by the transitional year X factor: that factor was locked in as it had been used to determine 2014-15 prices that were approved. Further, as the AER determined (in its final decision) that the actual revenue requirement for 2014-15 was lower than that approved for the transitional determination, the NER’s transitional requirement for a true-up in relation to 2014-15 revenues: see r 11.56.4(h)-(i), meant that there were revenues received in 2014-15 that had to be returned to customers and therefore reflected in future years’ X factors.
511 By reason of these circumstances, it was difficult to apply the AER’s usual smoothing approach, which aimed not only to smooth within the regulatory control period, but to minimise any step change in revenues from the end of the regulatory control period (2018-19) to the start of the next regulatory control period (2019-20). The NER’s transitional requirements removed the usual requirement to avoid such step changes, but the AER considered that as a matter of policy and consistency with the NEO it should still avoid too big a potential step change in revenues across regulatory control periods. Accordingly, the AER widened its usual tolerance limit of a +3 percent step change to as much as 10 percent for each of the Networks NSW DNSPs (Ausgrid: 10 percent for distribution and transmission; Endeavour: 10 percent; Essential: 10 percent): see Ausgrid Final Decision, Attachment 1 at p 1-15; Essential Final Decision, Attachment 1 at p 1-12; and Endeavour Final Decision, Attachment 1 at p 1-11.
512 Within these constraints (ensuring NPV neutrality, dealing with transitional year issues, and avoiding large revenue step changes at the end of the period) the AER smoothed revenues as much as it could by determining X factors that would not result in revenue falling and then rising again in subsequent years. The best profile that was achievable within the constraints required a significant revenue reduction in 2015-16 (as determined in the AER Final Decisions) which would then allow revenues to remain relatively flat for the rest of the period.
513 Between the Draft Decisions and Final Decisions, the AER made modest adjustments to the way smoothing occurred in response to concerns from the distributors, the effect of which was to allow the difference in end of period unsmoothed/smoothed revenues to be increased up to 10 percent. Although this facilitated a slightly smaller X factor for 2015-16 by increasing the 2016-17 X factor, the AER could not shift revenue reductions further into the future. A reduction to revenue in 2017-18, would require an increase in 2018-19 or, alternatively, a step change in revenue larger than 10 percent in 2019-20.
514 Hence, in the placeholder determination for the transitional regulatory control period applying to Ausgrid, Essential and Endeavour, the AER determined the annual revenue requirement for the Networks NSW DNSPs for the 2014-15 year as follows:
(a) Ausgrid (distribution): $1,956.45m;
(b) Ausgrid (transmission): $252.31m;
(c) Endeavour: $949.45m; and
(d) Essential: $1,291.72m.
515 However, in the Final Decisions in respect of the 2014-15 period, the AER determined the notional annual revenue requirements for the 2014-15 regulatory control year to be:
(a) Ausgrid (distribution): $1,546.00m;
(b) Ausgrid (transmission): $192.76m;
(c) Endeavour: $858.58m; and
(d) Essential: $976.15m.
516 As the AER says, because r 11.56.4(h) of the NER requires the AER, in the Final Decisions, to adjust the total revenue requirement for the subsequent regulatory control period to account for the difference between the placeholder annual revenue requirement determined for the transitional regulatory control period and the annual revenue requirement for the transitional regulatory control period as determined in the Final Decision, there were revenues received in 2014-15 that had to be returned to customers during 2015-19 and therefore reflected in future years X factors.
517 The result was that the AER determined the X factor for the 2015-16 regulatory year in a manner that gave immediate effect to a substantial amount of the difference between the annual revenue requirement approved in the AER’s placeholder determination for the transitional regulatory control period (the 2014-15 year) and the annual revenue requirement determined for the 2015-16 regulatory year. The effect of the X factor decision is to substantially reduce revenues in the 2015-16 regulatory year, followed by modest reductions (or in the case of Essential modest increases) year on year.
518 In order to offset the over recovery of revenue in 2014-15, in all of the 2015-19 regulatory years, the annual expected revenues are below the AER’s determined annual revenue requirement (except for Endeavour for the 2015-16 year, where the smoothed expected revenue is above the annual revenue requirement).
519 The expected annual revenue for the final year of the regulatory control period (2018-19) conforms with the AER’s +10 percent tolerance for revenue changes between regulatory control periods and equates to a 13.5 percent nominal increase in price in the first year of the next regulatory control period.
The Grounds of Review
520 Networks NSW says that the X factor decisions involve an incorrect exercise of discretion and/or an unreasonable decision: s 71C(1) of the NEL. They say that the magnitude of the reduction in revenue between 2014-15 and 2015-16:
(a) does not promote efficient investment in, and operation of electricity services for the long term interests of consumers with respect to safety quality and reliability and is directly contrary to incentive regulation contrary to the NEO;
(b) is inconsistent with s 7A(2) of the revenue and pricing principles as the decision does not provide Networks NSW with a reasonable opportunity to recover at least its (AER’s determined) efficient costs;
(c) inconsistent with s 7A(3) of the RPP as it does not promote efficient investment because it requires Networks NSW DNSPs to incur significant debt and equity investment to continue to operate given its actual costs; and
(d) gives rise to price shocks and pricing volatility which could be ameliorated if a more graduated reduction in revenue was implemented.
It is fair to observe that (a) and (b) are really formulaic.
521 Networks NSW also says that the AER’s Final Decisions place the Networks NSW businesses under immediate financial strain, and amplify the other errors by the AER which relate to the reduction of revenue and place the businesses at financial risk.
522 There is little debate about the immediate effect of the X factor decisions of the AER.
523 They give rise to significant price decreases in 2015-16, potentially followed by nominal price increases from 2016-19. Networks NSW says this leads to pricing volatility which is not in the long term interests of consumers with respect to price and not in accordance with the NEO. They also give rise to annual expected revenues in all regulatory years of the 2015-19 regulatory control period below the AER’s determined annual revenue requirement for each of those years. Again, Networks NSW says that is not in the long term interests of consumers of electricity with respect to safety, quality, reliability and security of supply of electricity. They say that a more graduated reduction in annual expected revenues that permitted recovery of revenues closer to the annual revenue requirements in the earlier years of the regulatory control period would assist in providing the Networks NSW DNSPs with an opportunity to adjust to significant revenue reductions, and to improve tariff efficiency and equity without imposing unacceptable price shocks. It is argued that the decisions to impose a single real reduction in revenue requirements for 2015-16 limits their ability to deliver the long term benefit of tariff reform to customers because it is not possible to develop and implement a full tariff reform agenda for 1 July 2015, particularly given the need to engage with customers and other stakeholders prior to any decision on tariff reform.
524 It is evident that there are significant qualitative assessments underlying the respective positions of Networks NSW and the AER.
525 As the Tables in the Networks NSW submission indicate, because of the significant over-recovery of revenue in the transitional year, the smoothing imposed by the AER will impact largely in the 2015-16 year by ($449m) distribution and ($197m) transmission in the case of Ausgrid, being in excess of 24 percent and 27 percent of the unsmoothed revenue requirement for that year, and then ranging between 6 percent and 3 percent for distribution, and at 2 percent transmission for the following years.
526 In the case of Endeavour, the reduction is over 17 percent on the 2015-16 year, and then 3 percent for each of the following years. In the case of Essential, the reduction is over 31 percent in the 2015-16 year and then insignificant (a minor increment) in the following years.
527 In each instance, obviously there is a significantly reduced cash flow in 2015-16, following the revenue allowed in 2014-15 which on the other hand was excessive because of the assessed cost levels in the current regulatory period. Such a dramatic change, not smoothed over a period of years, is said to contravene the NEO by not promoting efficient investment in, and provision of, electricity services in the long term interests of consumers. It is also said to be directly contrary to the incentive regulation structure under the NEL (as discussed in the Introduction section of these reasons).
528 As noted, they are qualitative assertions, readily understood. The AER, for its part, says it was constrained to that “smoothing” decision by the effectively shortened regulatory period; by the “true-up” requirement for the 2014-15 revenues (which was supported by Networks NSW): see Consultation Paper on savings and transitional arrangements draft national electricity amendment (economic regulation of network service providers) rule 2012, 25 October 2014 at pp 4-5; by achieving net present value neutrality between smoothed and unsmoothed revenues; and by avoiding large revenue step changes at the end of the regulatory control period.
529 The proposition that the AER’s decision would not best contribute to the achievement of the NEO is a complex one. The AER selected smoothed later years as best contributing to the achievement of the NEO, including (as it considered) the avoidance of significant step changes in revenues at the end of the regulatory control period.
530 The Tribunal does not need to resolve that dispute because “the AER accepts that, in the event that the Tribunal finds reviewable error in relation to any of the AER’s other constituent decisions affecting revenue, the X factor will have to be reapplied by the AER on remittal”. As appears later in these reasons, the Tribunal does intend to remit the Final Decisions concerning the DNSPs for reconsideration by the AER.
531 It should be noted that the parties agree on the desirability of avoiding price shocks and volatility. When raised during the consumer consultation period, the strong message from consumers, both personally and through representative bodies, was that price shocks should be avoided where possible – eg Mr G Brody representing the Consumer Law Action Centre submitted:
… it’s better for consumers to have a smooth cost of the bill. If bills go up and down and create price shocks then that can cause as much problem for an individual household as, you know, overall high costs.
532 The AER says that if the 2015-16 revenue was to be higher than was set in the Final Decisions, the subsequent three years would need to be lower than was set in the Final Decisions and that this would lead to a final year difference greater than 10 percent, in circumstances where the costs beyond the 2015-19 regulatory period were uncertain. The AER says that this would lead to a price shock at the beginning of the next regulatory period (2019-20). The Tribunal notes that the AER had considerable flexibility to spread the reduction in revenue over a longer period (without disadvantaging consumers in present value terms) and that, if there is any option once it has re-worked the opex allowance, the Tribunal has not determined that it should necessarily default to making almost the whole of any reduction in the (next) first period.
533 That also means that the Tribunal does not have to form concluded views on Networks NSW’s submissions that the way the AER implemented the X factor imposed significant and inappropriate financial strains on each of them, even to the point of their respective financial viability being at risk.
534 There is one aspect of the AER’s submission on this topic which the Tribunal does, however, need to address.
535 The Networks NSW submission is that the X factor decision is inconsistent with s 7A(2) of the RPP because it would mean that each of them did not have a reasonable opportunity to recover at least the efficient costs of that operator. For the reasons given, the factual proposition does not need to be resolved.
536 The AER says, in addition, that it need not act consistently with the RPP at all times. It points out that s 16(2) provides:
In addition, the AER –
(a) must take into account the revenue and pricing principles –
(i) when exercising a discretion in making those parts of a distribution determination or transmission determination relating to direct control network services; or
(ii) when making an access determination relating to a rate or charge for an electricity network service; and
(b) may take into account the revenue and pricing principles when performing or exercising any other AER economic regulatory function or power, if the AER considers it appropriate to do so.
Hence, it says, it “may take into account” the RPP (relevantly s 7A(2)), but it is not bound to do so.
537 It may be that the difference in views between the Networks NSW and the AER is semantic. The Tribunal, of course, accepts that there are matters of judgment about how the RPP (or a particular element of one of the principles) should be taken into account. It does not accept that, as perhaps the AER is saying, the NEO in its application may give rise to a result which means that a DNSP is not given a reasonable opportunity to recover at least its efficient costs in providing the direct control network services. As the Tribunal has sought to express in its Introductory remarks, it does not regard ss 7 and 7A as other than complementary so that the NEO may give rise to a reviewable regulatory decision which in fact is inconsistent with the RPP or one of the elements of the RPP.
538 The Tribunal does not, in the circumstances, need to determine whether the basic assertion by Networks NSW is correct. When the AER revisits and re-determines the opex allowances, it will then have to apply the X factor. It will do so at a time, and in relation to revenue streams, which will require it to make a fresh decision on the X factor. The Tribunal is anxious not to inhibit the AER at this point in exercising its discretion in that regard.
539 In the earlier part of these reasons for decision, there is extensive reference to the NEO and to the RPP in ss 7 and 7A of the NEL respectively. Section 7A(3) provides that:
A regulated network service provider should be provided with effective incentives in order to promote economic efficiency with respect to direct control network services the operator provides. The economic efficiency that should be promoted includes –
(a) efficient investment in a distribution system or transmission system with which the operator provides direct control network services; and
(b) the efficient provision of electricity network services; and
(c) the efficient use of the distribution system or transmission system with which the operator provides direct control network services.
540 Specifically, s 7A(3) provides that the DNSPs should be provided with effective incentives in order to promote economic efficiency in the provision of their network services. As discussed above, and as contended for by Networks NSW (and ActewAGL), it is clear that the structure of the RPP under the NEL reflects the concept of “incentive regulation”.
541 As part of that incentive regulation, an EBBS makes provision for sharing between a DNSP and its customers the efficiency gains or losses derived from the difference between a DNSPs actual opex and the forecast opex allowance for a regulatory control period. The EBSS is a forward-looking scheme. A DNSP is told at the commencement of the regulatory period what to aim for and at the conclusion of the regulatory period, it is told how well it did in relation to the efficiencies. The EBSS incentivises a DNSP by allowing it to keep any yearly gain derived from the difference between its actual opex and its forecast opex, not just in that year, but until the conclusion of the regulatory period. Thus, as the AER’s Final Decision New South Wales distribution determination 2009-10 to 2013-14, 29 April 2009, observed (at p 245):
The scheme will not have a direct financial impact on the NSW DNSPs until the 2014–19 regulatory control period, when the DNSPs will receive carryover benefits/penalties for efficiency gains/losses made during the next regulatory control period.
542 The AER further explained the role of the EBSS in the following paragraphs commencing on p 1 of the AER’s Efficiency benefit sharing scheme for the ACT and NSW 2009 distribution determinations, 29 February 2008 (the 2008 EBSS), the relevant EBSS for the Tribunal’s review:
The purpose of the EBSS is to share efficiency gains and losses between DNSPs and distribution network users. In the absence of an EBSS, the share of efficiency gains and losses received by a DNSPs declines as the regulatory control period progresses and, consequently, the incentive for the DNSP to improve the efficiency of its operating expenditure (opex) declines also.
The EBSS allows a DNSP to retain the benefits of an efficiency gain for the length of the carryover period regardless of the year of the regulatory control period in which the gain was initiated. After the length of the carryover period the benefits of an efficiency gain are ‘shared’ with distribution network users. By doing so the EBSS provides a DNSP with a constant incentive to improve the efficiency of its opex and thus reveal their efficient level of opex.
543 Section B.2.1 of the written submission of Networks NSW refers to a range of extrinsic materials which confirm and explain incentive regulation. It is not necessary to refer to them in detail. For the immediate purpose of addressing the EBSS allowance, s 7A(3) is clear.
544 That reflects the requirement in r 6.3.2(a)(3) of the NER that the building block determination for a DNSP must specify, for a regulatory control period, amongst other things, how any applicable EBSS is to apply to the DNSP.
545 Rule 6.4.3 then explains the building block approach. Firstly, it specifies that the annual revenue requirement of a DNSP must include the approach under which the building blocks include the revenue increments or decrements (if any) for each regulatory year of the period arising from the application of any EBSS: see r 6.4.3(a)(5). Secondly, for that purpose, it cross refers to r 6.5.8: see r 6.4.3(b)(5).
546 Rule 6.5.8 then addresses in detail the EBSS. The AER emphasises that r 6.3.2(a) says that it is for the AER, in its relevant Final Decisions, to specify how the applicable EBSS is to apply, and that r 6.12.1(9) indicates that its relevant Final Decisions are predicated on its decision as to how any applicable EBSS is to apply to a DNSP.
547 It pithily asserts that those rules, including r 6.5.8(c), means that the EBSS must only reward “real efficiency gains”.
548 This topic also concerns only the three Networks NSW DNSPs. It was not the subject of submissions by PIAC, nor by any of the other interveners.
549 The Final Decisions of the AER in relation to each of the Networks NSW entities is reflected in dollar terms in the following table, comparing the Revised Regulatory Proposal of each of those entities and the Final Decision:
Revised Regulatory Proposal
$166m worse off
$103.6m worse off
$74.2m better off
The expressions “reward” and “worse off” or “better off” are those used by the AER. The position of Essential is partly addressed separately in the Tribunal’s reasons dealing with its application. As can be seen, the Final Decisions for both Ausgrid and Endeavour resulted in a significantly smaller allowance (or potential allowance) for EBSS than was claimed in their respective Revised Regulatory Proposals. In the case of Essential, the AER in the Final Decision in effect waived retrospectively the imposition of that penalty.
550 It is also convenient to recall that, an underlying theme of the Networks NSW submission is that the AER’s approach in its Final Decisions concerning those three entities was flawed because it did not allow appropriately for opex, as well as the EBSS, to be consistent with incentive regulation. That is, at a higher level of reasoning, merely a qualitative complaint. Moreover, by inviting the substitution of the Tribunal’s assessment of what is appropriate to achieve effective incentive regulation, that approach tends to seduce attention away from the NEO in s 7 and the RRP in s 7A, and from the manner in which the NER (as prescribed by the AEMC) provide for and describe the way in which the NEO is to be achieved. However, it must be borne in mind that it is necessary to pay close attention to the relevant provisions to address a particular complaint or complaints.
551 That is what the Tribunal has sought to do in relation to opex (above) and in relation to each of the elements of the Final Decisions about which Networks NSW, and ActewAGL and JGN complain.
552 In relation to the EBSS, Ausgrid and Endeavour say that the AER suspended the operation of the EBSS for the 2015-19 regulatory control period so that it no longer has a “functional role” in the long term regulatory structure of DNSPs as contemplated by r 6.5.8 of the NER. In its general submissions, they say:
However, as the AER observed, Ausgrid and Essential “will already bear any costs in transitioning to efficient levels … there does not seem to be a strong reason to provide it with additional incentive to become more efficient”. In other words, looking forward there is no role for incentives anymore because the AER is forcing (what it considers to be) an optimum efficiency, by the use of benchmarking, on Ausgrid and Essential, in the short term (indeed, immediately).
553 As noted, it is common ground that one of the building blocks for the annual revenue requirement of a DNSP is the revenue increment of decrement (if any) for a particular year arising from the application of any EBSS. So much is required by r 6.4.3(a)(5) of the NER. The relevant increment or decrement are prescribed or anticipated by r 6.4.3(b)(5).
554 In the development and implementation of the EBSS, r 6.5.8(c) requires the AER to have regard to:
(1) the need to ensure that benefits to electricity consumers likely to result from the scheme are sufficient to warrant any reward or penalty under the scheme for DNSPs;
(2) the need to provide DNSPs with a continuous incentive, so far as is consistent with economic efficiency, to reduce opex;
(3) the desirability of both rewarding DNSPs for efficiency gains and penalising DNSPs for efficiency losses;
(4) any incentive that DNSPs may have to capitalise expenditure; and
(5) the possible effects of the scheme on incentives for the implementation of non-network alternatives.
555 Thus, the EBSS is to create a continuous incentive for a DNSP to find efficiency gains by permitting the DNSP to retain the benefit of the gain for 5 years regardless of the year in which the gain is realised, with consumers having the benefit thereafter. The adverse is that the EBSS provides a disincentive for efficiency losses by providing that the DNSP is penalised for five years for any inappropriate increase in expenditure. These matters are achieved by providing for carryover gains or losses into the next period.
556 The 2008 EBSS describes the EBSS at section 2.1 as rewarding “sustained efficiency gains through the operation of a symmetrical carryover mechanism”. Hence, a DNSP is either rewarded for opex reductions against forecasts or penalised for opex that exceeds forecast expenditure.
557 The example of the EBSS in Appendix A to the 2008 EBSS shows that this approach would lead to a sharing ratio of 70 percent of the efficiency gain to be returned to consumers over a 15 year period, with 30 percent of the gain being retained by the DNSP (and the same pattern with efficiency losses).
558 Consequently, the 2008 EBSS provides for the calculation of carryover amounts (either gains or losses) to be applied as a building block element in the calculation of allowed revenue for the regulatory control period commencing on 1 July 2014. The 2008 EBSS contains a formula for calculation of the carryover amounts, where:
(a) the efficiency gain or loss for the first year (2009-10) is the forecast opex minus actual opex for that year;
(b) the efficiency gain or loss for each subsequent year is: (forecast opex minus actual opex for that year) – (forecast opex minus actual opex for the previous year).
559 The 2008 EBSS provided that a DNSP could propose a range of additional cost categories to be excluded from the EBSS. That was to ensure that efficiency gains would be measured as the difference between forecast and actual expenditure, subject to adjustments designed to remove the impacts of agreed uncontrollable costs, non-network alternative opex and recognised pass-through events, and changes in capitalisation policies, demand growth and regulatory responsibilities: see at p 12.
560 For the 2009-14 revenue determination, the AER excluded five specific cost categories of opex from the operation of the EBSS for the next regulatory control period on that basis. They were debt raising costs, self-insurance costs, insurance costs, superannuation costs relating to defined benefit and retirement schemes, and non-network alternative costs. The 2009-14 revenue determination also specified forecast total opex amounts for each year for the purposes of the EBSS (that is, total forecast opex minus the excluded costs).
561 It should also be noted that Networks NSW’s forecast, and approved, opex for that determination period included provisions, calculated on an accruals basis, which included provisions for employee benefits.
562 It is not suggested that the provisions so made did not accord with Australian Accounting Standard Board standard No 119, requiring an entity to recognise a liability when an employee has provided service in exchange for employee benefits to be paid in the future and on the other side of the ledger an expense when the entity receives that economic benefit. Hence, cash payments in relation to employee benefits such as long service leave payouts either in service or when an employee exits the employment, reduce the provisions as they are made.
563 Networks NSW says that, in accordance with the 2008 EBSS, they included in their revenue proposals a calculation of EBSS carry-over amounts flowing from the differences between forecast opex and actual opex in the 2009-14 period. However, a not insignificant proportion of the component of forecast and actual opex comprised movements in provisions for employee benefits and other miscellaneous matters which were excluded by the AER from its allowance for EBSS in the current regulatory period.
564 That is because, the AER says, the Networks NSW entities changed the assumptions used to calculate their estimates, and so their provisioning for future payments for existing liabilities, resulting in artificial efficiency gains or losses which were claimed as “real efficiency gains” and losses but which did not represent genuine business outcomes.
565 The AER points out that the 2008 EBSS included the following statements:
The measurement of gains and losses should not be affected by artificial means such as the shifting of costs between years, but should represent genuine business outcomes that have arisen in the ordinary course of conducting the business in a prudent and diligent manner.
Adjustments will be made where necessary to correct for variances in costs categories and methodologies, and errors.
In calculating carryover gains or losses, the AER must be satisfied that the actual and forecast opex accurately reflects the costs faced by the DNSP in the regulatory control period.
566 The question whether the AER, in proceeding on the basis set out above, acted without exposing a ground of review is at the heart of this issue.
567 By way of introduction to the AER’s decision, the AER says that it has a considerable degree of discretion in how it applies the EBSS. As noted, that discretion, it argues, is apparent inter alia from rr 6.3.2(3) and 6.12.1(9) which provide that its building block determination for a DNSP must specify how the EBSS is to apply to the DNSP, thus recognising its discretionary decision-making, and secondly by r 6.5.8(c) which provides that it must have regard to a number of factors when implementing the EBSS, including “the need to ensure that benefits to consumers likely to result in the scheme are sufficient to warrant any reward or penalty” for the DNSPs. It is not, therefore, simply a mechanical exercise of adopting, relevantly, the provisioning of the DNSP for its liabilities from time to time.
568 Consequently, the AER’s approach to the EBSS for the current regulatory period in relation to each of the Networks NSW entities is driven, at least to a significant extent, by its understanding that the claim for reward under the EBSS is unrelated to any real efficiency gains.
569 As noted above, the Networks NSW entities reduced their opex in the last regulatory period by changing their estimates for provisioning of future payments to employees of entitlements such as provision for long service leave. The AER considered that the change in estimates was not driven by real efficiency gains, but by a substantial change in the assumptions underlying those estimates.
570 The Tribunal accepts that position was available to the AER. It is not a position taken by the AER which the Tribunal regards as involving factual or other error on its part.
571 The AER drew attention to the fact that Cumpston Sarjeant in its “Response to Queries on Essential Energy Entitlements Valuation” 19 July 2012 (the Cumpston Sarjeant 2012 Essential Report) described the changed assumptions as “outside the range of realistic long term outcomes” and “unprecedented”. It says that its decision was consistent with the NEO, and with the encouraging of Networks NSW to pursue efficiency gains.
572 The AER’s reasoning involved a number of steps. Having identified the change in provisioning, resulting in a change in opex, it considered whether the change in expenses represented by the change in provisioning (largely based upon a change in the assumptions as to the quantification of provisioning for employee entitlement expenses or liabilities) should be rewarded or penalised under the EBSS.
573 It proceeded by considering whether, in those circumstances, the claim represented a fair sharing of efficiency gains and losses between DNSPs and network users, reflecting r 6.5.8(a). It considered whether, having regard to the need to ensure that benefits to electricity consumers likely to result from the scheme are sufficient to warrant any reward or penalty under the scheme for service providers (r 6.5.8(c)(1)), and more generally the desirability of both rewarding the service provider for efficiency gains or penalising it for its efficiency losses (r 6.5.8(c)(3)), it was appropriate to adopt the claim of Networks NSW. It considered whether the claim as so made accorded with the requirement of the 2008 EBSS, so as to satisfy it that the actual and forecast opex accurately reflected the costs faced by Networks NSW in the regulatory control period.
574 Having regard to the analysis of the reasons for those changes referred to above, the AER decided that the change in expenses attributable to those provisions did not represent real business outcomes, but were attributable to changes in underlying assumptions. Bluntly, it says, the expenses did not reflect costs actually faced by each of the Networks NSW entities in the regulatory control period. As there were no actual efficiency gains, and the changes in opex were as a result of different assumptions which might or might not prove to be correct, it did not consider it was appropriate to reward Networks NSW for those changes as efficiency gains.
575 The AER then took the additional step of saying that it considered the more appropriate way to reflect the cost faced by each Networks NSW was to use a cash accounting methodology. Under that methodology the AER would account only for expenses actually recorded in respect of payments actually made. It says it did not thereby alter the reported opex, affected (as it accepted) by the change in assumptions underlying the value of provisions for the amount of provisioning.
576 It reached that view notwithstanding that, as Networks NSW pointed out, the incentive to move to efficient costs is consistent over an entire regulatory period. That is because historical opex towards the end of a then current regulatory period is a key input for forecasting opex allowances for the new regulatory period so any incentive to reduce opex below the regulatory allowance diminishes towards the end of the then current regulatory period. It is, therefore, a fair observation (as Networks NSW made) that the EBSS represents a consistent incentive across regulatory periods. In that light, Networks NSW says that, in reality, the AER in the relevant Final Decisions, simply decided to abandon the EBSS in the next regulatory period for Ausgrid and Essential, and in relation to all the Networks NSW businesses it excluded retrospectively an additional cost category so that none of those businesses were deemed to be entitled to include, as part of efficiency gains or losses, what appears as movements in provisions.
577 Networks NSW identifies two relevant decisions in relation to the EBSS made by the AER, which it says require reconsideration by the Tribunal as they will make out the grounds of review asserted:
(1) the determination that differences between forecast opex and actual opex arising from what the AER identified as changes in provisions, largely in provisions for employees’ entitlements, would be excluded from the calculation of the EBSS carry-over amounts to be included as a building block in the annual revenue requirements for Networks NSW (the EBSS Decision); and
(2) secondly, in relation to both Ausgrid and Essential, the AER decision that having regard to its benchmark analysis the EBSS for 2013 should not apply in the 2014-19 regulatory period, so there should be no calculation of carry-over amounts arising from actual opex in 2015-19 then to be applied in the 2020-24 regulatory period (the EBSS Suspension Decision).
Networks NSW described the basis of the EBSS Suspension Decision as being that the operation of an incentive scheme in the form of the EBSS was inappropriate where the business was not being given an incentive to itself move to an efficient level of expenditure, but rather was being moved directly to a benchmark level of efficient expenditure by the AER. Consideration of the EBSS Suspension Decision is addressed later in these reasons.
578 The main debate is whether the AER was correct in adjusting the provisions expense reported in Networks NSW reported opex, so that it reflected liabilities actually settled and paid during a particular year, rather than changes in provisions for liabilities for accounting purposes made during that year. The AER, by making that adjustment, deducted an amount from the reported opex equivalent to the movement in provisions as illustrated in the course of submissions.
579 It should be noted that the AER did not, as suggested by Networks NSW at one point, coarsely simply exclude a category of expense. That is, of course, to draw a distinction between an expense actually paid on the one hand and a liability incurred, to be quantified in the future but the subject of a present estimate, on the other.
580 The AER referred to, and was probably prompted towards its analysis, by the apparent volatility in Ausgrid’s provisioning for employee benefits. In the 2009-10 and 2010-11 financial years, that provisioning was constant. It was increased by a very significant sum for 2011-12 and then somewhat reduced but by a materially significant sum for 2012-13. The accounting justification for those adjustments for 2011-12, is found, for example, in the Ernst & Young Report, June 2012, “Ausgrid – Actuarial assessment for specified employee entitlements as at 31 December 2011” (the EY 2012 Ausgrid Report). For 2012-13, the reduction is explained by the Cumpston Sarjent Report “Ausgrid – Actuarial assessment of long service leave and other employee entitlements as at 31 December 2012” (the Cumpston Sarjeant 2012 Ausgrid Report), including different assumptions to assess the salary promotional scale.
581 The AER submission points out that, as a result of those reports being given effect to, for the 2012 financial year Ausgrid recorded an actuarial adjustment expense for long service leave of about $58m in nominal terms, and for the succeeding financial year recorded a negative actuarial adjustment expense for long service leave of about $40m in nominal terms. That fall in expenses is, in essence, claimed by Ausgrid to be properly treated as an efficiency gain under the EBSS. Similar analyses were made in relation to Essential’s provisioning over that period and Endeavour’s provisioning over that period.
582 In the case of Essential, the adjusted present value of the provisioning for long service leave and other employee benefits very significantly increased between 2010-11 and 2011-12, and then significantly reduced (but well above the 2010-11 level) for the 2012-13 year. The reasons for those changes can be seen as attributable to the use of different discount rates in the latter two years, and the long term growth assumption being higher for the latter two years compared to earlier years. The significance of those changes is also illustrated by the change in the relationship between the “discounted” present value of those provisional liabilities and their nominal value.
583 Similar observations can be made about Endeavour’s treatment of provisions during the 2009-14 regulatory period. The “matched pairs” link under which it was assumed that the discount rate was always higher than the wages growth rate by a more or less constant percentage was broken.
584 The AER, in its Final Decisions, excluded the allowances claimed by Networks NSW businesses for efficiency gains for that final year as a result of movements in provisioning on the basis that they should not be treated as actual opex for EBSS calculations. It removed the movement in provisioning from each of the Networks NSW DNSP’s reported actual opex when calculating the EBSS carry over amounts because, it considered, the changes in provisioning were driven largely by changes in the discount rate and, at least in the case of Endeavour and Essential, by changes in salary growth assumptions used to value the provisions for long service leave. Those changes were, it considered, the result of accounting methodology and/or the result of assumptions made by the service provider or its actuary when there should be a minimal effect on rewards or penalties a service provider receives under the EBSS by the changes in provisioning. That is because, as the AER said:
The fundamental requirement for the EBSS under the NER is to derive efficiency gains and losses from the comparison of forecast and actual opex over the period, not merely accounting gains or losses.
See generally: Final Decision Ausgrid distribution determination 2015-16 to 2018-19, Attachment 9 - Efficiency Benefit Sharing Scheme at p 9-17; Final Decision Endeavour Energy distribution determination 2015-16 to 2018-19, Attachment 9 – Efficiency Benefit Sharing Scheme at p 9-18; and Final Decision Essential Energy distribution determination 2015-16 to 2018-19, Attachment 9 – Efficiency Benefit Sharing Scheme at p 9-16.
585 Having taken that approach, the AER also then rejected the submission of Networks NSW that, even if it was appropriate to exclude the changes in provisioning from actual opex, the forecast opex (from the earlier 2009-14 Determination), should be used to compare actual opex for the purpose of the EBSS adjusted to remove any movement in provisions embedded in the forecast. That contention was included in the Revised Regulatory Proposals of each of Networks NSW. The AER declined to do so, on the basis that its approval of the forecast opex for 2009-14 was not an approach made with reference particularly to provisioning of that character, so any attempt to identify what was “implicitly forecast at the time for provisions” would not be “robust given the hypothetical nature of this exercise”: see eg Final Decision Ausgrid distribution determination 2015-16 to 2018-19, Attachment 9 - Efficiency Benefit Sharing Scheme at p 9-17 to p 9-18.
The grounds of review
586 As with other grounds of review, there is some debate about how properly to characterise the errors which Networks NSW asserted in terms of s 71C of the NEL. The Networks NSW submission says that the AER made a number of material errors of fact in its findings of fact, namely:
.2 in concluding that the changes to provisions excluded by the AER from the application of the EBSS were produced predominantly by changes in assumptions;
.3 in concluding that changes to provisions were not actual costs incurred in delivering network services;
.4 as to the level of efficiency gains or losses, including by assessing such gains or losses:
.4.1 without taking into account changes in provisions;
.4.2 by excluding changes in provisions from actual expenditure but not excluding changes in provisions from forecast expenditure; and/or
.4.3 by assessing certain costs on a cash basis and other costs on an accruals basis.
587 Alternatively, it says that the AER’s decision involved an incorrect exercise of a discretion, or was an unreasonable decision because it was irrational, illogical and arbitrary, and inconsistent with the requirements of r 6.5.8(c) because:
(a) by excluding changes in provisions, the AER arbitrary, irrationally and illogically excluded actual costs (including the incurring of liabilities for holiday pay and long service leave) from the assessment of efficiency gains or losses;
(b) the AER’s approach was internally inconsistent by:
(i) excluding changes in provisions from actual expenditure but not excluding changes in provisions from forecast expenditure; and/or
(ii) assessing certain costs on a cash basis and other costs on an accruals basis.
588 Finally, as a further alternative, it says that the decision to exclude opex, in respect of changes of provision, constitutes an amendment to the operation of the 2008 EBSS by identifying an additional cost category for exclusion from the calculation of the carry-over amounts, when the 2008 EBSS itself, and the Rules, do not permit the amendment of the operation of the 2008 EBSS retrospectively in the Final Decisions. Consequently, it is said the Final Decisions on this topic involve a misconstruction and misapplication of the NER, and therefore an incorrect exercise of discretion and, alternatively, an unreasonable decision.
589 The AER’s broad position is that its approach involved no relevant “finding of fact”, but an exercise of judgment as to what methodology best gives effect to the aims of the EBSS. Then it says that its decision to adopt its approach was a reasonable decision, without demonstrable error which could enliven any of the asserted available grounds of review. That invites on the part of Networks NSW the proposition that the AER Final Decisions each rested on a fundamental step which was illogical, irrational or arbitrary so that the Final Decisions themselves had that character and therefore were unreasonable in relation to each of them. It also provokes a response on the part of Networks NSW that, to the extent that the AER’s decision involved an exercise of discretion, the discretion was incorrectly exercised.
590 It is accepted that, if the Tribunal accepts that the AER erred in its decision on opex, the challenge to the EBSS Suspension Decision in relation to Ausgrid and Essential will necessarily fall away as that part of their Final Decisions will be required to be varied pursuant to s 71P of the NEL. On the other hand, it is implicit from the absence of any detailed submissions separately addressing the EBSS Suspension Decision, that if the AER decision in relation to the EBSS is not shown to be in error, it will not be necessary separately to address the EBSS Suspension Decision.
591 The Tribunal does not comment on the extent to which either the EBSS Decision or the EBSS Suspension Decision would require reconsideration or amendment by reason of its decision in relation to opex generally.
592 However, as the fundamental approach to the EBSS allowance is obviously critical to any further revised final decision on the part of the AER, the Tribunal proposes to address the contentions concerning the EBSS Decision.
593 It is convenient to address the contentions of Networks NSW in the sequence in which they appear in its written submissions.
594 The first is to address the alleged error that the AER retrospectively excluded the particular category costs in any event.
595 Much of the Tribunal’s reasons for its conclusion on this series of alleged errors emerges from the discussion above.
596 The Tribunal does not accept that the AER’s process of reasoning was retrospectively to exclude a category of costs in implementing the EBSS for the 2014-19 regulatory period.
597 As the AER says, it accepted that the costs which the provisioning allowed for – the payment of employee benefits such as holiday pay and long service leave entitlements – were to be accounted for. It departed from the previous practice of allowing for those costs on the basis of the provisions for them in the accounts, and instead used the measure of the actual payments to the employee creditors as and when they were paid.
598 It would seem that that change in the means of measuring those costs was prompted (at least in part) by the changes in the Networks NSW provisioning referred to above. Where there is such a dramatic change in provisioning, it is hardly surprising that the AER should consider whether the method of measurement of those costs by provisioning was the most appropriate way to measure them for the purposes of the NEL and in accordance with the incentivising of efficiency gains built into the EBSS. As a matter of practical commonsense, it is hard to see (for example) how Ausgrid’s adjusted provisioning between 2011-12 and 2012-13 in fact represented any actual efficiency gain, so that the reduction in provisioning in the latter 12 month period should be taken into account in the application of any EBSS designed to serve r 6.5.8. Of course, that is a simplistic view, but it is nevertheless one which is not demonstrably fallacious.
599 If the AER were to question the appropriateness of measuring the cost of payments for employees for holiday pay and long service leave simply by the provisioning of each of the Networks NSW DNSPs, where those costs can vary dramatically by assumptions made from time to time by the provider (as they did), it was reasonable to investigate how the measurement of provisioning costs represented (as the 2008 EBSS expressed it at p 3) “genuine business outcomes that have arisen in the ordinary course of conducting the business in a prudent and diligent manner”. Indeed, the 2008 EBSS at p 5 refers to the AER making adjustments to correct for the variances in methodologies.
600 Networks NSW says that the consequence of the change in methodology by excluding an additional category of costs, at the point of the relevant Final Decisions, is not to incentivise Networks NSW (or to disincentivise Networks NSW) because the costs which are reflected in the provisioning have already occurred by the “promise of rewards or penalties”.
601 However, the Tribunal does not consider that that is a correct way to characterise what the AER has done. The AER does not resile from the proposition that the costs incurred by Networks NSW for labour on account of holiday pay and long service leave approved and payable should not be taken into account in determining the appropriate payments or allowances going forward for the 2014-19 regulatory period. It is not obliged to accept the provisioning estimates of Networks NSW for that purpose. It has chosen to make the appropriate allowance prospectively, based upon the actual payments made from year to year for those costs in the previous regulatory period.
602 In taking that step, at least in principle, the Tribunal does not consider that the AER has made any error of fact in its findings of fact.
603 The detailed grounds of review of Networks NSW are set out above. It appears that, at this point in the AER’s reasoning, the relevant asserted error of fact is that the AER concluded that changes to provisions were not actual costs incurred in delivering network services. However, whatever may be the available accounting methods and assumptions from time to time for the making of provisions, or the appropriateness of the reported provisioning in accordance with accounting standards, it is not the case that the provisioning of those costs meant that they were “actual costs” incurred in delivering network services; they were estimates of liabilities incurred and to be paid in due course determined in accordance with accounting standards and based upon assumptions made by each of Networks NSW about the various elements going to the making up of the amount which, ultimately, would be paid to meet those liabilities.
604 The next step in Networks NSW’s contentions is that the AER erred in concluding that the changes to provisioning by each of the Networks NSW entities were produced predominantly by changes in assumptions. That is the second of the two alleged errors in a sequence of four alleged errors.
605 It is clear that the movements in the liability to employees during a financial year may increase, or accrue, by reason of the employee having worked during that year. If the number of employees was unchanged, and none took leave, clearly the accrued or accumulated liability to those employees would increase over that year. The following confidential Figure 1 in the Networks NSW submissions on the EBSS forcefully makes that point.
[CIC] Figure 1: AER adjustment of Ausgrid opex for provisions for employee benefits [CIC]
606 The Networks NSW submission also points out (as the AER recognised) that changes in provisions may be induced by a change in the value of existing provisions such as a change in the discount rate (to determine the present value of the accrued liabilities when anticipated to be paid), or changes in integers (such as the assessment of the employee attrition rate and so changes in the amount and timing of the long service leave payments), and changes because of changed external conditions (such as changes in the discount rate) affecting the net present value of those accrued liabilities.
607 Then it is said that there was no material before the AER which suggested that relevant assumptions were being manipulated or that the calculations as to provisioning were otherwise than the best assessment of the liabilities of each of the Networks NSW entities, based upon independent third party actuarial assessment.
608 It is not a necessary part of the AER’s role to determine that the way in which the Networks NSW provisioning was made involved any improper manipulation of data, although (as it pointed out) there was some material in the Cumpston Sarjeant 2012 Essential Report at pp 2 and 9 which did not clearly support certain integers used in the provisioning calculation of Essential. It should be noted that that report (as its title says) is confined to the circumstances of Essential. It is a significant departure from that expert’s advice, as it meant that for 2011-12 the selected discount rate and selected salary growth assumptions resulted in a discounted present value of employee entitlements of 128.8 percent of their nominal value, compared to 98.4 percent of their nominal value on the assumptions suggested by that expert.
609 As noted above, the 2012-13 provisioning estimate for Endeavour, with an altered discount rate on the same salary growth assumption, resulted in a reduced figure for provisioning which (at least in a significant measure) constituted the claimed efficiency gained under the EBSS for that year.
610 The Tribunal has concluded above that the AER did not err by looking at, and behind, the provisioning by Networks NSW entities to decide whether their provisioning was an appropriate basis for measuring efficiency gains by those businesses.
611 In the Tribunal’s view, the AER was not in error in deciding to look more closely at the alternatives to accepting the provisioning for such liabilities. As an accounting exercise, of course, double entry accounting necessarily means that an increase in provisioning requires the complementary recording of an expense, even though the payment of that expense is not then made and may ultimately vary from the provisional liability. That is influenced by the timing of the employee in taking the leave, the applicable salary rate, and the like. Hence, the provisioning for such liabilities is generally based upon actuarial assessment, in turn based upon – amongst other things – assumptions as to the applicable discount rate to determine net present value of future liabilities and the future salary growth, and affected at least by age and years of service.
612 It is not unreasonable to expect, in the absence of particular circumstances, a stable relationship in the long term between the discount rate (as it varies from time to time) and the rate of future salary growth. That relationship was changed by changes in those elements, but not in a complementary way, so that it led to the significant variations in provisioning between 2011-12 and 2012-13 as referred to. It should be noted, as counsel for the AER acknowledged, that the precise critical assumptions by Ausgrid, Endeavour and Essential in their respective provisioning processes were not routinely the same and that at least in the case of Ausgrid, there was no consistent direct “matching” relationship between the two significant assumptions referred to.
613 Nevertheless, it follows from that discussion that the AER did not, in the view of the Tribunal, err by concluding that the changes to provisions were largely the consequence of changes in critical assumptions.
614 The material before the Tribunal includes the correspondence between the AER and the respective Networks NSW commencing mid-2014 as to whether the changes in certain assumptions (and the consequential disjunction of the complementarity between the two assumptions or the consequences of the two assumptions particularly discussed above) were matters which were in fact reflective of real efficiency gains or losses. That material was not shown to have been overlooked by the AER; indeed the strong inference is that it was carefully considered by the AER.
615 In the circumstances, the Tribunal is not persuaded that the AER erred by concluding that the changes to provisions by the Networks NSW entities were produced predominantly by changes in assumptions of the character referred to. Insofar as there is said to be an error of fact in that conclusion of the AER, that is not made out.
616 The third contention of Networks NSW is that the AER erred in concluding that changes in provisions were not “actual costs incurred in delivering network services”. They say that it is an unconventional position for the AER to determine that a business entity, adopting an approach in accordance with standard and universally recognised accounting practice, and in a conventional way, should therefore exclude from having such properly recorded costs accounted for in an assessment of “actual opex”.
617 Reliance was placed upon the decision of the Full Court of the Federal Court in Paciocco v Australia and New Zealand Banking Group Limited  FCAFC 50 in support of the position taken by Networks NSW. It is convenient at this point to remark upon that decision. The Tribunal does not regard it as directly relevant to the question which the AER (and the Tribunal) have to address: whether the changes to provisioning should be the subject of rewards under the EBSS, where those changes occur as a consequence largely of differences in assumptions made from time to time about critical elements in determining the net present value of future liabilities. The context for that question, of course, is whether such changes in provisioning (where the changes are the consequence of different assumptions and the relativity of the critical assumptions was not maintained as had previously been the case) should attract the benefit of efficiency gains under the EBSS.
618 The AER decided to reject the provisioning approach previously adopted and exposed by the adjustments made by Networks NSW (in particular by the Essential accounts and methodology for provisioning) and it decided not to directly reflect actual efficiency gains or losses by the accounting provisioning for those future liabilities. Its use of the term “actual costs” is in that context. It does not mean that the AER simply chose to ignore the costs incurred by Networks NSW, as reflected by the payments made by the Networks NSW businesses for those liabilities, or to ignore the fact that there are future liabilities which are reflected in the provisioning estimates.
619 The AER presented a graph prepared from figures in the Ausgrid Revised Regulatory Proposal and then presented in Figures 3 and 5 of the written Networks NSW submissions on the EBSS, showing the Ausgrid proposal for opex to be used for the EBSS proposal and the AER’s allowance of opex to be used for EBSS purposes. One line is calculated using the cash accounting methodology instead of Ausgrid’s accrual accounting methodology. It shows the extent to which the AER has in fact allowed for opex for that purpose. It cannot be said that the AER did not allow for those costs by the method which it adopted.
620 There was some debate about whether the “actual costs” allowed by the AER were in fact correctly measured. It cannot be said that the AER did not accept that liabilities for accrued leave and other employee entitlements were not actual liabilities. Its approach was to measure those liabilities, to be discharged at some time in the future and to be quantified ultimately by the movement in earnings, promotion and other variables, at “actual”, that is, paid amounts year by year. To the extent that the two critical projected elements of discount rates and wage movements (for the purposes of the provisioning) varied from time to time they would then be reflected in the amounts actually paid to meet those accrued liabilities from time to time. It is the Tribunal’s view that that was an approach reasonably open to the AER. That is not to gainsay the observations in the Ernst & Young report Advice on movement in provisions, January 2015 expressing at p 3 and elsewhere the proper accounting method for those liabilities, and its view that the “true economic cost” to Networks NSW was or included the provisioning for entitlements to be paid in the future.
621 But, it has not been shown that the AER, for the purposes of the EBSS, was in error in looking behind the provisioning. It did that. As a result of looking behind or into the reasons for the changes in provisioning, it did not accept that the provisioning by Networks NSW, and more specifically the movements in provisioning, represented efficiency gains (or losses). The Networks NSW contentions did not set out to show that the changes in the critical assumptions for provisioning (which the Tribunal has accepted, as identified by the AER) were themselves correct so that the AER, upon analysis, could not have concluded that they were the only, or visually the only, assumptions which could properly have been made by Networks NSW. Indeed, the Tribunal in the course of submissions was taken to material which would suggest that the assumptions made by Networks NSW from time to time were not the only reasonable assumptions on those matters.
622 Finally, in relation to this contention, the Tribunal notes that the comparison of the allowance for opex for EBSS purposes between the Ausgrid proposal for opex for that purpose and the Ausgrid Final Decision (as based on Figures 3 and 5 in the Networks NSW principal submissions) indicates that the cash accounting methodology of AER does not mean that Ausgrid (or Endeavour or Essential) were deprived of the benefits of the EBSS incentives, save for those which might have flowed from the changes in provisioning for the 2011-12 and 2012-13 years, largely influenced by the changed assumptions referred to above.
623 The fourth general category of attack adopted by Networks NSW is based upon the assertion of a change in methodology for calculating opex (in relation to, or for the purposes of, the EBSS). It is said that the AER should have excluded any movement in provisioning embedded in the forecast opex for the 2009-14 regulatory control period, so that looking forward there would have been a proper comparison of “apples with apples”.
624 In support of that contention, counsel for Networks NSW took the Tribunal to the AER’s Final Decision Efficiency Benefits sharing scheme for the ACT and NSW 2009 distribution determinations, February 2009, particularly Section 5.2: Measuring efficiency and the EBSS, and Section 5.5: Adjustment of actual and forecast opex. It is not necessary to refer in detail to all the passages referred to. The conclusions sections 5.2.3 and 5.5.3 are as follows:
5.2.3 AER conclusions
The AER considers it appropriate to utilise a rule of thumb in assessing efficiency gains under the EBSS. However, the AER also considers that the EBSS should, as far as possible, reflect efficiency gains and losses by DNSPs. To this end the AER will allow forecast opex to be adjusted for actual demand growth for the purpose of calculating carryover amounts. The AER will also consider for exclusion from the EBSS cost categories proposed by DNSPs in their regulatory proposal before the commencement of the regulatory control period and must be determined as uncontrollable by the AER in its final determination.
5.5.3 AER conclusions
The AER will make adjustments to forecast and actual opex for the purposes of calculating carryover amounts where it has been explicitly stated in the final determination at the beginning of the regulatory control period that those specific adjustments will be applied to the EBSS for that period.
Any cost categories that a DNSP considers to be uncontrollable and that should be excluded from the operation of the EBSS must be proposed in the DNSP’s regulatory proposal prior to the commencement of the regulatory control period. These cost categories will only be excluded if the AER considers them to be uncontrollable and their exclusion prudent.
The AER retains the right to exclude further cost categories from the operation of the EBSS. These cost categories must be outlined in the final determination at the beginning of the regulatory control period.
625 The Tribunal does not regard that material as supporting this contention of Networks NSW. Nor does it regard that material as demonstrating, or assisting to demonstrate, a ground of review as claimed by Networks NSW. The form of the discussions preceding those conclusions is to identify and give credit for real efficiency gains. Hence, the exclusion of uncontrollable costs. Indeed, at p 5 of the 2008 EBSS, the AER observed that it would react to changes in the methodology adopted by a DNSP to calculate opex. It is only a semantic debate to say that the changes in assumptions by Networks NSW did not, in the present circumstances, amount to a change in methodology.
626 It is also important to note that the AER for the 2009-14 regulatory control period did not simply adopt the Networks NSW proposed opex allowances, but substituted its own assessment.
627 The Tribunal does not, therefore, conclude that this contention demonstrates that a ground of review has been made out by Networks NSW.
628 For those reasons, Networks NSW has not demonstrated any ground for review, or more accurately the Tribunal is not satisfied that any ground of review exists, in relation to the AER’s Final Decisions concerning the EBSS.
629 In addition, having regard to the reasons why the claimed EBSS efficiency gains should not have been accepted (as found by the AER and accepted by the Tribunal), the Tribunal in any event would not have been satisfied that the restoration of the EBSS gains or rewards as claimed by Networks NSW would, or would be likely to, result in a materially preferable NEO decision.
630 The changes in provisioning do not represent in fact any real efficiency gain in the long term interests of consumers. They represent an accounting provision based upon changed assumptions, the reasons for which are not obvious or obviously correct, and which would introduce into the EBSS calculations a volatility which – as the material presently stands – is not necessarily a realistic reflection of the extent of the liabilities incurred. The changes between the three years specifically discussed above are sufficient to make that point.
631 In a sense, it is premature to make such a conclusion (because it is at the ultimate step of the Tribunal’s determination that s 71P(2a) and (2b) really come into the Tribunal’s consideration). In this instance, however, it can be said that if the EBSS were the only matter in respect of which a ground of review were made out, the Tribunal would not, for the reason given, be satisfied in terms of s 71P(2a)(c) so that it would not vary the AER Final Decisions. And, it can also be said that, even if this and other grounds of review were made out, as a constituent component of the Final Decisions, the AER’s decision on the EBSS would not weigh in the scales towards a favourable ultimate decision about whether to vary or set aside the Final Decisions.
632 The return on equity topic gives rise to common interests between the Network Applicants under the NEL and under the NGL in the case of JGN. They sensibly presented their collective submissions through the same counsel.
633 While, in broad terms, their contentions were supported by the Vic/SA Interveners and by Ergon, it is on occasions necessary to separately address particular aspects of the submissions of the Vic/SA Interveners and Ergon.
634 As an applicant, PIAC did not apply to have the Network NSW Final Decisions in respect of this topic varied or set aside. As an intervener, however, PIAC made submissions on particular matters in respect of the topic confined to the Networks NSW Final Decisions.
635 As noted, the annual revenue requirement for a DNSP for each regulatory year of a regulatory control period must be determined using a building block approach, which includes a building block for return on capital for that year: r 6.4.3(a)(2) of the NEL. Rule 6.4.3(b)(2) of the NEL provides that the return on capital is calculated in accordance with r 6.5.2. Rule 76 of the NGR is to the same effect.
636 Rule 6.5.2 of the NEL relevantly provides:
6.5.2 Return on capital
Calculation of return on capital
(a) The return on capital for each regulatory year must be calculated by applying a rate of return for the relevant Distribution Network Service Provider for that regulatory year that is determined in accordance with this clause 6.5.2 (the allowed rate of return) to the value of the regulatory asset base for the relevant distribution system as at the beginning of that regulatory year (as established in accordance with clause 6.5.1 and schedule 6.2).
Allowed rate of return
(b) The allowed rate of return is to be determined such that it achieves the allowed rate of return objective.
(c) The allowed rate of return objective is that the rate of return for a Distribution Network Service Provider is to be commensurate with the efficient financing costs of a benchmark efficient entity with a similar degree of risk as that which applies to the Distribution Network Service Provider in respect of the provision of standard control services (the allowed rate of return objective).
(d) Subject to paragraph (b), the allowed rate of return for a regulatory year must be:
(1) a weighted average of the return on equity for the regulatory control period in which that regulatory year occurs (as estimated under paragraph (f)) and the return on debt for that regulatory year (as estimated under paragraph (h)); and
(2) determined on a nominal vanilla basis that is consistent with the estimate of the value of imputation credits referred to in clause 6.5.3.
(e) In determining the allowed rate of return, regard must be had to:
(1) the relevant estimation methods, financial models, market data and other evidence;
(2) the desirability of using an approach that leads to the consistent application of any estimates of financial parameters that are relevant to the estimates of, and that are common to, the return on equity and the return on debt; and
(3) any interrelationships between estimates of financial parameters that are relevant to the estimates of the return on equity and the return on debt.
Return on equity
(f) The return on equity for a regulatory control period must be estimated such that it contributes to the achievement of the allowed rate of return objective.
(g) In estimating the return on equity under paragraph (f), regard must be had to the prevailing conditions in the market for equity funds.
Return on debt
[Subparagraphs (h)-(l) are under the subheading “Return on debt” and are relevant to that topic. They are separately the subject of consideration in the next principal section of these reasons for decision.]
Rate of Return Guidelines
(m) The AER must, in accordance with the distribution consultation procedures, make and publish guidelines (the Rate of Return Guidelines).
The Rate of Return Guidelines must set out:
(1) the methodologies that the AER proposes to use in estimating the allowed rate of return, including how those methodologies are proposed to result in the determination of a return on equity and a return on debt in a way that is consistent the allowed rate of return objective, and
(2) the estimation methods, financial models, market data and other evidence the AER proposes to take into account in estimating the return on equity, the return on debt and the value of imputation credits referred to in clause 6.5.3.
(n) There must be Rate of Return Guidelines in force at all times after the date on which the AER first publishes the Rate of Return Guidelines under these Rules.
(o) The AER must, in accordance with the distribution consultation procedures, review the Rate of Return Guidelines:
(3) at intervals not exceeding three years ...
637 Rule 87 of the NGR is relevantly in much the same terms.
638 In its application of r 6.5.2 of the NEL and r 87 of the NGR the AER followed a six-step methodology outlined in Chapter 5 of its Better Regulation Rate of Return Guideline, December 2013, (the RoR 2013 Guideline) to arrive at a return on equity of 7.1 percent.
639 The AER’s return on equity of 7.1 percent is to be contrasted with:
(a) Ausgrid, Endeavour and Essential’s proposed 10.11 percent;
(b) ActewAGL’s proposed 10.71 percent; and
(c) JGN’s proposed 9.83 percent.
640 The 2012 Rule Amendments significantly altered the process previously prescribed for determining the return on capital.
641 Prior to 2012, r 6.5.2 of the NER required the return on equity to be determined using the Sharpe Lintner Capital Asset Pricing Model (SL CAPM), and r 87 of the NGR required the return on equity to be determined using “a well accepted financial model, such as the Capital Asset Pricing Model”. The 2012 Rule Amendments removed these requirements and instead required that regard must be had to relevant estimation methods, financial models, market data and other evidence: NER r 6.5.2(e)(1); NGR r 87(5)(a).
642 A further significant change is the insertion of the allowed rate of return objective (RoR Objective), being the rate of return for a regulated service provider which is commensurate with the efficient financing costs of a BEE with a similar degree of risk as that which applies to the regulated service provider in respect of the provision of standard control services/reference services: NER r 6.5.2(b); (c) and NGR r 87(2) and (3). The relevant Rules now require that return on equity is to be estimated such that it contributes to the achievement of the RoR Objective and in estimating the return on equity, regard must be had to prevailing conditions in the market for equity funds: NEL 6.5.2(f); (g); NGR r 87(6) and (7).
643 The 2012 Rule Amendments include a requirement that the AER must publish Rate of Return Guidelines in accordance with its consultation procedures. These guidelines must set out the methodologies that the AER proposes to use in estimating the allowed rate of return, and the estimation methods, financial models, market data and other evidence the AER proposes to take into account: NER r 6.5.2(m) and (n); NGR r 87(13) and (14).
644 The RoR 2013 Guideline referred to above, published by the AER in December 2013, result from the 2012 Rule Amendments requirement. It is accepted that the RoR 2013 Guideline are not binding on the AER when it comes to make an individual determination and that there is no requirement for the AER to provide persuasive evidence to depart from the RoR 2013 Guideline. The AER is permitted to make a decision that is not in accordance with the RoR 2013 Guideline, but if it does so, then it must state its reasons for departing from its guidelines: NER r 6.2.8(c); NGR r 87(18).
645 As the AER pointed out, one benefit of the changes was to provide a common framework for the determination of the rate of return under the NER and the NGR.
646 References by the parties to the AEMC’s 2012 Rule Amendments accurately reflect that. In principle, the AEMC sought to achieve a process by the AER to get to the best estimate of the rate of return that can be obtained which reflects efficient financing costs of the service provider at the time of the regulatory determination.
647 The AEMC said at p 43:
A rate of return that reflects efficient financing costs will allow a service provider to attract the necessary investment capital to maintain a reliable energy supply while minimising the cost to consumers.
648 The AEMC noted that its regulatory approach could best achieve the NEO, the NGO and the RPP. The broadening of available methods, to be adopted by the AER at its election, was intended to fulfil that objective. It made the point at p 68 that “achieving the overall objective has primacy”. Thus, it saw the RoR 2013 Guideline, and the process by which they were to come into existence, as representing the correct balance between flexibility and regulatory certainty following the consultation required of and by the AER before fixing upon the Guidelines.
649 The AER has appropriately extracted from the 2012 Rule Amendments the following propositions summarising how it intended the 2012 Rule Amendments, in particular r 6.5.2 of the NER and r 87(2) of the NGR, to operate:
(a) the RoR Objective has primacy in any estimation of the rate of return on equity (pp 18, 36 and 38-39);
(b) the AER’s obligation to “have regard to” the material referred to in NER 6.5.2(e) when determining the allowed rate of return is subject to its obligation under NER 6.5.2(b) to determine the allowed rate of return such that it achieves the RoR Objective (and equally under NGR r 87(3) and 87(2)) (pp 36-37);
(c) the AER must actively turn its mind to the factors listed, but it is up to the regulator to determine whether and, if so, how the factors should influence its decision (if at all) (pp 36-37);
(d) it is important that the AER be given flexibility to adopt an approach to determining the rate of return that is appropriate to market conditions (p 44);
(e) it is important for the AER to be transparent in its approach to determining the rate of return in order to maintain the confidence of service providers, investors and consumers in the process (pp 23 and 24);
(f) it is important that all stakeholders (including consumers) have the opportunity to contribute to the development of the RoR 2013 Guideline and its evolution through periodic review every three years (pp 45-46);
(g) the RoR 2013 Guideline should include details as to the financial models that the AER would take into account in making a determination, and why it has chosen those models over other models (p 70);
(h) the RoR 2013 Guideline should provide a service provider with a reasonably predictable, transparent guide as to how the AER will assess the various estimation methods, financial models, market data and other evidence in meeting the overall RoR objective. The Guideline should allow a service provider to make a reasonably good estimate of the rate of return that would be determined by the AER if the Guidelines were applied (p 71); and
(i) while while the RoR 2013 Guideline are not determinative, these should “provide a meaningful signal as to the regulator’s intended methodologies for estimating return on equity” and be capable of being given “some weight” to narrow the debate about preferred methodologies and models. They should be used as a starting point in making a regulatory determination (p 71).
650 It is apparent also that the AEMC did not consider that the rate of return estimates should be driven by a single financial model, whether the SL CAPM or another model, or by one estimation method. The available relevant evidence should be considered. As the DNSPs and JGN pointed out, the AEMC recognised that, in any event, other models may be useful as all have weaknesses to some degree, including that they are all based on certain theoretical assumptions, so that no one model can be said to provide the right answer.
651 Indeed, it is commonly accepted that the AEMC’s view (see the AEMC’s 2012, Economic Regulation of Network Service Providers, and Price and Revenue Regulation of Gas Services, Draft Rule Determinations, 23 August 2012, at p 48) that “estimates are more robust and reliable if they are based on a range of estimation methods, financial models, market data and other evidence” is a sensible one.
652 Following the 2012 Rule Amendments, the AER from December 2012 undertook a careful consultation process, including through its Issues Paper of December 2012 then its Consultation Paper of May 2013, and its draft RoR 2013 Guideline of August 2013 before publishing the RoR 2013 Guideline on 17 December 2013.
653 In the RoR 2013 Guideline, the AER indicated that it proposed to apply a six-step methodology in order to determine the estimated rate of return on equity. Those six steps are:
(1) identify relevant material: in this step the AER identifies relevant methods, models, data and evidence;
(2) determine role: the AER then assesses each piece of material against a set of criteria that it set out in the RoR 2013 Guideline (discussed below). These criteria are then used to determine what role each piece of material would play in the determination of the return on equity. Under the AER’s approach, each piece of material could either be:
(i) used as the foundation model (noting there could be only one); or
(ii) used to inform the foundation model; or
(iii) used to inform the overall return on equity; or
(iv) not used in any way;
(3) implement foundation model: in this step the AER determines a range and point estimate for the foundation model return on equity, based on the information from step two;
(4) other information: the AER uses other information to inform the overall return on equity estimate;
(5) evaluate information set: the AER evaluates outputs from steps three and four above, identifying patterns and investigating conflicting information; and
(6) distil return on equity point estimate: the AER uses the foundation model point estimate informatively to determine a starting point. Based on the information from steps four and five, it selects a final return on equity value as the foundation model point estimate, or a multiple of 25 basis points (from within the foundation model range).
654 The RoR 2013 Guideline identified that the SL CAPM was to be used as the foundation model, the Black CAPM was to be used to inform the parameter estimate of the equity beta for use in the SL CAPM, dividend growth models (DGMs) were to be used to inform the parameter estimate of the market risk premium (MRP) for use in the SL CAPM, with no role for the Fama French three factor model (the Fama-French model). Figure 5.2 identified a broad range of other information which was relevant or potentially relevant to the AER’s task and is reproduced below:
Table 5.2 Role of other information.
Material (step one)
Role (step two
Commonwealth government securities
Inform foundation model parameter estimates (risk free rate)
Observed equity beta estimates
Inform foundation model parameter estimates (equity beta)
Historical excess returns
Inform foundation model parameter estimates (MRP)
Survey evidence of the MRP
Inform foundation model parameter estimates (MRP)
Inform foundation model parameter estimates (MRP)
Other regulators’ MRP estimates
Inform foundation model parameter estimates (MRP)
Inform foundation model parameter estimates (MRP)
Inform foundation model parameter estimates (MRP)
Inform the overall return on equity
Takeover and valuation reports
Inform the overall return on equity
Brokers’ return on equity estimates
Inform the overall return on equity
Other regulators’ return on equity estimates
Inform the overall return on equity
Comparison with return on debt
Inform the overall return on equity
Brokers’ WACC estimates
Other regulators’ WACC estimates
655 The AER made Final Decisions in relation to the return on equity for each of the Network Applicants in substantially the same terms. The AER’s Final Decisions on the return on equity for each of the Network Applicants are set out in Attachment 3 to each of the AER’s Final Decisions for the 2015-16 to 2018-19 period, each dated 30 April 2015, except for the JGN Final Decision dated 3 June 2015.
656 The AER Final Decisions rejected the Network Applicants’ proposal that the return on equity be calculated by reference to four models, being the SL CAPM, the Fama-French model, the Black CAPM and a CAPM informed by the dividend growth models (DGM).
The AER’s Foundation Model approach
657 Instead, the AER adopted the “foundation model” approach to estimating the return on equity as it regarded that approach to be consistent with the RoR 2013 Guideline. The AER used the SL CAPM as the foundation model, as it considered it to be superior to all other models for estimating the expected return on equity by reference to the BEE.
658 The AER considered that the SL CAPM:
(a) was the current standard asset pricing model of modern finance, both in theory and practice;
(b) has been in use for a long period to estimate expected equity returns and transparently represents the key risk and reward trade-off at the heart of the AER’s task;
(c) is widely accepted; and
(d) is consistent with the approach employed by financial market practitioners.
659 The AER did consider other models, including the Black CAPM, the DGM and the Fama French Model. The AER said it used the theory behind the Black CAPM to inform the equity beta to be used in the foundation model, and used the DGM to inform the MRP. It is noted that the Network Applicants contend that the AER erred at this point, because:
(i) it “disregarded” estimates from other models, provided through other experts reports, and did not itself use other models with its own estimated input data; and
(ii) gave “entirely subsidiary roles” to the other models, rather than using them to determine the return on equity.
660 It is therefore appropriate to note how the Final Decisions record the AER’s regard to other models.
661 The AER considered that the Black CAPM relaxes one of the key assumptions of the SL CAPM, namely the assumption that investors can borrow and lend unlimited amounts at the risk free rate. It is accepted that this leads the SL CAPM to underestimate the return required for low-risk investments. In place of that assumption, the AER said the Black CAPM assumes that investors can engage in unlimited short selling. It regarded this assumption as not reflecting how the stock lending markets work because short sellers are required to post collateral in the form of cash or equity when lending stock. It also noted that, in place of the risk free asset in the SL CAPM, the Black CAPM substitutes the minimum variance zero beta portfolio, which requires estimating an additional parameter (the zero beta expected return) in order to use the Black CAPM to empirically estimate the point estimate for return on equity.
662 However, consistent with the RoR 2013 Guideline, the AER said it did use the theory underlying the Black CAPM to inform the estimate of the equity beta. As the Network Applicants contend, all that the AER did was use the weakness of the SL CAPM identified by the Black CAPM as a rationale for selecting a beta at the top of the range suggested by the SL CAPM. As considered below, this approach may be justifiable, but it is hardly “using the theory”. It did not use the Black CAPM empirically to estimate the return on equity for the BEE, thereby not accepting the arguments advanced by the service providers to the contrary. It gave reasons for that conclusion.
663 DGMs use dividends forecasted by market analysts to derive the return on equity by assuming that the market value of the equity in the business is equal to the present value of future dividends. In the RoR 2013 Guideline, the AER determined that it would limit the use of the DGMs to the function of informing the MRP in the SL CAPM.
664 It did not use a DGM return on equity for the BEE. The AER’s reasons for that were:
(a) there was not sufficiently robust data of dividend yields for Australian energy network service providers;
(b) it was also unclear whether there was a sufficiently robust method for estimating the dividend growth rate for Australian energy network service providers; and
(c) the sensitivity of a DGM to its input assumptions limits its usefulness as a foundation model.
The AER noted that simple DGMs generated returns on equity for energy infrastructure businesses which significantly exceeded the average return on equity for the market. This, it said, did not make sense because a regulated natural monopoly was much less risky than the overall market, and should therefore have a lower return on equity.
665 The AER also considered that the Fama French Model was not appropriate to use. The risk factors used by the model are return on the market, firm size (measured by market capitalisation) and the ratio of book value to market value. By reference to the RoR 2013 Guideline, the AER considered:
(a) there was little evidence of companies or regulators using the model;
(b) empirical implementation of the model is relatively complex and opaque;
(c) its estimates are sensitive to the chosen estimation period and methodological assumptions;
(d) there is a lack of theoretical foundation for the factors; and
(e) the instability of parameter estimates and the backward looking observation of risk factors do not mean that those factors will apply on a forward looking basis.
666 The AER, then, needed to address the risk free rate, the equity beta, and the MRP for the purposes of its modelling.
667 In relation to the risk free rate, the AER was satisfied that the yields on Commonwealth government securities with a 10 year term to maturity represented a widely accepted proxy for the risk free rate. That is not contentious.
668 In relation to the MRP, after observing that the MRP cannot be directly observed, the AER considered a range of conceptual and empirical evidence to enable it to determine a point estimate that had regard to prevailing conditions in the market for equity funds.
669 The evidence that the AER had regard to in estimating the MRP was historical excess returns, DGM estimates (from its preferred construction of the DGM), survey evidence of the expectations of investors and market practitioners, conditioning variables (dividend yields, credit spreads and implied volatility) and recent decisions by Australian regulators.
670 The AER noted that there was no consensus among experts on which method produces the best estimate of the MRP and that estimates of it are diverse and can vary over time. As noted, the AER used DGM estimates (from its preferred construction of the DGM) to inform the estimate of the MRP, having regard to evidence that the output from the models is very sensitive to input assumptions and likely to show an upward bias in current market conditions. In that, it was supported by advice from McKenzie and Partington, in their report to the AER: Part A: Return on Equity, October 2014, at p 9 (the 2014 McKenzie Partington report).
671 In relation to the equity beta, the AER also noted that the equity beta also cannot be directly observed, and considered a broad range of information in order to inform its estimate. The evidence that the AER had regard to in estimating equity beta included empirical estimates based on Australian energy network firms, conceptual analysis of a BEE’s systematic risks relative to the market average, international empirical estimates and the theory of the Black CAPM. Consistent with the RoR 2013 Guideline, it appears to the Tribunal that the Black CAPM was not “used” to produce a beta estimate; only the idea that SL CAPM understates beta was used to justify selecting a high-end value.
672 The AER adopted an equity beta point estimate of 0.7 from a range of 0.4 to 0.7. It was satisfied that an equity beta of 0.7 is reflective of the systematic risk a BEE is exposed to in providing regulated services, and was likely to contribute to the achievement of the RoR Objective. The position adopted in the Final Decisions is the same as that set out in the RoR 2013 Guideline; that is, the AER did not depart from the RoR 2013 Guideline in its determination of the equity beta of the BEE.
673 Through the above assessments in the Final Decisions, the AER adopted a risk free rate of 2.55 percent, an equity beta of 0.7 and a MRP of 6.50 percent as the input parameter values for the SL CAPM. The return on equity estimated by the SL CAPM using these parameter values was 7.1 percent.
The principal concerns of the Network Applicants
674 The Network Applicants are critical of the outcome of the AER approach: a return on equity of 7.1 percent.
675 They submit, firstly the AER’s approach to the assessment of the equity beta is flawed because the AER wrongly ring-fenced the range of equity beta before considering other evidence to select within that range. Secondly, they submit the adjustment to the SL CAPM equity beta was in error because the AER’s adjustment in the light of the Black CAPM was arbitrary and not based on any empirical evidence.
676 They are also critical of the AER’s conclusion on the MRP because, they say, it unduly weighted historical average express returns and disregarded or discounted other relevant evidence, including the “Wright approach” (in very simple terms, an alternative means of calculating the MRP in the SL CAPM recommended by Professor Wright, see, for example: Wright, Review of risk free rate and cost of equity estimates: A comparison of UK approaches with the AER, October 2012). In this regard, they also say the AER erred in adjusting estimates of the MRP for the value of imputation credits by applying an incorrect formula in making that adjustment.
677 Finally, at a more general level, they contend that the estimate fixed for return on equity was not reasonable, and can be shown to have failed the appropriate cross-checks.
The AER’s consideration of other material
678 That complaint flows from the next step in the AER’s methodology. That was to undertake an examination of other information that could inform an overall estimate of return on equity as outlined in Figure 5.3 from Attachment 3 to the Ausgrid Final Decision. The AER considered the spread between debt and equity risk premiums, and return on equity estimates from the Wright approach, valuation reports, broker reports, and other regulators’ decisions. The AER considered that the range of MRPs derived from this other information supported its use of the foundation model and its foundation model return on equity estimate. Although investors’ required return on equity is unobservable, the AER noted that the observable MRPs for debt did not materially change in the face of declines in the risk free rate.
679 At its conclusion, the AER considered that an expected return on equity derived from the SL CAPM should be the starting point for estimating the return on equity, and that the other information did not indicate that the equity MRP estimate should be uplifted or downshifted to better contribute to the achievement of the RoR Objective.
680 The AER was satisfied that an expected return on equity estimate of 7.1 percent derived from its implementation of the SL CAPM would contribute to the achievement of the RoR Objective and was commensurate with the prevailing market conditions (as required by NEL r 6.5.2(f) and (g) and NGR r 87(6) and (7)).
681 PIAC, as might have been expected, carefully addressed the AER’s methodology and conclusions and the contentions of the Network Applicants critical of those methods and conclusions.
682 Equally of present significance, PIAC contended that the AER erred in selecting a point value of the equity beta of 0.7 from the range 0.4-0.7.
683 In preparing the RoR 2013 Guideline, and reflected in it, is the work of Professor Henry: Estimating β, 2009 (Henry 2009 Report) for the AER. It is referred to in the RoR Explanatory Statement. Professor Henry noted that the empirical estimates then available justified a point estimate of about 0.55 for equity beta. The AER nevertheless then, as it did in the Final Decisions, adopted a range of 0.4-0.7 as a reasonable range for equity beta of a BEE.
684 PIAC notes that after the RoR 2013 Guideline was published, Professor Henry provided a paper Estimating β: an update, April 2014 to the AER (Henry 2014 Report). The Henry 2014 report reported equity beta estimates consistently falling within the range of 0.4-0.7, but added that “most estimates clustered around 0.5”. That was expressly noted in the Final Decisions.
685 Based on the Henry 2014 Report, PIAC contends that the correct equity beta should have been 0.5, and so the AER erred by selecting a value of 0.7 at the top of the range which the RoR 2013 Guideline contemplated. It then says that, having regard to the obligations imposed on the AER by s 16(1)(d) of the NEL (as its applications concerned only the Final Decisions of the AER in relation to Networks NSW, it did not address the NGL), the AER should have proceeded on the basis of the equity beta of 0.5 to reach the materially preferable NEO decision.
Does s 71O of the NEL preclude PIAC from taking this position?
686 Before considering the issue whether PIAC is correct in its contention that the AER should have used an equity beta of 0.5, it is necessary to address the AER’s submission that PIAC is precluded from taking the point it has raised by reason of s 71O of the NEL. That section prescribes matters that may and may not be raised in a review.
687 Section 71O(2)(c) and (d) are the relevant provisions. They provide:
(2) In a review under this Subdivision, the following provisions apply in relation to a person or body, other than the AER (and so apply at all stages of the proceedings before the Tribunal):
(c) an affected or interested person or body (other than a provider under paragraph (a) or (b) may not raise in relation to the issue of whether a ground for review exists or has been made out any matter that was not raised by the person or body I a submission to the AER before the reviewable regulatory decision was made;
(d) subject to paragraphs (a), (b) and (c) –
(i) the applicant, or an intervener who has raised a new ground for review under section 71M, may raise any matter relevant to the issues to be considered under section 71P(2a) and (2b); and
(ii) any person or body, other than the applicant or an intervener who has raised a new ground for review under section 71M, may not raise any matter relevant to the issues to be considered under section 71P(2a) and (2b) unless it is in response to a matter raised by –
(A) the AER under subsection (1)(b)(iii); or
(B) the application under subparagraph (i); or
(C) an intervener under subparagraph (i).
688 As observed, PIAC is an applicant in respect of the NSW DNSP Final Decisions and an intervener in the NSW DNSPs’ applications. Also as observed, while PIAC as an applicant did not seek to have the Final Decisions on the return on equity varied or set aside, it has as an intervener made submissions on the appropriate retur