ESMA registers NEX Abide Trade Repository AB as a trade repository

The European Securities and Markets Authority (ESMA), the EU supervisor of trade repositories (TRs), has registered NEX Abide Trade Repository AB as a TR under the European Market Infrastructure Regulation (EMIR), with effect from 24 November 2017. NEX Abide Trade Repository AB is based in Sweden and covers the following derivative asset classes: commodities, credit, foreign exchange, equities and interest rates.

EMIR introduced provisions to improve transparency, establish common rules for central counterparties (CCPs) and for trade repositories and to reduce the risks associated with the OTC derivatives market. It provides for the obligation to centrally clear OTC derivative contracts or to apply risk mitigation techniques such as the exchange of collateral. It also provides for the direct supervision and the registration of TRs by ESMA as well as the recognition of non-EU TRs.

TRs are commercial firms that centrally collect and maintain the records of derivatives contracts reported to them. The registration of this TR means that it can be used by the counterparties to a derivative transaction to fulfil their trade reporting obligations under EMIR.

In order to be registered as a TR a company must be able to demonstrate to ESMA that it can comply with the requirements of EMIR, including, most importantly, on:

  • operational reliability;
  • safeguarding and recording; and
  • transparency and data availability.

The NEX Abide Trade Repository AB registration brings the total number of TRs registered in the EU to eight TRs, which can be used for trade reporting.

For more details on the list of registered TRs and the derivative asset classes which are covered by the registration, please refer to the following list.




ESMA registers NEX Abide Trade Repository AB as a trade repository

Download PDF



Questions and answers on the projects of common interest (PCIs) in energy and the electricity interconnection target

I. Projects of common interest (PCIs) in energy

 

1.  What are projects of common interest in energy?

Projects of common interest (PCIs) are key infrastructure projects aimed at completing the European energy market in order to help the EU achieve its energy policy and climate objectives: affordable, secure and sustainable energy for all citizens, as well as the long-term decarbonisation of the economy in accordance with the Paris Agreement.

2. Which projects are on the PCI list?

The 2017 list includes 173 projects, of which 110 are electricity and smart grids projects, 53 gas projects, and 6 oil projects. For the first time as well, the list of PCIs provides for four cross-border carbon dioxide network projects.

Electricity and smart grids projects will allow for the integration of renewable energy amounts and their transportation over long distances. They will also reinforce security of supply through increased grid resilience and flexibility, and will allow for involvement of demand response.

Gas PCIs are still needed to achieve diversification and to complete the integration of the energy markets in the EU and beyond, thus enhancing energy security and competitiveness.

The selected PCIs in the oil sector will also address the need to diversify oil supplies in the Central-Eastern EU region in order to improve energy security.

The full list of projects can be found here: https://ec.europa.eu/energy/sites/ener/files/documents/annex_to_pci_list_final_2017_en.pdf

3. Which criteria does a project have to meet to be included on the PCI lists?

For a project to become a PCI, it should be an energy network infrastructure that:

a)  has a significant impact on at least two EU Member States,

b)  enhances market integration and contributes to the integration of Member States’ networks,

c)  increases competition on energy markets by offering alternatives to consumers,

d)  enhances security of supply,

e)  contributes to the sustainability objective, e.g. by supporting renewable generation.

Only those electricity and gas projects included in the latest Ten-Year-Network Development Plans prepared by the European Network of Transmission System Operators for gas and for electricity (ENTSOG and ENTSO-E) are allowed to become PCIs, as these Plans highlight the projects’ socio-economic benefits on the EU energy system.

4. What are the benefits of being on the list?

PCIs benefit from a number of advantages:

a)  streamlined permit granting procedures (a binding three-and-a-half-year time limit);

b)  improved, faster and better streamlined environmental assessment;

c)  a single national competent authority (one-stop-shop) coordinating all permit granting procedures;

d)  a procedure allowing for the allocation of investment (construction) costs among Member States benefiting from the PCI;

e)  under specific conditions, possibility of receiving financial assistance under the Connecting Europe Facility (CEF) in the form of grants and innovative financial instruments.

5. The 2015 PCI list contained 195 projects. Why are there fewer projects this year?

This list contains 22 fewer PCIs than the (second) list adopted in 2015, mainly because 30 projects (22 gas and 8 electricity) from the 1st and 2nd lists will have been completed by early 2018, but also because of a clearer focus on key priorities.

Thanks to recent progress made, gas infrastructure has become in general more resilient and better integrated in most parts of Europe. However, the Eastern Baltic Sea region and the Central South-Eastern part of Europe remain more vulnerable to potential supply shortages. This vulnerability can be effectively addressed by the selected projects as they will provide the necessary diversification and integration.

6. Will the projects on the list automatically get EU funds?

Projects selected as PCIs can automatically benefit from many advantages stemming from the Trans-European Network – Energy (TEN-E) Regulation, including an accelerated permit granting and improved regulatory treatment.

PCI status is a precondition for grants under the Connecting Europe Facility (CEF), but it does not guarantee the award of a CEF grant. In more precise terms, grants to support studies can be requested for all PCIs (except for oil projects). However, when it comes to grants for works, only some projects included on the PCI list will need – and will be eligible for – financial assistance; many of them can be implemented on a commercial basis without EU funds – or other public subsidies.

Projects selected as PCIs, upon meeting specific additional criteria, will be able to apply for the CEF support in a separate procedure.

Oil projects are not eligible for any funding under the CEF.

7. When will PCIs be able to apply for EU funding?

The next call for proposals to receive grants under the Connecting Europe Facility (CEF) is scheduled for spring 2018. Project promoters will be able to apply for grants for studies and grants for construction works.

Grants for works, however, will be available only to those PCIs that are not commercially viable despite their positive impacts, such as ending the energy isolation of Member State(s), solidarity, security of supply or technological innovation. Objective cost-benefit analysis will be used to determine these parameters.

In addition to CEF funding, PCIs can also apply for support under other EU’s programmes, such as the European Fund for Strategic Investment (EFSI), and the European Structural & Investment Funds (ESIF) – in particular the European Regional Development Fund (ERDF).

8. Which projects have received co-funding under CEF?

In the 2014-2016 period, 74 PCIs have benefited from the CEF programme. The allocated grants for works and studies are worth € 1.6 billion in total (out of the total CEF energy budget of € 5.35 billion).

In the first three years of the CEF funds, more focus was given to gas projects critical for ending the energy isolation of the Eastern Baltic Sea region and ensuring security of supply in South Eastern Europe. This was justified by the high urgency and maturity of the projects.

As of 2017, CEF will concentrate more on electricity and smart grids projects focusing on the integration of renewables across borders as well as innovation, the digitalisation and smartening of the grid.

9. When will the next PCI list be published?

A new list of PCIs is prepared every two years, therefore the next, fourth, list will be published in 2019.. Project promoters may propose adding new projects to the current list. Completed projects will be removed from the list, as well as projects which have proven unfeasible or no longer needed given developments in the energy system. This allows the EU to adapt the list to evolving market conditions and infrastructure needs.

10. What is the Commission doing to make sure that the projects on the list will actually be completed?

The Commission cooperates closely with the project promoters, Member States and regulators to ensure the implementation of the PCIs in good time. Where and when needed, the Commission can propose corrective measures.

The monitoring of the PCIs is carried out by the Trans-European Energy Networks (TEN-E) Regional Groups and is based on annual reports prepared by the Agency for Cooperation of Energy Regulators (ACER) and the national competent authorities. These reports also recommend ways to overcome delays and difficulties in implementing PCIs.

To accelerate the development of PCIs in specific European regions facing particular challenges, the Commission has established four High-Level Groups.[1] These Groups are an important contribution to the PCI development process as they foster high-level political commitment, help reach agreement on regional action plans and facilitate smart mobilisation and optimisation of EU financial support.

11. How is the PCI list established?

At a first stage, project promoters submit their projects for selection as PCIs.

Secondly, so-called Regional Groups (chaired by the Commission and including representatives from the Member States, transmission system operators and their European networks, project promoters, national regulatory authorities, as well as the Agency for the Cooperation of Energy Regulators (ACER)) assess the projects’ compliance with the criteria and their European added value. Stakeholders are invited to take part in these meetings and bring their insight on the infrastructure bottlenecks and on the candidate PCIs into the assessment process.

Based on this assessment the Regional Groups adopt the regional lists of PCI candidates.

The Commission adopts the list of PCIs in the form of a delegated regulation on the basis of the regional lists.

12. Do stakeholders participate in the PCI selection process?

Yes. The PCI identification process involves stakeholders active in the field of energy, such as consumer and environmental protection organisations. These groups have dynamically participated in the meetings of the Regional Groups.

Furthermore, extensive public consultations have been organised to obtain the public’s views on the pertinence and added value of the proposed projects.

In addition, several bilateral meetings have been held between the interested stakeholders and project promoters, which have allowed for in-depth and constructive discussions on the projects characteristics and their potential social and environmental impact.

13. What are the next steps?

The list of PCIs (in the form of a delegated Regulation) will be submitted by the Commission to the European Parliament and Council, who have two months to decide whether to accept or object to the list. This period of two months may be extended by and additional two months upon their request. If within this timeframe neither the Parliament nor the Council rejects the list, it will enter into force. It should be noted that Parliament or Council can only accept or reject the list as such, but not request amendments to it.

14. Where can I obtain additional information about the PCIs?

More information can be obtained on the Commission website dedicated to PCIs.

This website includes an interactive map of the PCIs.

 

II. Electricity interconnection target

 

1. Why is a target needed?

Interconnections are the hardware of the electricity system and they contribute to ensure security of electricity supply, a well-functioning internal market for better competitiveness and sustainability thanks to the integration of increasing amounts of renewable energy.

–   In 2002, the European Council called for all Member States to achieve an interconnection target of at least 10% of their installed electricity production capacity by 2020. This means that each Member State should have in place electricity cables that allow at least 10% of the electricity that is produced in the country to be transported across its borders to its neighbouring countries.

–   The 10% electricity interconnection target has provided political momentum to advance key cross-border projects. It is an important and useful policy tool that helps guide the development of trans-European electricity infrastructure.

–   Today’s Communication addresses the issue of what should happen after 2020. It translates the 15% headline target for 2030 into three operational thresholds, each of which providing a good indicator for the energy policy objectives of security, market integration and sustainability. This will help monitor progress of electricity network development and point to where urgent action is needed.

 

2. Will reaching the target make a difference on our energy bills?

The greater flexibility and availability for energy suppliers will reduce wholesale prices. With greater competition, these savings will be passed on to the consumer, translating into lower household bills. A study from 2013 estimated that a fully integrated energy market could save EU consumers a total amount of €12-40 billion every year. These benefits are much higher than the cost of grid investments.

 

3. Why isn’t there a similar target for gas?

In fact there are already binding infrastructure standards enshrined in the Regulation concerning measures to safeguard the security of gas supply (Regulation (EU) 2017/1938). That said, gas is different from electricity, there is a high share of imports, thus gas is transported through pipelines crossing several borders before it is consumed. Therefore an infrastructure target similar to electricity would not reflect the real needs.

4. How will the electricity interconnection target be reached?

The Commission has asked the European Network of Transmission System Operators for Electricity (ENTSO-E) to measure electricity interconnection levels annually and to report to the Commission and ACER.

The main tool for reaching the interconnection targets is through the implementation of Projects of Common Interest (PCIs). The third list of PCIs was adopted today. All these projects benefit from accelerated licensing procedures, improved regulatory conditions, and in some instances have access to financial support. The PCIs will significantly contribute to the achievement of the target. The Commission also underlines that whenever new projects are being developed to reach the target, the final decision for new infrastructure should be always based on a thorough cost-benefit analysis. Only interconnectors that can demonstrate potential benefits outweighing costs should be developed.

The implementation of the PCIs is being monitored in the Regional Groups and in addition, challenges to the implementation of PCIs is regularly discussed and addressed at the Copenhagen Infrastructure Forum.

5. Is the electricity interconnection target a legally binding target?

The interconnection targets provide important political guidance to close infrastructure gaps in Europe by triggering urgent action by the relevant Member States, transmission system operators, project promoters and regulators.

Projects that are necessary to reach the interconnection targets will be closely followed by the Commission, the Regional Groups established under the TEN-E Regulation and the political High Level Groups.

In addition, challenges to the implementation of PCI projects, such as of financial, regulatory, technical or public acceptance nature, will be regularly discussed and addressed at the Copenhagen Infrastructure Forum.

[1]         https://ec.europa.eu/energy/en/topics/infrastructure/high-level-groups




Benoît Cœuré: Policy analysis with big data

Speech by Benoît Cœuré, Member of the Executive Board of the ECB, at the conference on “Economic and Financial Regulation in the Era of Big Data”, organised by the Banque de France, Paris, 24 November 2017

The recent financial crisis, and the euro area sovereign debt crisis that followed, were characterised by periods of increased heterogeneity, market fragmentation and sudden turns in economic activity. This often made it difficult for economic policymakers to understand and assess in real time the underlying forces driving economic behaviour. Both traditional statistical datasets and our models proved at times inadequate to support the decision-making process, reflecting long time lags, linear assumptions and the absence of more granular information.

These events increasingly boosted the efforts in policy circles to obtain timelier and richer data for policy analysis, in short big data.[1] This push towards more granular information was not a revolution, however. It can be argued that big data, under different guises, have been used as an input into policymaking since Adolphe Quételet’s Mémoire in 1848. Since then, big data has been central to business cycle analysis, from the early work of Clément Juglar to the contributions of both Wesley Mitchell and the Cowles Commission, right up until today. According to central bank mythology, former Federal Reserve Chair Alan Greenspan would sit in his bathtub perusing sheets of statistics. And indeed, economic historians have analysed how the power of governments has been shaped by statistics – and vice versa.[2]

The most recent push towards more granular data was thus an evolution rather than a revolution, triggered in part by the emergence of new opportunities – themselves a reflection of rapid technological progress – and the experience gained over several years of crisis management.

This evolution is already bearing fruit. Policymakers today have access to a large number of micro datasets, often very different in nature and scope. Some are the result of new financial regulations. Others are by-products of increased use of technology. What they have in common, however, is that, if used appropriately and responsibly, they can help policymakers to extract more timely and diverse economic signals, and thus are a meaningful complement to existing official data.

In my remarks this morning, I will take stock of the progress made at the ECB, and in the central banking community more generally, on the use of more granular data in the conduct of monetary policy.[3] I will show that micro data collected by central banks themselves, in particular data on transactions between banks, have already proven to be an additional valuable guide for policymakers in devising policy responses.

And I will show that data generated through the greater use of technology in our daily lives have an enormous potential to help policymakers overcome prevailing constraints on timely data availability, understand better the consequences of their policies and calibrate them accordingly, while also creating challenges.

Central banks as big data collectors

Let me start with the role of central banks as producers of big data.

Central banks do not have to be at the forefront of data collection, and we should not seek to displace private sector efforts. That said, there are areas in which central banks have started, or are about to start, collecting large amounts of data to help them monitor developments in financial markets and allow them to extract richer information about the transmission of monetary policy, which in turn helps us calibrate our policies.

As a showcase for these efforts, I would like to discuss two such initiatives, namely the money market statistical reporting (MMSR) data, which the ECB, in collaboration with the wider Eurosystem, has been collecting since July 2016, and the so-called AnaCredit project, short for “analytical credit datasets”, which will produce its first results over the course of next year.[4]

MMSR data contain confidential daily information on the individual euro-denominated loans in the euro money market from the 52 largest euro area banks, which collectively account for approximately 80-85% of the total balance sheet of euro area banks. At present, this means collecting information on 10,000 daily transactions in the unsecured money market, with a daily volume of around €100 billion. We also collect data on around 30,000 daily transactions on secured loans, worth around €500 billion.

Earlier this week, the project took an important step forward with the publication of the first set of euro area money market statistics, covering each of the Eurosystem’s reserve maintenance periods in 2017.[5] The published data cover the unsecured market. The aim is to publish data on the secured market in 2018 once we are satisfied that the data are of sufficiently high quality.

The sheer volume of data, combined with the high frequency with which it is collected, means that the standard verification process involving human beings is not feasible. Carrying out checks by algorithm, using machine learning techniques and artificial intelligence is one way of ensuring that data remain of high quality. The Eurosystem has a steep learning curve in front of itself.

These efforts will no doubt pay off. Indeed, MMSR data have proven highly useful to policymakers in their short period of existence. Let me give you an example. The data help us assess the impact of the ECB’s asset purchase programme (APP) on market functioning, as I explained in more detail in a speech I gave last week in Brussels.[6]

MMSR data show that since the ECB launched the APP there has been a marked shift away from trades backed by general collateral, which are traditionally used to manage cash. Instead, there has been a growing share of repos off the general collateral curve, termed specials. Take for example the German Bund market. Traditionally, only around 5% of bonds in the German repo market traded as special, but in the second half of last year, that share rose to 50%.

Being aware of the distribution of trades and the premium placed on special bonds enables us to assess the impact of the APP on market functioning. To ease potential frictions, we decided last December to allow cash to be used as collateral in our securities lending programme, and to permit APP purchases below the deposit facility rate. As a result of these decisions, the “specialness” premium on long-term German bonds has declined, and the share of bonds that trade special is now around 30%.

A further notable area of use for MMSR data has emerged from growing concerns about the reliability and robustness of current risk-free benchmark rates for the euro area.[7] Banks have become increasingly reluctant to participate in benchmark panels, owing to concerns about potential litigation, compliance risks and costs. The resulting uncertainty in the integrity of reference rates represents a potentially serious source of vulnerability and systemic risk.

Against this backdrop, the ECB recently announced that it intends to provide a new overnight unsecured interest rate, using MMSR transaction data. The new rate is intended to complement existing benchmark rates produced by the private sector and serve as a backstop reference rate, with the aim to start publishing by 2020.[8] This move is motivated by our desire to mitigate the potential adverse impact on the monetary transmission mechanism and on financial stability from the lack of reliable benchmarks. The market facilitating role played by the ECB in this field is consistent with the tasks conferred upon it by its Statute.

The process will involve broad consultation with market participants, end-users and other public authorities. A first public consultation will be launched before the end of 2017.

The use of MMSR data implies that the new overnight interest rate will differ from the current EONIA benchmark in a number of ways.[9] Accordingly, any decision pertaining to this new benchmark rate does not bear direct consequences for the choice of our operational target.

Let me now say a brief word about the second initiative I mentioned earlier, namely the AnaCredit project, which will push the frontier of big data use at the ECB further out.[10]

AnaCredit will deliver loan-by-loan information, mostly on a monthly basis, on credit to companies and other legal entities extended by euro area banks. Early estimates point to around 70 million exposures reported every month, representing loans granted by about 4,500 credit institutions to more than 15 million counterparties.[11] In view of these large expected data volumes, a state-of-the-art IT infrastructure is currently under development, which will also take into account the need to ensure the adequate protection of confidentiality.

AnaCredit data collection has been designed to produce a complete picture of the credit exposure of the reporting population. The information collected comprises almost 100 different “attributes” covering various aspects of the credit exposure, such as the outstanding amount, maturity, interest rate, collateral or guarantee, information on the counterparty, and many other things.

Access to highly granular loan-level information will be a major step forward in helping policymakers to analyse and monitor credit developments and to assess the impact of their decisions on bank lending. As you know, bank lending plays a key role in the euro area, where the share of loans in the total external financing of small and medium-sized enterprises (SMEs) is considerably higher than in the Anglo-Saxon world, where market financing plays a more important role.

Granular AnaCredit data will help us look beyond aggregates and extract the underlying developments. Granular loan-by-loan data will allow us to know the characteristics of specific groups of counterparties involved in each transaction, without of course revealing their identities. This means that we will be able to assess the driving forces behind aggregate developments and distinguish genuine and healthy growth from potential exuberance. This is crucial for policymakers.

And, finally, although no data will be collected specifically for supervisory purposes in the initial stages of the project, the information will also be very useful in many areas of banking supervision. I am thinking in particular of the information on the link between lenders and the unique identification of counterparties across the entire lender population.

The use of big data in policy analysis

Let me now turn to the broader issue of the use by central banks of technology-driven data.

An early and prominent example of online data being used for policy analysis is the “Billion Prices Project”, which was launched by the Massachusetts Institute of Technology (MIT) in 2008.[12] Today, the project publishes daily online price indices for more than 20 countries, based on a technology called “web scraping”, where price information is collected automatically by machine from hundreds of retailers that also have physical outlets. By 2015, about 15 million prices were being collected every day from 900 retailers.

The advantages for policymakers are plain to see. Online inflation data, if of high quality, are much timelier than current price statistics and may cover a much larger number of products. Eurostat currently collects, via the national statistical institutes, roughly three million prices for the Harmonised Index of Consumer Prices every month in the 28 EU Member States and it publishes the index 17 days after the end of each month.

Online price data can therefore be used to improve short-term forecasts and to check the robustness and reliability of current price indices. Mixed-frequency models, for example, could be used to enhance existing forecast models that are based on monthly data.

Sveriges Riksbank has already tested the use of online price data. The results indicate that the data add value when it comes to forecasting short-term developments in consumer prices for fruit and vegetables.[13] Preliminary explorative analysis at the ECB confirms the predictive power of online price data, provided they are sufficiently granular.

This suggests that online prices may complement existing price data. Daily online data can also add significant value to policymaking in periods of high uncertainty. As research has shown, would the full Billion Prices Project data have been available in 2008, the turning point in US inflation following the demise of Lehman Brothers could have been identified months before it showed up in the official US consumer price Index.[14]

Online prices are, however, only one source of potential interest for policymakers. Large-scale barcode scanner data, for example, are an alternative avenue that researchers at the ECB have explored in the past to investigate factors that determine prices and the degree of price dispersion, much in the tradition of our long-standing efforts under the umbrella of the Eurosystem Inflation Persistence Network.[15] The data, which consist of 3.5 million observations on the price and quantity of individual products sold, have confirmed that competition at producer and retail level is a key factor affecting micro price-setting.

Insights gained from the use of online and scanner data have also encouraged the ECB to make micro-price research a strategic research priority between 2018 and 2020. The first step is to pick up where the Inflation Persistence Network left off, by resuming the collection of micro data underlying the Consumer Price Index and the Producer Price Index and complementing them, where possible, with scanner and online price data.[16] Collaboration within the Eurosystem will once again be vital for the success of this ambitious project.

The ECB, with the help of the national central banks, also draws on big data to help improve business cycle analysis. Google search data, for example, have been suggested as a potentially valuable source of data for policymakers. Back in 2012, Hal Varian and co-author showed that Google searches can help predict economic activity.[17]

Building on this idea, ESCB staff have explored the possibility of nowcasting unemployment using volumes reported by Google Trends for a large number of search queries broadly related to unemployment. They find that many of the search terms indeed correlate with unemployment and may reduce forecasting errors by up to 80% by comparison with naive benchmark models.[18] Job search and related Google search quotes have also been found to be strong predictors of variations in subjective wellbeing.[19] Similar initiatives are underway for more leading indicators.

Related to this, electronic payment data from credit cards and cash withdrawals from ATMs have been shown to help forecast private consumption as well as GDP growth, provided that they are made available in a timely manner.[20] These data are also less prone to revision than traditional national accounts data. Our internal findings so far suggest that, in some euro area countries, the correlation between payment data and private consumption is encouragingly strong, making their systematic use in forecasting an option for policymakers.

Technological progress in exploiting textual information can provide similar benefits. Research shows that information extracted from business newspapers can be used to nowcast quarterly GDP growth and outperforms traditional forecasting methods at turning points.[21]

Similar methodologies can also be used to extract information on how central bank communication is perceived by the public. Specifically, text mining techniques are employed by the ECB to determine whether its communication has been interpreted as dovish or hawkish by the media and to assess its impact in unconventional times.[22] Given the increasing role of communication in determining the monetary policy stance, such analysis may usefully complement the feedback extracted from financial markets.

Challenges in using big data for policy analysis

Overall, these examples highlight the fact that big data can help policymakers overcome some of the shortcomings of traditional macroeconomic time series. In short, these are: improved timeliness, richer detail on interactions and the ability to test and develop old and new theories using data not previously available. For sure, big data have also provided plenty of opportunities to deepen our understanding of behavioural economics, and how psychology can drive macroeconomic developments.

At the same time, some of the data raise new challenges, many – but not all – of them are related to the statistical production process itself.

For example, the sheer scale and variety of big data result in the difficulty of processing unstructured data. The database interfaces used by central banks were designed in the past mainly to handle time series data. Strides have certainly been made to cope with cross-sectional data, such as those used in supervision. But putting in place effective data processing methods and interfaces that permit the retrieval and visualisation of big data represents a considerable IT challenge.

It is also a challenge in term of human resources. When recruiting staff, central banks have not traditionally sought experience with techniques developed to cope with big data. Building up expertise in this new field takes time, and it also involves strategic planning on the recruitment side. But central banks are not the only employers wanting these skills, and competing against the private sector for the limited pool of highly sought-after experts is likely to prove challenging.

Apart from these operational considerations, there are also analytical, legal and ethical concerns related to the use of such data.

For example, perennial pitfalls involved in analysing traditional data also apply to big data. One misconception about big data is the common view that we no longer need to concern ourselves with sample bias, as large volumes of data take precedence over standard sampling theory, and that such data provide a census.[23]

Any statistician will know that this is of course not true. Take a dataset that includes all tweets ever made. Despite being a census of tweets, people who tweet are likely to differ in age, preferences and behaviour from those that do not, so several segments of the population are underrepresented, or do not appear in the sample.

Making assumptions about household behaviour without adequately re-weighting the sample is likely to result in biased and inaccurate estimates. Traditional datasets, designed to be representative of the entire population, might therefore sometimes be less subject to sample biases, although they are significantly smaller in size.

Similarly, big data also involve analytical challenges related to, for example, econometric identification. This is also referred to as the “curse of dimensionality”, (sometimes called the “large p, small n paradigm”) where there are many parameters but few observations.[24] Dynamic factor models, or Bayesian shrinkage, are methods which can help address the difficulties arising from the high dimensionality of data and, in fact, the term “big data” first appeared in econometrics in 2000 in relation to these models.[25] But some of these methods are still being developed, in particular for cases involving many observations – a situation which we will increasingly face over time as our datasets grow even larger.

Furthermore, the ability of big data to link individuals across various datasets raises important questions about privacy and confidentiality. A special Eurobarometer survey in 2015 showed that two-thirds of respondents were concerned about not having complete control over the information they provided online. A quarter of respondents considered it a risk that personal information could be shared with third parties, such as government agencies.

With this in mind, the European Parliament in April last year approved the EU’s General Data Protection Regulation (GDPR), which will come into force in May 2018. This Regulation will provide strong personal data protection, while giving individuals greater control over their own data.[26] But statistical agencies and central banks are already taking serious steps to ensure that privacy standards are fully upheld when they make use of new big data, such as credit card transactions or Google search data. Anonymity in transaction data helps build trust and allows policymakers to leverage new technology to help take better policy decisions for the common good.

Differences in national data protection regulations, in turn, together with differences in data standards, also pose a challenge, in the form of our global ability to collect, aggregate, disseminate and share data. In today’s interconnected world, a global approach to data collection is a prerequisite for developing a coherent view of the global economy and financial system, and for identifying vulnerabilities. But many legal hurdles to cross-border data sharing remain to be overcome, and implementation gaps to be filled, if the intended benefits of the data harmonisation reforms in the financial field are to be fully achieved.[27]

Regulation also plays a big role in decisions about accessibility. Consider the revised Payment Services Directive (PSD2), which will enter into force in a few weeks. One outcome of the new Directive is that banks will soon lose the monopoly they have on their customers’ transaction data.

The implication is that new fintech firms will be able to use these data to design new tailor-made products and to allow real-time access to financial services anywhere and at any time. This can be expected to foster productivity. At the same time, regulators need to be mindful of the potential financial stability implications, in particular if incumbents fail to rise to the challenge, and fintech firms thus start to crowd banks out of a large range of financial services. Granting access to big data therefore also has the potential to shift economic structures.[28]

It even has the potential to change the course of policy. So far, policy has largely relied on data produced by governments, often in the form of abstract concepts, such as GDP.[29] But the more widespread ability to create individual statistics threatens the hegemony of policymakers in defining the parameters of policy. One risk is that individuals will use other, readily available sources of data that could undermine confidence in official statistics and, hence, in the central bank’s commitment to price stability.

Just as there are concerns about “fake news” dominating social media, there is a risk of “fake”, or at least poor quality, statistics driving out better quality ones in public discourse. And just as false reports may quickly go viral, reliance on internet search data in the assessment of the economic situation may become self-fulfilling. Actions by economic agents could become less anchored to actual activity and more prone to manias and panics, with obvious implications for economic and financial stability.

The challenges here are twofold for policymakers. The first is to meet the demand for more individually tailored information. Some statistical offices, for example, already offer the possibility to calculate a personalised inflation index using official price quotes, but with expenditure weights provided by the user. The second challenge is to recognise that public perceptions of what we should target may differ from the abstract definitions we employ. Over time, engagement with the public may result in changes to definitions, and even in the conduct of policy.

This brings me to my last challenge, which has a philosophical dimension to it. Over time major central banks have tended to become more transparent, in the belief that communication would aid the effectiveness of monetary policy.[30] This has led to growing interaction between central banks and financial markets, simply because communication is not a one-way street.

You can see what I’m getting at. New technological advances may create a new “monkey in the mirror” mimetic loop[31], this time not through financial markets, but through big data. That is, we may one day be tempted to draft our monetary policy statements and speeches in the light of how they will be comprehended and interpreted by artificial intelligence algorithms. So in the future big data may work two ways, with central banks acting on, and reacting to, it, with consequences which remain to be understood.

Conclusion

Let me conclude.

Central banks have made considerable progress in recent years in integrating big new datasets into their policy analysis and decision-making. Granular data collected by central banks themselves have, in particular, become an indispensable source of information for policymakers.

Yet, it is probably fair to say that we are still exploring how to use much of the big data that new technology opens up to us. Although evidence is growing that online data may provide tangible benefits for short-term forecasting, more research is needed to ensure that the data are of sufficient quality and reliability to systematically inform policymaking.

The potential of such data to enrich central bank analysis in the future is considerable, however, as are the challenges that come along with it. For this reason, I encourage all of you to continue your efforts and to work in the same collaborative spirit as you have done so far.

Thank you.




Lighting the way: violence against women must stop

Tomorrow, on 25 November, the world marks the International Day for the Elimination of Violence against Women. On the eve of this occasion Vilnius Town Hall and EU House, where the European Institute for Gender Equality (EIGE) is located, will be lit up in orange. This is how EIGE will place its host city on the map of the global awareness initiative “Orange the World” and calls for action to stop violence against women. Presidential Palace and the Ministry of Foreign Affairs will also join the campaign and light their buildings in orange.

 “Violence against women is both a cause and a consequence of persisting gender inequalities in all aspects of life. Women’s lower status in society can place them at risk of violence and the experience of violence can undermine their ability to access education, obtain or keep a job. This is a problem also in Lithuania”, said Virginija Langbakk, EIGE’s director.

As part of the Gender Equality Index 2017, EIGE developed a way to measure the extent of violence against women in the European Union. The research made by EIGE measures the prevalence but also the severity and disclosure of violence against women.

The research reveals that one in three women in Lithuania have experienced violence since the age of 15 and in most of the cases by more than one perpetrator. Evidence shows that violence is widespread not only in intimate relationship, but also in public spaces such as schools or workplaces. In more than half of the cases, violence caused long lasting physical and psychological harm to the victim.

The majority of women who disclosed their experience of violence, claimed to have told about the incident to the police or someone close to them. However, violence still takes very serious forms in Lithuania. The number of femicide by intimate partner is the highest among all the 13 EU Member States, which collect this data.

 “These figures show how important it is to address the reported violence immediately. It is essential to build mechanisms to help women in need and gain their trust. The violence unfortunately tends to continue and an assault that might seem of lesser importance can lead to severe consequences”, Langbakk stressed.

In Lithuania, attitudes towards violence against women is also a concern. Eurobarometer survey in 2016 revealed that almost half of Lithuanians consider that the victim often provokes violence. This number is almost three times higher than the EU average. In addition, 42 % of Lithuanians said they think women often make up or exaggerate claims of abuse or rape. This number is twice as big as the EU average.

The violence we know about is only a fraction of reality. In the EU as a whole, almost one in two victims do not disclose violence to anyone. Recent allegations of sexual harassment in Hollywood and European Parliament encouraged women to come forward with their own stories.  The testimonies highlighted how common it is for women to experience harassment and stay silent about the incident. Most often, the perpetrators are men in a position of power.   

Combating violence against women is a priority to the European Commission. Commission launched a campaign of focused actions to further its commitment to eradicating all forms of violence against women a year ago. In support of Commission’s efforts, EIGE has prepared factsheets on gender-based violence in all the 28 member states. You can find the factsheet for Lithuania here.

Vilnius Town Hall and the EU House will be lit up in orange from Friday, 24 November, 17:00 to Sunday, 26 November. The Presidential Palace will also be lit up in orange on 25 November and the Ministry of Foreign Affairs of Lithuania be lit up until 10 December.

Gender Equality Index is a unique tool that measures the complex concept of gender equality and assists in monitoring progress of gender equality across the EU over time. EIGE released third edition of the Index on 11 October. Explore here.

The measurement structure for the domain of violence measures three aspects: prevalence, severity and disclosure of violence against women. On a scale of 1 to 100, 1 represents a situation where violence is non-existent and 100 represents a situation where violence against women is extremely common, highly severe and not disclosed.

The score for Lithuania is 25 compared to 27.5 for the EU as a whole.

Scores of the satellite domain of Violence are not included in the final Index score due to the lack of comparable data.

For more information, please contact Donata Matuleviciene, donata.matuleviciene@eige.europa.eu  or phone +370 5 2157 449

Download the report here

Download the factsheet here

Explore Lithuania’s score here

Follow us on Facebook and Twitter for more updates and news

Subscribe to our monthly newsletter here