MarketData valuation
Company Profile

Data valuation

Data valuation is a discipline in the fields of accounting and information economics. It is concerned with methods to calculate the value of data collected, stored, analyzed and traded by organizations. This type valuation depends on the ownership, reliance, uniqueness, scarcity, type, reliability and field of data.

History
In the 21st century, exponential increases in computing power and data storage capabilities (in line with Moore's law) have led to a proliferation of big data, machine learning and other data analysis techniques. Businesses increasingly adapt these techniques and technologies to pursue data-driven strategies to create new business models. Traditional accounting techniques used to value organizations were developed in an era before high-volume data capture and analysis became widespread and focused on tangible assets (machinery, equipment, capital, property, materials etc.), ignoring data assets. As a result, accounting calculations often ignore data and leave its value off organizations' balance sheets. Notably, in the wake of the 9/11 attacks on the World Trade Center in 2001, a number of businesses lost significant amounts of data. They filed claims with their insurance companies for the value of information that was destroyed, but the insurance companies denied the claims, arguing that information did not count as property and therefore was not covered by their policies. A number of organizations and individuals began noticing this and then publishing on the topic of data valuation. Doug Laney, vice president and analyst at Gartner, conducted research on Wall Street valued companies, which found that companies that had become information-centric, treating data as an asset, often had market-to-book values two to three times higher than the norm. On the topic, Laney commented: "Even as we are in the midst of the Information Age, information simply is not valued by those in the valuation business. However, we believe that, over the next several years, those in the business of valuing corporate investments, including equity analysts, will be compelled to consider a company's wealth of information in properly valuing the company itself." == Characteristics of data as an asset==
Characteristics of data as an asset
A 2020 study by the Nuffield Institute at Cambridge University, UK divided the characteristics of data into two categories, economic characteristics and informational characteristics. Economic characteristicsData is non-rival. Multiple people can use data without it being depleted or used up. • Data varies in whether it is excludable. Data can be a public good or a club good, depending on what type of information it contains. Some data can reasonably be shared with anyone who desires to access it (e.g., weather data). Other data is limited to particular users and contexts (e.g., administrative data). • Data involves externalities. In economics, an externality is the cost or benefit that affects a third party who did not choose to incur that cost or benefit. Data can create positive externalities because when new data is produced, it combines with already existing data to produce new insights, increasing the value of both, and negative externalities, when data may be leaked, breached or otherwise misused. • Data may have increasing or decreasing returns. Sometimes collecting more data increases insight or value, though at other times it can simply lead to hoarding. • Data has a large option value. Due to the perpetual development of new technologies and datasets, it is hard to predict how the value of a particular data asset might change. Organizations may store data, anticipating possible future value, rather than actual present value. • Data collection often has high up-front cost and low marginal cost. Collecting data often requires significant investment in technologies and digitization. Once these are established, further data collection may cost much less. High entry barriers may prevent smaller organizations from collecting data. • Data use requires complementary investment. Organizations may need to invest in software, hardware and personnel to realize value from data. Informational characteristicsSubject matter. Encompasses what the data describes, and what can it help with. • Generality. Some data is useful across a range of analyses; other data is useful only in particular cases. • Temporal coverage, Data can be forecast, real-time, historic or back-cast. These are used differently, for planning, operational and historical analyses. • Quality. Higher quality data is generally more valuable as it reduces uncertainty and risk, though the required quality varies from use to use. Greater automation in data collection tends to lead to higher quality. • Sensitivity. Sensitive data is data that could be used in damaging ways (e.g., personal data, commercial data, national security data). Costs and risks are incurred keeping sensitive data safe. • Interoperability and linkability. Interoperability relates to the use of data standards when representing data, which means that data relating to the same things can be easily brought together. Linkability relates to the use of standard identifiers within the data set that enables a record in one data set to be connected to additional data in another data set. == Data value drivers==
Data value drivers
A number of drivers affect the extent to which future economic benefits can be derived from data. Some drivers relate to data quality, while others may either render the data valueless or create unique and valuable competitive advantages for data owners. • Exclusivity. Having exclusive access to a data asset makes it more valuable than if it is accessible to multiple license holders. • Timeliness. For much data, the more closely it reflects the present, the more reliable the conclusions that can be drawn from it. Recently captured data is more valuable than historic data. • Accuracy. The more closely data describes the truth, the more valuable it is. • Completeness. The more variables about a particular event or object described by data, the more valuable the data is. • Consistency. The more a data asset is consistent with other similar data assets, the more valuable it is (e.g., there are no inconsistencies as to where a customer resides). • Usage Restrictions. Data collected without necessary approvals for usage (e.g., personal data for marketing purposes) is less valuable as it cannot be used legally. • Interoperability/Accessibility. The more easily and effectively data can be combined with other organizational data to produce insights, the more valuable it is. • Liabilities and Risk. Reputational consequences and financial penalties for breaching data regulations such as GDPR can be severe. The greater the risk associated with data use, the lower its value. The process of realizing value from data can be subdivided into a number of key stages: data assessment, where the current states and uses of data are mapped; data valuation, where data value is measured; data investment, where capital is spent to improve processes, governance and technologies underlying data; data utilization, where data is used in business initiatives; and data reflection, where the previous stages are reviewed and new ideas and improvements are suggested. == Methods for valuing data==
Methods for valuing data
Due to the wide range of potential datasets and use cases, as well as the relative infancy of data valuation, there are no simple or universally agreed upon methods. High option value and externalities mean data value may fluctuate unpredictably, and seemingly worthless data may suddenly become extremely valuable at an unspecified future date. Information-theoretic measures, such as entropy, information gain, and information cost, are useful for anomaly and outlier detection. In data-driven analytics, a common problem is quantifying whether larger data sizes and/or more complex data elements actually enhance, degrade, or alter the data information content and utility. The data value metric (DVM) quantifies the useful information content of large and heterogeneous datasets in terms of the tradeoffs between the size, utility, value, and energy of the data. Such methods can be used to determine if appending, expanding, or augmenting an existent dataset may improve the modeling or understanding of the underlying phenomenon. Infonomics valuation models Doug Laney identifies six approaches for valuing data, dividing these into two categories: foundational models and financial models. Foundational models assign a relative, informational value to data, where financial models assign an absolute, economic value. Foundational modelsIntrinsic Value of Information (IVI) measures data value drivers including correctness, completeness and exclusivity of data and assigns a value accordingly. • Business Value of Information (BVI) measures how fit the data is for specific business purposes (e.g., initiative X requires 80% accurate data that is updated weekly – how closely does the data match this requirement?). • Performance Value of Information (PVI) measures how the usage of the data effects key business drivers and KPIs, often using a control group study. Financial modelsCost Value of Information (CVI) measures the cost to produce and store the data, the cost to replace it, or the impact on cash flows if it was lost. • Market Value of Information (MVI) measures the actual or estimated value the data would be traded for in the data marketplace. • Economic Value of Information (EVI) measures the expected cash flows, returns or savings from the usage of the data. Bennett institute valuations Research by the Bennett Institute divides approaches for estimating the value of data into market-based valuations and non-market-based valuations. • A consumption-based approach builds on the principles in the modified cost value approach by assigning data users different weightings based on the relative value they contribute to the organization. These weightings are including in the modelling of data usage statistics and further modify the measured value of data. • Data hub valuation uses a cost-based approach that measures the cost of data hubs where large repositories of data are stored, rather than measuring the cost of separate datasets. The data hub cost can then be modified, as in the consumption based and modified cost value approaches. Another hub valuation approach uses a modified market value approach, by measuring savings to users from accessing data via hubs versus individually accessing data from producers, and user willingness-to-pay for access to data hubs. • A stakeholder approach engages key stakeholders to value data, examining how data supports activities which external stakeholders identify as creating value for them. It uses a model that combines the total value created by the organization, a weighted list of value creating initiatives (as defined by external stakeholders) and an inventory of data assets. This approach was developed in a collaboration between Anmut, a consultancy firm, and Highways England, a public sector agency for which data valuations based on market value, income gains or economic performance are less meaningful. The approach can also be applied in the private sector. Companies performing Data Valuations getting Data Assets on Balance Sheet Data Capitalisation Partners (DCP: Data Capitalisation Partners' mission is the listing of Data on Balance Sheet as an independent asset class by applying GAAP, IFRS & AASB Accounting Standards and International Valuation Standards Counsel (IVSC) compliance rigour to issue mathematically defensible valuations that transform unrecognised intangible data assets into collateralisable financial assets to significantly increase Enterprise Value. DCP does this by leveraging Specialised Data Asset Valuation methods for Collateralisation of Data & Capital Markets. • Data Capitalisation Partners delivers IVSC-compliant benchmarks to provide definitive Data Asset values, equipping organisations with rigorous methodology and supporting Accounting frameworks to achieve formal Balance Sheet recognition of Data Assets, ensuring transparency for Auditors, Regulators and Global Capital Markets. The team have realized over $8.2 Billion in data asset value for companies.. • DCP Valuation methodologies ensure full IVSC compliance, providing mathematically defensible Data Asset pricing, the foundation of regulatory and audit transparency. Establishing rigorous, evidenced-based values for intangible Data Assets mitigates systematic risk and proactively counters formation of 'asset bubbles.' This precision-driven framework fundamentally de-risks the capital structure, safeguarding investor interests while providing the critical transparency required in corporate and institutional debt and equity markets. Data Valuation as a Service provides: • A data valuation report from 17 different data valuation methodologies and calculations to create a defensible valuation of your data unique to your company and its data. • An interrogation of data via data due diligence and for strategy, security, governance, monetization, substantiation, security, privacy and people • A data monetization strategies review against each use case in order to glean as much current and future value of data as possible. • Analytic evidence of data value as well as model forecasts for data drivers, use case, and monetization impacts to the data valuation of your data ==References==
tickerdossier.comtickerdossier.substack.com