ENTRIES TAGGED "publishing data"
To weather industry disruption, publishers essentially need to become leaders of big data and data intelligence companies.
In a recent interview at PBS’ Media Shift, Jason Ashlock, founder and president of Movable Type Management, addressed the changing roles of publishers and argued that they’re not innovating fast enough. Ashlock argues that we’re in the age of the author and direct audience engagement and that publishers need to become conduits for this engagement and curators of communities in addition to curators of quality books.
In a similar vein, Forbes writer Suw Charman-Anderson recently argued that publishers need to become retailers. As retailers, publishers put themselves in a position to collect customer data, which in turn puts them in a better position to offer customers unique additional value and experiences. O’Reilly’s Joe Wikert is a big proponent of the direct sales channel.
To weather the disruption in the industry, publishers do need to become strong multi-media companies as Ashlock suggests and retailers as Charman-Anderson suggests, but more than that — more to the root of that — publishers essentially need to become leaders of big data and data intelligence companies in order to capitalize on the benefits of these business models. They need to learn how to sift through and analyze data to extract meaning, quite a different business than traditional publishers are used to and not an easy task. As Adam Frank argues at NPR, making use of big data and data intelligence requires specialists who understand the intricacies and nuances of data, who know how to “separate the chaff from the real, useful insights.”
The digital transformation in publishing is bringing forth more than new reading platforms, gadgets and distribution options — it also brings a wealth of data publishers have never before had access to, data that can be applied to new marketing and production strategies, and used to help create more efficient business models.
As data becomes more and more central to publishing ecosystems, traditional methods of metric collection and analysis are proving insufficient. This need for new measurement techniques has given rise to a new metrics approach called “alternative metrics.” I reached out to Todd Carpenter, Executive Director of NISO, to find out what’s behind the changing data needs and more about how altmetric applications can benefit publishers. Carpenter will explore this topic further at TOC Frankfurt on October 9, 2012. Our interview follows.
What are alternative metrics?
Todd Carpenter: Alternative metrics — referred to as “altmetrics” — are a suite of assessment criteria and measures that are being developed, particularly in the scientific and academic communities, to assess the importance of a particular work of scholarly output in a new way.
Traditional metrics have been downloads, citations, or sales — generally based on publication-level data. For example, the Thomson Reuters Impact Factor, one of the most widely used metrics in scholarly publishing, measures quality at the journal level by measuring the number of citations to it in other journal articles. As academic publishing has expanded and diversified, these traditional metrics have been increasingly criticized for issues such as their granularity (i.e., measuring at the publication level, not the item level), or their bias toward citation, which is a common practice among researchers but doesn’t reflect more applied, practical, or public use.
The scope of measures that could be considered altmetrics is actually quite broad, ranging from analysis of usage data to social media references; Google Page Rank; deep statistical data analysis techniques, such as betweenness centrality; and other relatedness statistical measures. Also considered for inclusion in alternative metrics are measures of non-traditional types of content production, such as the release of scientific data sets, blog posting, or social media activity — none of which are addressed in traditional metrics.
James Levy explains how publishers can use the Hiptype platform to make smarter publishing decisions.
As our industry shifts from print to ebooks we’re discovering a wealth of new data to study. Retailers hold most of the cards for this data, but a startup named Hiptype is looking to change that. In the interview below, Hiptype’s president and CEO James Levy (@jamtoday) talks about how their platform works and how it can lead to making smarter publishing decisions.
A call for NewCo to expand its focus, ereading data is influencing content, and Hugh McGuire talks ebooks at TEDxMontreal.
B&N plans to open Nook stores worldwide; Joe Wikert says their store focus need a technology turn. Elsewhere, WSJ reporter Alexandra Alter looks at data generated by ereading, and Hugh McGuire argues ebooks belong on the web at TEDxMontreal.
The FAA looks at passenger gadget use, Joe Wikert digs deep into data, and PEW reports on news.
Airline passengers in the U.S. may soon be allowed to use ereaders and tablets during takeoff and landing. In other news, O'Reilly's Joe Wikert reveals the power of direct sales, and the key phrase for news is "mobile."
The second in a series looking at the major themes of this year's TOC conference.
Several overriding themes permeated this year’s Tools of Change for Publishing conference. The second in a series looking at five of the major themes, here we take a look at data in publishing — how publishers can benefit, practical applications, and innovative ways it can be used.
Peter Collingridge on real-time data and analytics in publishing.
Peter Collingridge, cofounder of Enhanced Editions, says big data can be eye opening for publishers. In this interview, Collingridge talks about the role of real-time data and analytics in publishing and about a new market intelligence service for books.