“Quantitative Formalism: An Experiment” and a related thought about pop music

A few years ago I did a lot of reading about algorithms and machine learning as it related to the arts and popular culture. What immediately sprang to mind when I read the opening of “Quantitative Formalism: An Experiment” was an article from WIRED in 2011 about at team from The University of Bristol that worked on developing an equation that could predict a hit song.

http://www.wired.com/2011/12/hit-potential-equation/

At the top of the article you’ll see a video that shows the “evolution of musical features” as they relate to hit songs. Since we will soon be considering ways to display results from our data sets I thought this might be interesting to take a look at.

The short article considers both the Bristol team’s work and other similar projects related to predicting the popularity of new pop music. While this is not scholarly work, I thought it was interesting to share and consider how this type of enquiry is being used outside of the academy.

 

A few resources from the recent NEH Project Director’s Meeting

I recently attend an NEH Project Director’s meeting in DC and wanted to share news of a few projects that I thought might interest the group:

Massmine.org

(in pre-release)
http://www.massmine.org/

splash

“MassMine is a social media mining and archiving application that simplifies the process of collecting and managing large amounts of data across multiple sources. It is designed with the researcher in mind, providing a flexible framework for tackling individualized research needs. MassMine is designed to run both on personal computers and dedicated servers/clusters. MassMine handles credential authorizations, data acquisition & archiving, as well as customized data export and analysis.”

Poemage: A Visualization Tool in Support of Close Reading

http://www.sci.utah.edu/~nmccurdy/Poemage/

Screen Shot 2015-10-10 at 7.42.03 AM

“Poemage is a visualization system for exploring the sonic topology of a poem. We define sonic topology as the complex structures formed via the interaction of sonic patterns — words connected through some sonic or linguistic resemblance — across the space of the poem. Poemage was developed at the University of Utah as part of an ongoing, highly exploratory collaboration between data visualization experts and poets/poetry scholars. Additional details are provided in the companion paper [to appear in IEEE Transactions on Visualization and Computer Graphics].”

New DHDebates CFP is Up

We’ve been looking at various barometers of the field in recent weeks (conferences, blog postings, tweets, etc.), so I wanted to mention that Lauren Klein and I posted the Call for Papers for the 2017 edition of Debates in the Digital Humanities today. Some of you might be interested in looking through it to get a sense of the questions we’re asking and how they map to current concerns in the field.

Screen Shot 2015-10-08 at 6.25.26 PM

Social Justice and the Digital Humanities

Hey all,

Just ran across a site that might be of interest to those interested in postcolonial, or more generally, social justice approaches to DH. Social Justice and the Digital Humanities is a site that emerged from one of the courses at HILT 2015 (Humanities Intensive Learning and Teaching). The three-day intensive course combined theory and praxis, similar to our class. Discussions included questions about: access, material conditions, method, ontologies and epistemologies. I like that much of the content are questions, followed by further readings/references. It also included a list of DH projects with a social justice approach in mind.

One project I thought was pretty cool is Map of Native American Tribes . Aron Carapella, a self-taught mapmaker, mapped out original names and locations of native american tribes before their contact with Europeans. This transformed the entire makeup of how we visualize the map of the United States.

I also found out about #transformdh, “an academic guerrilla movement seeking to (re)define capital-letter Digital Humanities as a force for transformative scholarship by collecting, sharing, and highlighting projects that push at its boundaries and work for social justice, accessibility, and inclusion”

-Maple

Future History on the Web

After reading “I nevertheless am a historian” I came across another article about a Texas high school and a mother who complained about her son’s history book. The book titled “Geography” claimed, immigrants were brought over to the new world to be workers, When in reality, the immigrants they were talking about were Africans. The book, further stated that Europeans were brought over, as indenture servants. The latter is partly true, but omitting most of the facts, especially, how Africans were workers, instead of slaves is a lie perpetrated by the book publishers, McGraw Hill, and the school advisory board that approved it. I could not help but think what I read in the Robert S. Wolff piece “People with little or no formal training in the discipline have embraced the writing of history on the web, which raises the question, whose histories will prove authoritative in the digital age?” The publishers have formal training, etc. Wow!
What becomes of history in the digital age remains to be seen, already we are seeing revisionists change history in printed books, and on the web. The fact that a big time book publisher like McGraw Hill would even put their name on a history book that changes and leaves out pertinent details, such as slavery, is an insult to the average person’s intelligence, and a crime against education.

I agree with Corey Meyer in the “Black Confederate Solider” piece, it helps when those who are claiming with authority about an historical account that one should work with original source materials, and have a understanding of the background of how to work with them.  I guess McGraw Hill did not get that memo…..

TACIT – A New Tool for Text Collection and Analysis

Thanks to Sava Saheli Singh, whose weekly round-up for the GC’s own Journal of Interactive Technology and Pedagogy brought a new tool to my attention: TACIT, Text Analysis, Collection and Interpretation Tool. From the website:

Though several limited-method tools for text analysis are already available (e.g. LIWC), and some have become part of standard statistical packages (e.g., SPSS Text Analytics), a unified, open-source architecture for gathering, managing and analyzing text does not exist.

The Computational Social Science Lab (CSSL) at the University of Southern California introduces TACIT: An Open-Source Text Analysis, Crawling and Interpretation Tool.
TACIT’s plugin architecture has three main components:

  • Crawling plugins, for automated text collection from online sources (e.g., US Senate and Supreme Court speech transcriptions, Twitter, Reddit)
  • Analysis plugins, including LIWC-type word count, topic modeling, sentiment analysis, clustering and classification.
  • Corpus management, for applying standard text preprocessing to prepare and store corpora.

TACIT’s open-source plugin platform allows the architecture to easily adapt with the rapid developments text analysis.

The tool is available on Github for those interested in checking it out. A related paper can be found on SSRN.

I have not used this tool, so if anyone here tries it out, please report back!

Clio, Mnemosyne and Internet

Cohen’s and Rosenzweig’s Introduction to Digital History: A Guide to Gathering, Preserving, and Presenting the Past on the Web, 2005 was a very interesting reading because it echoed with some texts I’ve read in the last few months.
In this post I would like to discuss two main assumptions on the potentiality of the web: infinite possibilities of recording everything, easy access to all information on the web.

Can we really record everything?
I thought yes. Then a few months ago I read this article in The New Yorker where the author tells the wonderful story of the Internet Archive and its extreme usefulness in tracking and recording the digital world. What made me understand this important task was a bunch of illuminating sentences:
“No one believes any longer, if anyone ever did, that “if it’s on the Web it must be true,” but a lot of people do believe that if it’s on the Web it will stay on the Web.”
– “Web pages don’t have to be deliberately deleted to disappear. Sites hosted by corporations tend to die with their hosts. When MySpace, GeoCities, and Friendster were reconfigured or sold, millions of accounts vanished.”
– “Facebook has been around for only a decade; it won’t be around forever. Twitter is a rare case: it has arranged to archive all of its tweets at the Library of Congress.”
– “The Web dwells in a never-ending present. It is—elementally—ethereal, ephemeral, unstable, and unreliable. Sometimes when you try to visit a Web page what you see is an error message: “Page Not Found.” This is known as “link rot,” and it’s a drag, but it’s better than the alternative. More often, you see an updated Web page; most likely the original has been overwritten.”
– “According to a 2014 study conducted at Harvard Law School, “more than 70% of the URLs within the Harvard Law Review and other journals, and 50% of the URLs within United States Supreme Court opinions, do not link to the originally cited information.”
– “Last month, a team of digital library researchers based at Los Alamos National Laboratory reported the results of an exacting study of three and a half million scholarly articles published in science, technology, and medical journals between 1997 and 2012: one in five links provided in the notes suffers from reference rot. It’s like trying to stand on quicksand.”

And as Cohen and Rosenzweig affirm, the dream of preservation ad infinitum is indeed a dream: “The current reality, however, is closer to the reverse of that—we are rapidly losing the digital present that is being created because no one has worked out a means of preserving it. The flipside of the flexibility of digital data is its seeming lack of durability—a second hazard on the road to digital history nirvana.”
We should not take for granted the storage power of the web. Therefore we should improve our efforts in archiving the digital. Not just for the sake of academic articles but mainly because much of our life is indeed on the web.

But then a second question, to which I really don’t have any answer, came up to my mind.
Do we have to record everything?
Should this be also a moral question? Is it right to record everything just because we can do it? If we try to make a parallel between a conversation among two friend in a bar, and a conversation among two friend on the FB wall of one of them, we can clearly see the difference of how something that was highly ephemeral becomes something that is highly easy to capture and preserve. Maybe this is not the right example, because things change over time and we are talking about two different spaces. But, the act of communication is the same.
Recently, in Europe there has been much discussion on the right to oblivion. What do you think about this?
As an historian, of course, I would love to have all the possibilities to reconstruct the past. But, we have to acknowledge the importance of oblivion in societies (much of history research is indeed constructed around the effort to understand social amnesia and the role of memory in defining individual and collective identities). Should we leave this right to the vagaries of history or people should have the right to rewrite their past?  Should oblivion even be a right?

Access
I guess that to this issue, there is alo a question of access. Much of these data, of personal information exchanged by people, is owned by private companies that use these same data to make more profit (see the well-establisehd practice of Terms of Service – By the way, here a wonderful tool to understand the different terms of service of various web services). In other words, it seems to me that when we talk about the web and the richness of its data we too quickly assume an easy access to these same data. As Jeremy Rifkin wrote in The Age of Access, in the next future, which is already now since the book was published in 2000, power will gravitate around those who control the access to information. (see also Cohen and Rosenzweig: “A more serious threat in digital media, which runs counter to its great virtues of accessibility and diversity, is the real potential for inaccessibility and monopoly.”)

Probably one effective answer is given by common creative and open sources projects (Internet Archive, Open Culture etc) and I would suggest that we need an education to open access. We should praise more open access projects rather than the new designs of Apple (hysterical obsession for 6-months pseudo new products) and Google (do we really need all this insistence on Google glasses? Another screen just in front or our eyes?).
What I’m trying to say is that maybe even in the digital world we need to buy less and share more in common owned virtual spaces.

These same issues are very clearly discussed also by Cohen and Rosenzweig: “open source should the slogan of academic and popular historians”. In other words, this is a strong call for more digital preservation. This book was published in 2005. Did something change in the meanwhile? In the last weeks we talked about the decreasing importance of digitalization projects and a growing insistence on “more complex” DH projects. Why is that? Is it because since 2005 the digitalization process reached an “acceptable” level, or because it has acquired a lower status in the academy, and in the emerging DH field, compared to more analytical DH projects?

Finally, here a wonderful example, according to me, of digital pedagogy in history and many other disciplines: https://www.youtube.com/watch?v=Yocja_N5s1I&list=PLBDA2E52FB1EF80C9

Best wishes,

Davide

Fashion History meets Digital Humanities

Hey guys,

I’m uploading a post from the blog Historic Dress that reflects on the relationship can exist between Fashion and Digital Humanities. It’s about a contribution that Historic Dress co-conspirator Jon Berndt Olsen gave to University of Massachusetts Amherst for the new First-Year seminar, “From Cotton to Kevlar: Fashion History meets Digital Humanities.”
I needed to post this because sometime I lose track of how DH can me interesting in every field of reality, especially mine. What’s DH’s impact in areas where it is not immediately connected?

The aim of this reading is  to try out some ideas that animate Historic Dress while also introducing students to digital approaches to humanities research and interpretation.

From Cotton to Kevlar: Fashion History Meets Digital Humanities

 

Thank you all,

Nico