Could not complete log in. Possible causes and solutions are:
Cookies are not set, which might happen if you've never visited this website before.
Please open https://media.dlib.indiana.edu/ in a new window, then come back and refresh this page.
An ad blocker is preventing successful login.
Please disable ad blockers for this site then refresh this page.
How widespread is social bookmarking? Or do people just google to find something they've found before? In November/December 2008 we conducted a survey of re-finding behavior in the IU community. This talk presents the results of the survey, highlighting differences between students and other groups of users. Technology use surveyed went beyond bookmarking to include personalization of OneStart or SecondLife and subscription to listservs and RSS feeds. These findings should be of interest to anyone creating websites in the IU community.
Sakai 3 will eventually mean a major change in IU's Oncourse. This talk will provide an overview of Sakai 3 current plans and directions. A demo will illustrate some of the major changes coming in Sakai, including a new emphasis on content authoring and social media. There will also be an opportunity to discuss the implications of Sakai 3 for libraries.
Since 2001, the Workshop in Political Theory and Policy Analysis and the Indiana University Digital Library Program have partnered to create the Digital Library of the Commons (DLC) - a gateway to the international literature on the commons. The DLC provides free and open access to full-text articles, papers, and dissertations. In this talk, we will discuss the new infrastructure that has been developed to preserve and make available the world's only dedicated collection on the study of the commons.
The Text Encoding Initiative Guidelines for Electronic Text Encoding and Interchange (TEI), first published in 1994, quickly became the standard for encoding literary texts. The TEI was widely adopted by libraries for its promise of discoverability, interoperability, and preservation of electronic texts, but the TEI's monolithic nature inspired the codification of library-specific practice. Since 1999, libraries have relied on the Best Practices for TEI in Libraries (http://purl.oclc.org/NET/teiinlibraries) to guide their work with encoded texts. In April 2008, the TEI in Libraries special interest group (SIG) and the DLF-sponsored TEI Task Force partnered to update the Best Practices. The revision was prompted by the release of P5, the newest version of the TEI, and the desire to create a true library-centric customization.
The revised Best Practices contain updated versions of the widely adopted encoding 'levels' - from fully automated conversion to content analysis and scholarly encoding. They also contain a substantially revised section on the TEI Header, designed to support interoperability between text collections and the use of complementary metadata schemas such as MARC and MODS. The new Best Practices also reflect an organizational shift. Originally authored by the DLF-sponsored TEI Task Force, the current revision work is a partnership between members of the Task Force and the TEI Libraries SIG, with the SIG taking the lead. As a result of this partnership, responsibility for the Best Practices will migrate to the SIG, allowing closer work with the TEI Consortium as a whole, and a stronger basis for advocating for the needs of libraries in future TEI releases.
If you work with encoded texts or simply want to learn more, please join me for the Best Practices for TEI in Libraries brown bag session. I will discuss the: motivations governing encoding in the context of libraries; historical context for the development of the Best Practices; and TEI Header and encoding levels recommendations.
Historically, libraries‰ÛÓ especially academic libraries‰ÛÓhave contributed to the development of the TEI Guidelines, largely in response to mandates to provide access to and preserve electronic texts. The institutions leveraged standards such as the TEI Guidelines and traditional library expertise‰ÛÓauthority control, subject analysis, and bibliographic description‰ÛÓto positively impact publishing and academic research. But the advent of mass digitization efforts involving scanning of pages called into question such a role for libraries in text encoding. Still, with the rise of library involvement in digital humanities initiatives and renewed interest in supporting text analysis, it is unclear how these events relates to the evolution of text encoding projects in libraries.
This paper presents the results of a survey of library employees to learn more about text encoding practices and to gauge current attitudes toward text encoding. The survey asked such questions as:
As library services evolve to promote varied modes of scholarly communications and accompanying services, and digital library initiatives become more widespread and increasingly decentralized, how is text encoding situated in these new or expanding areas?
Do we see trends in uptake or downsizing of text encoding initiatives in smaller or larger academic institutions? How does administrative support or lack thereof impact the level of interest and engagement in TEI-based projects across the library as whole?
What is the nature of library-led or -partnered electronic text projects, and is there an increase or decrease in local mass digitization or scholarly encoding initiatives?
Preliminary analysis shows, despite assumptions of decline, that over 80% of eligible respondents are actively engaged in text encoding projects, and many others are planning to embark on a new project. The presentation will unveil a full analysis.
As interest in digital scholarship becomes more widespread, and more units within the Libraries and across the IU campus become interested in developing digital collections, the DLP is re-evaluating how it can best support the growing needs of the university and the wider community, and adjust to developing expectations. In this talk, the Associate Director for Digital Library Content and Services will briefly discuss the history of digital collections development, both at IU and beyond, before examining the new paradigm for digital scholarship and describing the DLP's plan for supporting that new paradigm.
The RDF model underlying Semantic Web technologies is frequently described as the future of structured metadata. Its adoption in libraries has been slow, however. This is due in no small part to fundamental differences in the modeling approach that RDF takes, representing a "bottom up" architecture where a description is distributed and can be made up of any features deemed necessary, whereas the record-centric approach taken by libraries tends to be more "top down" relying on prespecified feature sets that all should strive to make the best use of. This presentation will delve deeply into the differences between these two approaches to explore why the RDF approach has proven difficult for libraries, look at some RDF-based initiatives that are happening in libraries and how they are allowing different uses of this metadata than was previously possible, and pose some questions about how libraries might best make use of RDF technologies
As part of an exhibition at the Lilly Library entitled The Globalization of the United States, 1789-1861 scheduled to open September 15, historian Konstantin Dierks and librarians Erika Dowell and Michelle Dalmau have partnered to create a digital counterpart to the physical exhibit that includes an interactive, map-based visualization. The visualization tracks several data points or ‰ÛÃfacets‰Û about U.S. interventions in the rest of the globe, from diplomatic missions to stationed military squadrons. As Dierks describes, it provides a tool for scholars and students to investigate how ‰ÛÃthe United States, no longer swaddled within the British empire, sought to recalibrate its interaction with the wider world as an independent nation.‰ÛÂ
This presentation will focus primarily on one component of the digital exhibit, the map-based visualizations, and how we in the libraries have been able to use this project as a use case for generalizing research-oriented treatment of geospatial and temporal data. By abstracting the data gathering and mapping processes and building workflows to support these activities, we have the beginnings of a services-oriented approach to map-based discovery and inquiry that could be leveraged by other digital research projects at Indiana University. As part of this presentation we will: a) evaluate the various map-based tools with which we experimented including SIMILE Exhibit, Google Fusion, Neatline, and Leaflet, b) review the metadata challenges particular to this project and how they can be abstracted for future projects, and c) relay lessons learned when working with historical maps. We will conclude by proposing a model established by Professor Dierk's project team, using a combination of tools and techniques referenced above, as a way forward in supporting map-based digital research projects more generally.
Isaac Newton is an iconic figure in the history of science but he had a mysterious side that remained hidden and unknown until the 1930s---he wrote more than 125 manuscripts on alchemy, comprising over 2300 pages and a million words. Alchemy was equated with sorcery and charlatanism over many centuries, so the academic world was surprised to learn of Newton's consuming interest and tended to ignore it until very recently. Our project is creating a scholarly online edition of Newton's alchemical manuscripts. The goal is to provide accurate transcriptions of the originals in TEI/XML documents with Unicode encodings. Alchemists used large numbers of special symbols in their cryptic literature to stand for substances, principles, processes, and devices. Newton's use of those symbols presented us with many unexpected challenges. We'll discuss those challenges and describe our use of font editors and symbol generators and XSL to create and serve the symbols, our use of XTF to make them searchable, and our work with the Unicode Consortium to create a new block of code points devoted to these historical alchemical symbols.
This talk reports on a survey conducted during Spring 2010, with responses
collected from over 100 people. The survey focused the current practices and
future plans of academic libraries regarding video streaming. In addition to
summarizing results of the survey, we will provide an overview of video
streaming plans at IU, with a focus on activities in support of Action 37 in
the Empowering People strategic plan.
‰ÛÜLibrary publishers often straddle the line between journal publisher and journal host, which presents challenges for ensuring that journals meet certain standards for quality and transparency. At Indiana University, we conducted a self-evaluation to determine whether our library-published open access journals were following best practices for scholarly journals. This presentation will discuss the methods and criteria used, and how we developed new tools and approaches to educating journal editors based on our findings.
The African Studies Program, Indiana University, Bloomington and the Indiana Consortium for International Programs
Summary:
Presented by Kwamina Panford, Department of African-American Studies, Northeastern University. Presented at Symposium: "A Contested Resource: Oil in Africa".
Metadata creations tools are in general surprisingly hard to use, often lacking basic features such as auto-complete, spell-check, and data validation. In addition, there is little research on how they could be made better and more efficient. The ideal metadata creation tool would make it easy to create robust, consistent metadata, and difficult to make a mistake. This presentation will examine research on data entry interfaces, look at the state of the art in metadata creation tools, and demonstrate some features that make metadata creation tools work well.
The African Studies Program, Indiana University, Bloomington and the Indiana Consortium for International Programs
Summary:
Welcome by Samuel Obeng, Director, African Studies Program, Indiana University Bloomington and Karen DeGrange, Executive Director, Indiana Consortium for International Programs. Presented at Symposium: "A Contested Resource: Oil in Africa".
The Troubadour Melodies Database is a Drupal-platform site that includes basic information about and transcriptions of the extant troubadour melodies as they are found in the 13th-14th century manuscripts preserving the tradition. The melodies are encoded using alpha-numeric strings designed for the font Volpiano, developed by David Hiley and Fabian Weber. The site gives basic information on the manuscripts and troubadours themselves as well as tables showing concordances and totals of melodies by troubadour, manuscript, genre, and catalog number. In addition to gathering the melodies and information about the corpus in one place, the database also provides the ability to search the melodies using a search tool based on Jan KolÌÀÂek's original Melody Search Tool, designed for his own chant database, which allows for three searches (beginning, anywhere, and end) of the melodies in the database. Further, having the melodies encoded has allowed for analysis and comparison of the melodies in terms of their characteristics using tools like AntConc to generate concordances, find collocates, etc. Modification of the Melody Search Tool's PHP script has also allowed the generation of intervallic profiles of the melodies, creating further opportunities for analysis for any melodies encoded in Volpiano.
This brief video presentation outlines the rationale for the "For Students, By Students" Exam Review Assignment, explains the process, and provides examples of student work.
2021 IAH Annual Business Meeting
President’s Report
Amendment of the IAH By-Laws
Election of New IAH Board Members and Officers
Awards Ceremony
Bennett-Tinsley Award for Undergraduate History Research and Writing
Walter K. Nugent Best Graduate Student Paper Award
James H. Madison Best Indiana Magazine of History Article (2020) Award
This video remediates some of the interactive features of the Shining Lights website. It includes a walkthrough of some of the most interactive and visually interesting pages on the website.
3JUP10ORP: Animation of the evolution of midplane and meridional densities in logarithmic scale for the 3JUP10ORP simulation. The axes have units of AU and the time is given in ORPs in the upper right of each panel. The series starts at t = 10.54 ORPs and proceeds to the end of the simulation at approximately 21 ORPs. The black diamond in each of the panels indicates the location of the planet.
1JUP10ORP: Animation of the evolution of midplane and meridional densities in logarithmic scale for the 1JUP10ORP simulation. The axes have units of AU and the time is given in ORPs in the upper right of each panel. The series starts at t = 10.54 ORPs and proceeds to the end of the simulation at approximately 21 ORPs. The black diamond in each of the panels indicates the location of the planet.
For more than sixty years, since Nagasaki in 1945, no nuclear weapons have been exploded in anger, despite several wars in which one side possessed nuclear weapons. The taboo is an asset to be preserved. New nuclear weapon states should recognize that the weapons have proven useful for deterrence; any other use will almost certainly bring universal opprobrium. Certain responsibilities will accrue to any new nuclear-weapon states: security against accident, sabotage, or unauthorized use. The United States was slow to recognize the need for such security, as it was slow to recognize the crucial importance of designing weapons safe from attack. Perhaps China, a mature nuclear-weapon state, is in a strong position to provide guidance to any nations contemplating nuclear weapons.
From Homer's Iliad to Milton's Paradise Lost to the Russian novel and contemporary global literature, major works, read from the angle of the practice of statecraft, offer insights on leadership, substance, and the structure of world affairs reaching beyond the methodologies of international security and policy studies. The meaning and prospects for "The Arab Spring" and the European Union may be more clearly discerned when viewed through this humanities-focused lens.
The uncertainties are great but so are the certainties. Venus and Mars show what too much, or too little, greenhouse gas can do to the possibility of life. Carbon dioxides ability to absorb infra-red radiation can be measured. But the analysis of how much warming, what changes in climate, what impact on agriculture, on health and comfort, how well different countries can adapt, is still in progress; too little is known to predict what concentration will be too much, so no global rationing scheme is likely. Certainly the worst climate impacts will be on the rural poor; economic development is an important defense. An ambitious program of research, development, and exploration for new economical energy sources and locations for carbon sequestration is urgently needed. The already-developed nations will have to provide financial and technological assistance to the less developed.
Mass usage of the Internet is in its second decade, and Professor O'Donnell's Avatars of the Word, a study of the place of media in cultural history, is just ten years old. What have we learned, what haven't we learned, and especially: what sense do we make of the scale and speed of change for our most traditional ways of building and preserving culture?
This presentation discusses likely futures in a world where governments make their own rules about emissions reduction. It examines whether there is a future for the UNFCCC, and looks at the role of the IPCC in supporting the work of the UNFCCC.
Discusses the importance of understanding nonverbal communication in therapeutic recreation. Describes types of nonverbal communication: visual cues, vocal cues, spatial cues, and temporal cues. Includes vignettes to illustrate each type of communication. The vignettes provide an opportunity for class discussion and are followed by panelists' summaries of nonverbal communication they saw in the vignette. Includes an initial list of objectives and a closing summary.
Outlines the steps important for safely transporting individuals with disabilities. The basics of infant and child restraint systems and lift operations are also discussed.
Describes the uses of recreation therapy in various settings and in fields such as oncology, pediatrics, drug rehabilitation, psychiatry, gerontology, and physical therapy. Details the processes of recreation therapy.
Do you know what the term ‰ÛÃAccessibility‰Û means? Do you want to learn how people with disabilities access digital content? Come learn about some of the most common hurdles and barriers that people with disabilities face when accessing digital content. The Assistive Technology and Accessibility Centers (ATAC) Staff will discuss accessibility best practices, standards, and give a short live demonstration of assistive technology. No prior experience with accessibility is necessary and we welcome all questions about accessibility, disabilities, and assistive technology.
The Medieval Electronic Scholarly Alliance (MESA) is a federated
international community of scholars, project, institutions, and
organizations engaged in digital scholarship within the field of
medieval studies. Funded by the Andrew W. Mellon Foundation, MESA
seeks both to provide a community for those engaged in digital
medieval studies and to meet emerging needs of this community,
including making recommendations on technological and scholarly
standards for electronic scholarship, the aggregation of data, and the
ability to discover and repurpose this data.
This presentation will focus on the discovery aspect of MESA, and how
it might serve the non-digital medievalist who may nevertheless be
interested in finding and using digital resources. Starting with a
history of medievalists and their interactions with digital technology
as told through three data sets (the International Congress on
Medieval Studies (first held in 1962), arts-humanities.net (a digital
project database in the UK, sponsored by JISC and the Arts &
Humanities Research Council), and two surveys, from 2002 and 2011, that looked specifically at medievalists' use of digital resources), I will draw out some potential issues that this history has for the current developers of digital resources for medievalists, and
investigate how MESA might serve to address these issues.
IUB Libraries' subscription to Shared Shelf, an image and media management software for hosting and cataloging locally owned images, is an exciting development for visual disciplines--but many decisions need to be made in order to effective deploy this tool. This presentation will discuss the collaborations and processes implemented for establishing sustainable policies and workflows. Our ultimate goal is to promote and integrate interdisciplinary image use, and we will discuss our backwards-design approach, including implementing a pilot project, evaluating legacy data, and establishing partnerships to reach out to faculty across campus.
Folk songs have been at the heart of the study of folklore since its beginnings, and the scholarship on song is one of the finest achievements of the field. But in recent years interest in songs, especially songs in English, has waned among scholars in both folklore and ethnomusicology. Despite some continuing important and innovative work, and public fascination with the subject, song no longer seems central to folklore studies. I will argue that song is a cultural universal, indeed a cultural imperative, and exists as a system similar to kinship systems, language, and economic relations. This will be a plea to resume interest in songs, and will suggest some means by which folklore studies might again assume responsibility for understanding the role of song in human history. (Sponsored by the AFS Fellows.)
Here, AFS continues the custom of including a public interview with a senior member of our field at the annual meeting. In this session, Robert Baron and Ana Cara will interview John Szwed, professor of music and jazz studies and director of the Center for Jazz Studies at Columbia University, about his life and work. (Sponsored by the AFS Oral History Project.)
This lecture presents results of a project on folk medicine among Latinx in Los Angeles in which 131 interviews were conducted with 49 individuals, more than half of whom were healers associated with botánicas. Contrary to a number of previous reports, research data reveal that the healers were not poorly educated, unsophisticated, or adversaries of biomedical care; that clientele were not exclusively Latinx; and that a number of long-standing assumptions in works on Latinx healing traditions should be reassessed. The present study of ethnomedical treatment offers insight into needs and concerns that could inform the healthcare profession in regard to one of the largest and most underserved populations in the US.
Some songs pertaining to the “música tropical” genre, or music exhibiting tropical rhythms from both the Pacific and Atlantic coasts, feature Afro-mestizo protagonists in their lyrics. My study explores the imaginaries constructing the subjectivities of Afro-mestizo men and women and posits that these gender constructions are different between the two sexes. Men tend to be depicted more harshly than women. Both, however, are depicted in a stereotypical and racist manner. My study incorporates feminist and critical race theories as well as postcolonial theories in the analy- sis and hermeneutics of the representation of Afro-mestizos in the lyrics of these songs.
Casey, Michael, Mobley, Robert, Figurelli, Daniel, Dunn, Jon
Summary:
Audio-Video Preservation Services (AVPS) is a new department in the Library Technologies division of IU Libraries, offering services to IU units that hold archival audio and/or video recordings. It is staffed by veterans of the now-completed Media Digitization and Preservation Initiative and offers expert preservation-quality digitization of a number of audio and video formats. In addition, AVPS is able to assist with audio and video collection management issues such as prioritization, selecting vendors, developing grant proposals, and quality control, among others. This presentation will explore AVPS origins, current objectives, and technical capabilities, with presentations by Mike Casey, Rob Mobley, Dan Figurelli, and Jon Dunn.
The Indiana Geological and Water Survey (IGWS) is a research institute of Indiana University and a state agency whose roots originate with the first survey of earth science resources of the state of Indiana in 1837 by David Dale Owen, our first State Geologist. The IGWS is responsible for maintaining the state’s geological “cabinet,” or collections of physical samples, maps, and data that are increasingly born-digital. There are now more than 2 million individual items in the IGWS sample collections and digital holdings of both digitized and born-digital data from oil, gas, and water wells; historic photograph collections; springs and caves; hydrologic and atmospheric data; and much more. To better maintain these collections, the IGWS created CARST—the CollectiveAccess ResourceSpace Tandem—an environment that integrates both a collections management system and a digital asset management system. These free, open-source software platforms are combined with metadata schemas, controlled vocabularies, and persistent, globally unique identifiers to facilitate long-term archival and discoverability of these important assets. This environment also easily integrates with data aggregators to benefit regional and nationwide studies on the availability of critical minerals and rare earth elements, the quantity and quality of water resources, and potential threats from geologic hazards like earthquakes and landslides.
Using inclusive vocabularies, defined here as those vocabularies representative of and created by historically marginalized communities, is helpful for providing options when creating original description. Is it possible to also supply these vocabularies for use as a navigation aid into a system that only makes use of commonly used controlled vocabularies that are not as inclusive or representative, such as Library of Congress Subject Headings (LCSH)? This talk will explore progress in learning about the scope of available controlled vocabularies and classification schemes from marginalized communities and a proof-of-concept project to engage such a vocabulary as an information retrieval aid for search and discovery against a system using only LCSH for subjects (a typical library catalog).
The two academic disciplines linguistics and literary studies are often part of one common study program, but they differ in many respects: Their object of study, the methods they use, the type of knowledge they aim to generate, and also the presentation of their work in academic writing. I explore these differences by examining a corpus of German PhD theses from the two disciplines.
The focus of this talk will be twofold: First, I will discuss how we can identify differences between corpora in a data-driven way, i. e. with only few theoretical assumptions. While many data-driven approaches rely on surface-based frequencies of words and sequences of words, I argue for the additional use of syntactic annotations for this purpose. Second, I will present and contextualize the differences between academic texts in linguistics and literary studies that can be detected in this way. I conclude by reflecting more generally on how the results of a data-driven analysis can be integrated into existing theories.
For some undergraduate students, it can be increasingly difficult to distinguish fact from fiction in an online environment. On top of this, students can be so overwhelmed by the massive amount of information that they have problems finding and identifying accurate information for their research. Enter the Critical Thinking Online Toolkit.
As a series of assignments and modules in Canvas, this Toolkit provides materials for instructors across all IU campuses to help students hone their information literacy skills: identify and evaluate valid sources of information, synthesize that information, and construct and communicate knowledge for their academic work and everyday lives.
Come learn about more about the Toolkit: what it is, where it is, and how it’s helping instructors across IU campuses engage their students to navigate and critically assess information in an online environment.
This presentation is a step toward understanding the problem of bias in metadata and how that impacts inclusivity in the research process. Original description provided for digital collection discovery and access as well as controlled vocabularies commonly used for subject headings (such as Library of Congress Subject Headings) have inherent biases which present challenges for researchers discovering and engaging with these collections, particularly researchers from underrepresented or historically marginalized populations. We’ll review work in this area to date and discuss possible approaches for where to go next to improve description and the academic research experience.
Even as JATS XML has become the standard format for academic publishing, the challenges involved in implementing a JATS XML-based publishing workflow have prevented many library publishers from moving beyond PDF-based publishing. The complicated apparatus of even the most basic scholarly articles complicates XML production considerably. In addition, most existing workflows are reliant on XML conversion tools or paid vendors to convert author submission documents into JATS XML. In either case, these XML documents are time-consuming to produce and often require additional editing and correction before publication. Book reviews, on the other hand, provide a less burdensome format for library publishers who wish to transition to XML publishing. With minimal training, editorial teams can format JATS XML book reviews in-house without resorting to paid vendors or conversion tools. This presentation outlines the successful onboarding of a JATS-only book review journal to the Open Journal Systems platform. To facilitate this, we created a simplified JATS XML template using the DAR tag subset specification to optimize machine readability, avoid redundancy, and ensure reusability. The onboarding process also required customization of the OJS interface and the creation of detailed documentation and training materials for the editorial team. Although the editorial team had no prior experience with OJS or JATS XML, they are now publishing full-text, machine-readable books reviews. As the result of our work, these book reviews will now be more easily indexed and permanently stored as markup in a digital preservation archive. The semantically tagged content will facilitate keyword searches and increase discoverability over the long term. Finally, as a machine-readable format, JATS XML is inherently accessible and includes elements that allow for accessibility tagging and for the creation of interfaces that are both Section 508 and WCAG compliant.
While the JATS XML format is widely used in scholarly publishing, many library publishers have been slow to implement this standard in their article production workflows. Due to the challenges involved in converting, editing, and rendering conventional article submission files into full-text XML galleys, library publishers often lack the resources and experience to adopt JATS as a publishing format. The complicated apparatus of even the most basic scholarly articles, such as abstracts, images, graphs, footnotes, and references, complicate XML production considerably. Book reviews, however, provide a less complex format for library publishers who wish to gain experience publishing in XML. Drawing on a recent experience onboarding an online book review journal to the Open Journal Systems platform, this presentation offers a practical guide to developing a JATS publishing workflow that is accessible for both library publishers and editorial teams with minimal prior knowledge of XML.
Graney, Brian , Lucaites, John Louis, Lynch, Shola
Summary:
Regeneration in Digital Contexts: Early Black Film, organized by the Black Film Center/Archive, brought together an interdisciplinary group of scholars, moving image archivists, and technology specialists in digital humanities for a two-day conference and workshop held in Bloomington, Indiana November 15-16, 2013.
The conference program, November 15, 2013, opened with a Welcome by Brian Graney, Archivist and Head of Public and Technology Services at the Black Film Center/Archive. The Introduction was provided by Associate Dean for Arts & Humanities, with responsibility for Undergraduate Education John Lucaites. Shola Lynch, curator of the Moving Image and Recorded Sound Archive at the Schomburg Center for Research in Black Culture, presented the Keynote Address
"Film, Race & Archives — The Odyssey of a Film Maker & Curator".
Inventories are one of the most useful types of documents available to book historians. They are essentially lists of person or organization’s goods, but these seemingly simple lists contain a wealth of data and information. For a private individual, an inventory of their household goods can point towards their wealth and status in society while an inventory of their books allows us to analyze their book ownership habits and potential reading. For members of the book trade, inventories of their businesses can tell us about the size and characteristics of their business, the typical tasks they performed, as well as what types of books they produced or sold. Aggregations of these book inventories help us understand the production, sale, ownership and reading of books in a given geographic and temporal space as a whole. However, the data found in inventories of early modern private libraries, booksellers, and printers are usually published by book historians as simple transcriptions of the documents (sometimes with metadata identifying the book described in each entry) in print or in online journals as PDFs. Whether in print or in PDF, this static presentation of inventory data makes it difficult for book historians to browse, search, aggregate, compare, and build upon each other’s data.
As part of my doctoral work investigating bookselling and private libraries in early modern Navarre, Spain, I am using TEI-XML and the open-source database builder Heurist to address these issues of dissemination, interoperability, and sustainability for book inventory data and to improve my overall process for conducting historical research. In this presentation, I will outline my current workflow for moving from historical documents in the archives to a final dataset. I will discuss my use of TEI inside and outside of the archive and the development of my Heurist database, Libros en Navarra | Books in Navarre (LN|BN), which stores data for private library and bookseller inventories documenting what books were present in Navarre during the 16th and 17thcenturies. I seek to show how these methods of digital scholarship provide a base which facilitates not only my research but hopefully the research of other book historians who in the future may wish to incorporate and transform my data in their own work.
Since the presidency of Franklin Roosevelt, the first 100 days of an administration has been used as a measuring stick to estimate the ability of a new president to govern. This is still true today. The first 100 days of the Trump administration invoked strong sentiment both for and against his policies. However, was the sentiment generally positive or negative or neutral? Using different sentiment analysis algorithms and Trump’s favorite social media platform, Twitter, we scraped over 181,000 English language tweets between January 20th, 2017 and April 29th, 2017 to get an idea of Twitter user sentiment regarding the new Commander-in-Chief during his first 100 days.
While our results reveal an interesting snapshot of the heightened emotions of the first 100 days of this presidency, they also raised some concerns regarding the bias inherent in the sentiment analysis process. More specifically, in the different dictionaries used to determine which words are “positive” and which words are “negative" issues of bias regarding race, gender, sexuality, and religion emerge. Therefore, it's important to "look underneath the hood," even when using a vetted dictionary, to examine the assumptions made, tweak the dictionary, and make transparent any assumptions left in the lexicon. We have parsed a further 16K tweets from the weekend of the Charlottesville protests to show what happens both before and after dictionary is tailored to an event focused on issues that are source of bias.
This presentation is the first step in an answer to Emily Drabinski’s 2013 challenge to library and information science (LIS) professionals to think about ways in which to ‘queer the library catalog,’ and to represent identity as historically constructed and described. Beginning with a brief outline of the troubled history between marginalized groups and LIS classifications, I examine some of the proposals suggested over the past half-decade by researchers—and their limitations. Instead of starting anew, or using ‘uninformed’ social tagging/folksonomies, I propose a ‘turn’ to the catalogs and controlled vocabularies of archives and special collections, which frequently reckon with unclassifiable material.
Following through on that turn, I will discuss how linked data and linked data vocabularies are currently being used by several digital archives—along with some possible lessons for the LIS field as a whole. The radical and subversive use of linked data by queer digital archives offers a partial solution to the conundrum of minoritized and historical representation in the catalog. Finally, I will conclude by describing my own experiences and considerations in the construction of a new linked data vocabulary.
This study examines the South Korean cyberfeminist community Womad, a community currently under fire in South Korea due to its exclusionary politics, i.e. its antagonism towards anyone (biologically) male and (ethnically) Korean. In this project, The study reveals how Womad’s “medium specificity” (i.e. its platform) interacts with South Korea’s “national specificity” in specific ways that results in the “technological assemblage” that is Womad. By amassing individual postings through data crawling, the project uses computerized text analysis to (1) reveal the technical difficulties underlying textual analysis of Korean online communities in general (2) examine the kinds of topics that this specific community engages with (3) investigate which topics are most popular/unpopular within the community (thus accruing more upvotes/downvotes).
The Vietnam War: Stories from All Sides began as an oral history project telling stories from American & Vietnamese veterans, refugees and others impacted by the war. Ron Osgood initiated the project through an Institute for Digital Arts & Humanities (IDAH) Fellowship and a New Frontiers Grant. Over the past 10 years, more than 150 oral history interviews have been recorded, a prototype website for educational use created and a documentary film produced.
Currently, Osgood is working with Jon Cameron to upload the 150 oral history interviews to the IU Libraries’ Media Collections Online service. In addition to providing long-term preservation for the media and an institutionally owned platform for hosting, website embeds are also being migrated from YouTube to Media Collections Online.
Audiovisual archivists agree that media holdings must be transferred to the digital domain as soon as possible in order to survive. Because this work requires significant resources, it must be conducted as efficiently as possible. One place to realize efficiencies is in the management of the digitization process. This presentation will explore managing effective and efficient 1:1 as well as parallel transfer media digitization workflows. Using the Indiana University Media Digitization and Preservation Initiative project as a case study, Mike Casey will discuss applying the theory of constraints and adapting software development methodologies to efficiently manage 1:1 digitization workflows. This will include a look at working with bottlenecks, scrum methodology, and the daily standup. Andrew Dapuzzo from Memnon Archiving Services will address issues in regulating parallel transfer workflows including the role of workflow management software, the importance of both human and machine quality assurance in each step of the process, the difficulty in maintaining obsolete machines, overall system design and Total Quality Management. The more efficient the digitization workflow, the more we are able to preserve with scarce resources.
As webserve has done away with Digital Media and encourages folks to put video on Kaltura, I have developed a plugin that allows you to playback Kaltura video in Omeka. This plugin, which will also allow the playback of video from Youtube as well, will allow those who need to use video in Omeka to use Kaltura to store their videos and playback in an Omeka site. I will demo the setup and use of this plugin in Omeka.
This interactive workshop will consider how Open Educational Resources (OER) can alleviate the high cost Indiana University Bloomington undergraduate students pay for course materials (an estimated $1,034 each academic year). Data suggests that students will forgo purchasing expensive course materials, even when they know it will impact their success in the classroom. This session will introduce OER and discuss its benefits, critically think about challenges to OER adoption, and formulate strategies to support IU instructors in finding high-quality OER, adapting them to fit students’ needs, and creating (even in collaboration with students) customized course materials. Please bring a laptop or similar device.
An exciting development for audio and video repositories is the emerging IIIF standard for time-based media. Join us to understand what IIIF is and why the Avalon project is collaborating with the IIIF-AV community. We will also discuss how we see the future of these two important open source projects and their contribution to a rich media viewing experience.
By preserving artifacts held by communities who are often hidden or erased from the dominant historical narrative and contextualizing these artifacts with oral histories, the History Harvest model, set forth by the University of Nebraska-Lincoln (UNL) in 2010 as a form of public humanities engagement, amplifies voices that would otherwise not be heard. A public history endeavor at Indiana University Bloomington offers a case study in which we reconsider History Harvest fundamentals, from technology needs to workflows, with human labor at the center, following minimal computing approaches. Our adaptations of the original model, eleven years after UNL’s launch of the History Harvest, are guided by the same principles of engagement, replication, and autonomy for the community members, students, and scholars alike, all of whom contribute to the telling of stories. In consultation with minimal computing “thought pieces” and related literature, we are working towards an approachable model, both in computer and human terms, for History Harvests. Our presentation will explore the human and technological aspects of minimal computing in the context of History Harvests, with a focus on how to scaffold limited resources like funding, lightweight technology and workflows, and properly support and acknowledge the limitless contributions of the cross-section of people involved in History Harvests.
In recent years, concern over the longevity of physical audio and video (AV) formats due to media degradation and obsolescence, combined with decreasing cost of digital storage, have led libraries and archives to embark on projects to digitize recordings for purposes of long-term preservation and improved access. IU's Media Digitization and Preservation Initiative (MDPI) is one of those projects, estimated to digitize 325,000 audio and video items and 25,000 film items overall. Beyond digitization, in order to facilitate discovery and research use, AV materials must also be described, but many items and collections lack sufficient metadata. Join us to learn more about a planned project at IU with experts from University of Texas at Austin School of Information and AVP to create a technology platform and workflow to support the incremental application of both automated and human-based processes to create and augment metadata at large scales for AV collections. The project is called AMP: Audiovisual Metadata Platform.
Academic libraries and archives are dealing with increasing numbers of digital audio and video (AV) files, acquired through both digitization of analog collections and acquisition of born-digital AV resources. While the emergence of low-cost storage options and maturity of streaming platforms has made it easier to store and deliver AV, these collections often lack metadata needed in order to make them discoverable and usable by researchers and other users. Since late 2018, the Indiana University Libraries have been working with partners at the University of Texas at Austin, New York Public Library, and digital consultant AVP to develop an open source software platform, known as AMP (Audiovisual Metadata Platform), that leverages automated machine learning-based tools together with human expertise to build workflows to create and augment metadata for AV resources to improve discovery, rights determination, and use. We will present an update on progress of the AMP project and its successes and challenges to date, including a demonstration of the AMP system and discussion of issues in system design, workflows, and the use of open source and commercial cloud-based machine learning tools. We will also discuss results to date of testing the AMP system using collections from the Cook Music Library and University Archives at IU and from the New York Public Library. This work is generously supported by a grant to IU from the Andrew W. Mellon Foundation.
Representation is one of the most powerful impacts that archives can make on communities. Ensuring that all people’s works, lives, and information is being preserved in an archive is what fuels a many modern day archivist. However, establishing equal representation of minorities and underrepresented groups is not enough to create a more inclusive world, archivists must also create ways for people to access that information. The creation of digital libraries and other online resources, allows for more people to use the resources collected, see themselves and their work represented, and gain an understanding of the artists who have come before them. The Ars Femina Archive (AFA), is housed at Indiana University Southeast, and is a collection of music composed by women from before the 1500s to the 1800s. This archive preserves and celebrates the impact that women in history have had on music. Women are largely underrepresented in the arts and especially in music, the AFA allows for people from around the world to research and access this collection of musical compositions created by women. This presentation will focus on the history of the collection, what is contained in the archive, its mission and how that mission is furthered by digitization, and the impact it has on scholarship and performance.
This fall, the IU Libraries is launching two exciting new services: IU DataCORE, for storage and access of IU research data, and Digital Collections, for managing and delivering digitized images, books, newspapers, sheet music, and archival collections . These IU-wide services were conceived as part of the Enterprise Scholarly Systems (ESS) initiative, a partnership between the IU Libraries, IUPUI University Library, and UITS. Both services are built using the Samvera Community’s open source Hyrax repository platform. They represent a new, modern way of managing and proving access to our unique digital collections using software collaboratively developed by several partner institutions including IU. This talk will provide an overview of both services, providing insight into their history, technologies, and plans for the future.
Word embeddings have recently been applied to detect and explore changes in word meaning in historical corpora (Hamilton et al., 2016; Rodda et al., 2017; Hellrich, 2019). While word embeddings are useful in many Natural Language Processing tasks, there are a number of questions that need to be addressed concerning the stability, accuracy and applicability of these methods for historical data. Previous studies mostly made use of exceptionally large corpora such as Google books (Hamilton et al., 2016). However, there is scarce literature on the stability and replicability of these embeddings, especially on small corpora, which are common in historical work. It also remains unclear whether methods used to evaluate embeddings in contemporary data can be used for historical data sets.
In the work presented here, we focus on three methodological questions:
How replicable and stable are the results of different word embeddings models?
How do we determine the accuracy of different embedding models on our historical data?
Given the low resource situation, can we find (enough) meaningful words in the embeddings to draw conclusions about semantic change? Do our findings correspond to prior knowledge?
We experimented with a historical corpus of medieval and classical Spanish that is an order of magnitude smaller than those used in previous studies, and obtained word embeddings using three commonly used word embedding models: SGNS (Mikolov et al., 2013), GloVe (Pennington et al., 2014), and SVDPPMI (Levy et al., 2015). We compare the results of different models and the solutions we developed to address the challenges found.