The Law Library acquired a large collection from William S. Hein & Co., Inc. to make all volumes of several collections (like the Federal Register) available in open access to researchers. Preparing these files by adding metadata for easy searching takes a lot of work, so this summer we asked law students and library students from across the country to help become our “crowd” in order to crowdsource metadata for a collection of 542 volumes of U.
I always think of SEO like the dentist—no one really likes it, but you need to do it. Yet, despite my lack of excitement for the topic, this will be at a minimum my second post (here’s the first about the relationship between creating good content and SEO practices. Today I want to dive a little more into often overlooked aspects of the content creation process and overall content maintenance.
Multilingual does not always mean multiple accounts or websites. Increasingly, multilingual content is delivered in an integrated way, with two (or more!) languages delivered on the same website, app, or social media platform. The World Digital Library (WDL) is one example of how multiple languages can be incorporated on single platforms. The WDL is a hub for cultural artifacts that includes books, manuscripts, maps, newspapers, journals, photographs, sound recordings, and films.
Metadata for website content is usually managed as part of the editorial process when documents are created and published with content management systems. There may be another source for this metadata, especially in regulatory agencies: internal databases that reference Web content in support of record keeping processes. These databases may contain public and non-public information that were never meant to be published for public consumption. “Metadata” is not typically how the content is described.
Metadata, tagging, content modeling … they’re not identical concepts, but they’re driven by the same basic principle: when you structure your digital information, it can be more easily searched, reused, connected, shared, and analyzed. If you’re new to structured content, where should you start? Ideally, your metadata strategy will be part of your overall content strategy. In practice, however, a lot depends on your agency’s culture, its technical resources, its existing practices, and the state of your content.
Imagine a world without Web pages, only intelligent, self-assembling chunks of content waiting to respond to your needs. The page is irrelevant, there may be no context beyond what is included in your content. The content has to survive on its own, perform its goals on its own. Originally when creating content, you would take into account the things that surround it on that page; they give it additional context and relevance.
Many forces are converging to strengthen the political, economic and commercial ties that bind the United States, Canada and Mexico. The GSA Office of Citizen Services and Innovative Technologies (OCSIT) has anticipated this drive toward collaboration for decades, building a network of links among the three nations’ Chief Information Officers and other national technology and data experts. Annual OCSIT-sponsored North American Day (NAD) talks have contributed to improved digital services in all three countries.
DigitalGov University has hosted some great events over the last year in partnership with Data.gov, the MobileGov Community and 18F to bring you information on opening data and building APIs. This month we’ve rounded up the events over the past year so that you can see what’s been offered. Use the comments below to offer up suggestions on what else you’d like to see on the schedule.
Data and code are the foundation, building blocks, and cornerstone of government digital services. They are the keys that open the door to a better digital government future and are fundamental in making government more open. No matter who you are or where you work in the federal space, data and code enable your projects to meet real needs. This month we’re featuring articles around the theme of data and code.
Audit. It’s a word that generally has no positive connotations whatsoever. We hear the word audit and we think of tax audits or timesheet audits, etc. The word normally strikes fear or dread in the hearts of most mortals. But it is also a task that all websites will need to perform from time to time, and hopefully after reading today’s column you can view content audits as positive opportunities and not as dreadful chores.
Data.gov is the central clearinghouse for open data from the United States federal government. It also provides access to many local government and non-federal open data resources. But how does this data get on to Data.gov? Data.gov does not host data directly, but rather aggregates metadata about open data resources in one centralized location. In addition to Data.gov’s recent webinars on how Data.gov harvests data, the Data.gov team has created a living resource to explain in further detail how data must be structured to connect to the Data.
In case you missed it: the Data.gov team recently hosted DigitalGov University webinars designed to help agencies and open data advocates better understand how to get data on Data.gov and how to implement the Open Data Policy’s metadata schema updates. These webinars were designed assist government data publishers in making more data discoverable to the American people. You can watch these webinars and check out additional supplemental resources below.
How can you find the top 5 users of your open data? We were recently asked this question on the Open Data listserv, and while this information can be a good measure of success for open data programs, we also figured some of the answers shared would be of interest to the broader community. This blog post seeks to summarize and clarify those answers. What Defines a Top Third-Party Developer?
Content models provide an opportunity for agencies to structure, organize, distribute, and better publish information in multiple forms and on multiple platforms. Federal agencies discussed why content models are important for future-facing content in our What Structured Content Models Can Do For You Webinars in May and June. The point—with good content models, a single piece of Web content becomes an adaptive information asset that can be leveraged anytime, anywhere.
Not sure how to get your datasets into Data.gov? We’ve put together an overview to show you how the process works. Agencies prepare their enterprise data inventories in data.json format and post them on their websites (agency.gov/data.json), pursuant to the Open Data Policy and following the guidance and using the tools available on Project Open Data. Data.gov also offers a tool called inventory.data.gov that can be used to assist agencies in creating their data inventories.
Structured content and open content models can help you create content that is platform-agnostic, format-free, and device-independent. This was the theme of the “What Structured Content Can Do For You: Article Model” webinar last month. [youtube=http://www.youtube.com/watch?v=kG25vyQ5Jps&w=600] Using a content model is less about how you are crafting your message and more about how the internet is going to react to your content or how you can manipulate it, according to Holly Irving from the National Institutes of Health, Russell O’Neill from the General Services Administration, and Logan Powell from U.
Improving the federal government’s ability to deliver digital information anytime, anywhere, on any device—via open content—is a key goal of the Digital Government Strategy. A content management system (CMS) can help your agency move to an open content model, making it easier for people to find, share, use, and re-use your information. The key steps in getting ready to move to a CMS include: Prepare Your Content Choose a CMS Migrate Your Content to a CMS Prepare Your Content Develop a Content Strategy A content strategy defines such things as topics, themes and purpose, and can also play a part in website governance, customer experience, metadata and search engine optimization (SEO).
Structured content refers to the concept of organizing and treating digital content like data. It’s a way of publishing content as modular, discrete pieces of information that are tagged with machine-readable descriptions. Structured content has the potential to transform how people find, understand, share, and use government information. Why Structured Content Matters Most digital content published by the federal government is still found on static HTML Web pages. This unstructured content doesn’t always adapt well to smaller screens, and it’s harder to discover, share, or reuse the information.
Wikipedia says that structured content refers to information that’s been broken down and classified using metadata. It can also refer to information that’s been classified using XML or other standard or proprietary forms of metadata. The National Institutes of Health’s National Cancer Institute (NIH/NCI) move to structured content has allowed them to easily deliver their content anywhere, anytime, and on any device. The Challenge The National Cancer Institute (NCI) has a huge base of information on cancer treatments that is continually updated.