This section is a growing list of common questions and answers to support agencies when implementing the Open Data Policy.
Technology moves much faster than policy ever could. Often when writing policy for technology, agencies are stuck with outdated methods as soon as they publish new policies.
This project is meant to be a living document, so that collaboration in the open data ecosystem is fostered, and the continual update of technology pieces that affect update can happen at a more rapid pace.
Help the United States Government make its Open Data Policy better by collaborating. Please suggest enhancements by editing the content here, or add tools that help anyone open data (See “How can I contribute?” below).
This project constitutes a collaborative work (“open source”). Federal employees and members of the public are encouraged to improve the project by contributing. This can be done in two ways:
Note: You will need to create a free GitHub account if you do not already have one.
Note: All contributors retain the original copyright to their contribution, but by contributing to this project, you grant a world-wide, royalty-free, perpetual, irrevocable, non-exclusive, transferable license to all users under the terms of the license(s) under which this project is distributed.
The project as originally published constitutes a work of the United States Government and is not subject to domestic copyright protection under 17 USC § 105. Subsequent contributions by members of the public, however, retain their original copyright.
In order to better facilitate collaboration, the content of this project is licensed under the Creative Commons 3.0 License, and the underlying source code used to format and display that content is licensed under the MIT License.
Anyone – Federal employees, contractors, developers, the general public – can view and contribute to Project Open Data.
Ultimately? You. While the White House founded and continues to oversee the project, Project Open Data is a collaborative work — commonly known as “open source” — and is supported by the efforts of an entire community. See the “how to contribute” section above to learn more.
At the onset, the General Services Administration is here to provide daily oversight and support, but over time, it is our vision that contributors both inside and outside of government can be empowered to take on additional leadership roles.
Yes! Simply follow the “advanced” instructions above to submit a pull request.
Release cycles vary from repo to repo. See the README file within the repo where you submitted a pull request to see how often code pushes and updates are done.
Agencies’ Information Resource Management (IRM) plans are comprehensive strategic documents for information management that are intended to align with the agency’s Strategic Plan. IRM plans should provide a description of how information resource management activities help accomplish agency missions, and ensure that related management decisions are integrated with organizational planning, budget, procurement, financial management, human resources management, and program decisions.
In 2012, OMB established PortfolioStat accountability sessions to engage directly with agency leadership to assess the maturity and effectiveness of current IT management practices and address management opportunities and challenges. As part of the annual PortfolioStat process, agencies must update their IRM Strategic Plans to describe how they are meeting new and existing information life cycle management requirements. Specifically, agencies must describe how they have institutionalized and operationalized the interoperability and openness requirements in this Memorandum into their core processes across all applicable agency programs and stakeholders. The FY13 OMB PortfolioStat Guidance was issued on March 27, 2013.
While ISO 32000 is an open standard, the Portable Document Format (PDF) does not achieve the same level of openness as CSV, XML, JSON, and other generic formats.
The common core metadata schema is based on existing vocabularies and easily mapped to NIEM, Information Sharing Environment, and FGDC.
A persistent identifier is a unique label assigned to digital objects or data files that is managed and kept up to date over a defined time period (e.g., Unique Investment Identifiers).
The core metadata schema was the result of recommendations from a government-wide Metadata Working Group at Data.gov combined with research of existing public schemas for data catalogs. Most of the elements trace their roots to the Dublin Core Library.
Submit a pull request for the metadata schema.
Yes, if your data management process includes rich metadata specific to the mission of your agency or the Line of Business in which your agency participates, publishing additional metadata that makes your data more useful to the public is welcomed and encouraged. Note that Data.gov will be harvesting only the metadata in this published schema unless specific arrangements are in place (e.g. geospatial FGDC/ISO).
Each agency is responsible for all data made public.
For general questions about Data.gov, please contact http://www.data.gov/contact-us. For specific information about the mosaic effect, please contact the Data.gov PMO at GSA.
Having the metadata available at the agency level provides agencies with a self-managed publishing capability. In addition, having the metadata in a machine-readable format opens the possibility that major search engines will index these metadata in a manner similar to site maps and allow the public to discover public data across the government using a search tool of their choice.
The agency.gov/open page contains information regarding an agency’s contributions to Open Government, while the /developer and /data pages pertain to APIs and Open Data, respectively. All three pages contribute to an open and transparent government in the United States.
In the near term, Data.gov will continue its current dataset publishing process. As agencies deploy agency.gov/data pages, the publishing process will become a harvesting of metadata from these agency data hubs.
No, the file should be located at the agency.gov/data web space. Each agency should publish their Public Data Listing at agency.gov/data.json.
Data.gov will (when possible) help agencies get started by creating a data.json file for each agency containing the metadata in the correct syntax. The agency will then begin to manage that file for future publishing of datasets.
A wide variety of tools are available to manage a data catalog, whether public-facing or for internal data managements. The records of metadata in the file can be managed by databases, spreadsheets, or even text editors. Data management systems should be able to export the metadata either in the desired format or in one which may be simply mapped with tools.
Having a contact point at the agency who can answer questions and receive comments about published data is extremely important to making your data more open and valuable to the public. This contact point can be centralized at the agency level, but it’s extremely value to have someone close to the source of the data who understands it well enough to help the public take full advantage of it..
Agencies should regularly add to and improve the entries in their data catalog, as well as ensure continuity of access to the data by involving primary users in the changes.
All executive agencies.
Agencies are required to implement the Open Data Policy within six months.
Each of these initiatives has a discreet, targeted focus, but all are aimed at increasing access and use of government data. Data.gov has provided a central place to find data and applications for publicly releasable information. New applications and services to better serve citizens have been produced as a result in the increase of information made available through Data.gov. The DGS/ODP policy establishes a framework to help institutionalize the principles of effective information management at each stage of the information’s life cycle. The framework can help agencies build information systems and processes in a way that increases information and system interoperability, openness, and safeguarding – mutually reinforcing activities that help to promote data discoverability and usability. NIEM, as a government-wide program provides tools to enhance the way many communities build standardized exchanges to increase mission performance. NIEM fully aligns to the DGS/ODP policy and can be seen one of the tools for implementation.
NIEM provides a commonly understood way for various organizations to connect data that improves government decision making for the greater good. By making it possible for organizations to share critical data, NIEM empowers people to make informed decisions that improve efficiency and advance and fulfill organizational missions.
NIEM is not a standard, database, software, or the actual exchange of information. Rather, NIEM provides the community of users, tools, common terminology, governance, methodologies, and support that enables the creation of standards. As a result, organizations can “speak the same language” to quickly and efficiently exchange meaningful data.
There are 14 domains or communities established within NIEM. These are the Biometrics, CBRN (Chemical, Biological, Radiological, Nuclear), Children, Youth, and Family Services, Cyber, Emergency Management, Health, Human Services, Immigration, Infrastructure Protection, Intelligence, International Trade, Justice, Maritime, and Screening Communities.
Additional tools and toolkits can be found at NIEM.gov. Any tools relevant to the NIEM community may also be registered in the NIEM Tools catalog to ensure reuse across the NIEM community at NIEM.gov.
Treating information as a national asset is core to the Open Data Policy and the National Strategy for Information Sharing and Safeguarding. Departments and agencies will need an end-to-end Data Strategy that accommodates both codified in IT governance. Both are aimed at liberating data from the bounds of the application into exposure for unintentional users and uses (as permitted by law and policy). NIEM has become a best-practice implementation of the new National Information Sharing and Safeguarding Strategy, and is fully supportive of the implementation of the Open Data Policy and is positioned to become an early adopter. NIEM provides a common data model, governance, training, tools, technical support services, and an active community.
NIEM adheres to the DGS/ODP Policy. NIEM Communities use open standards such as XML / XSD, and UML to assist in the development of standardized ways of exchanging information across and between government agencies. NIEM is vendor and product neutral. The adoption of the UML profile will allow additional open standards implementations of NIEM based exchanges as supported by community requirements. Additionally, some NIEM Communities submit their NIEM based information exchanges to external standards development organizations to increase industry adoption such as the NIEM Biometrics and NIST, NIEM Radiological / Nuclear and IEC.