Recently I was able to present at the American Geophysical Union’s annual conference. Not everything I work on is as easily made public, so this was a nice chance to share some of the team’s work and have good discussions with others working similar problems across multiple federal agencies, academia, and the public.

Download (PDF, Unknown)


The theme of the poster session I presented under was:

 Achieving the “R” of the FAIR Data Principles: Reusability Is the Biggest Challenge but the Most Rewarding! 

F.A.I.R. principles refer to the following:

To be Findable:

F1. (meta)data are assigned a globally unique and eternally persistent identifier.
F2. data are described with rich metadata.
F3. (meta)data are registered or indexed in a searchable resource.
F4. metadata specify the data identifier.


A1  (meta)data are retrievable by their identifier using a standardized communications protocol.
A1.1 the protocol is open, free, and universally implementable.
A1.2 the protocol allows for an authentication and authorization procedure, where necessary.
A2 metadata are accessible, even when the data are no longer available.


I1. (meta)data use a formal, accessible, shared, and broadly applicable language for knowledge representation.
I2. (meta)data use vocabularies that follow FAIR principles.
I3. (meta)data include qualified references to other (meta)data.


R1. meta(data) have a plurality of accurate and relevant attributes.
R1.1. (meta)data are released with a clear and accessible data usage license.
R1.2. (meta)data are associated with their provenance.
R1.3. (meta)data meet domain-relevant community standards.

Our poster discussed two different projects we did to improve the F of F.A.I.R, “to be findable” by reusing metadata to create new metadata.

As described in the poster above, one of these focused on and was a failure due to our inability to routinely get an accurate download link across all tens of thousands of the datasets on ingests metadata about datasets from over 80 different NASA data sites. Although the standard required metadata fields on include “download link” many of the URLs in that field are actually to a front-page or a page where one link out of 50 links on the page is the download link. These limitations are minor for humans, but insurmountable if you’re trying to automate web scraping across 80+ sites.

If we had been able to programaticly get direct links to download each file, our plan would have been to download files, run a series of Python profiling script to generate descriptions of files in terms of file type, file size, number of instances, etc. From these metadata, we had the idea that we could create a model to identify which datasets would be better for the general public and which would have been better scientists and other high-end end-users. We wanted to have this ability, because many of the people that land on from google are general public users who are interested simple easy to use NASA data. Most of the datasets on are large or complicated datasets, so those results tend to swamp the small, easy datasets that would actually be useful for the majority of the people who land on

The other was focused on creating new and better keyword tags for the open-source code projects listed on This was a success! You can see the keywords created by the machine-learning model on right now. Links on the poster further describe the model. Anthony Buonomo did the heavy lifting on the STI tagger model.


Categories: code


Leave a Reply

Avatar placeholder

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.