University of California
Informatics and GIS Program

IGIS Blog

ESRI Imagery Education Summit

ESRI held its first ‘Imagery Education Summit' in Redlands California this week, and even though I came with high expectations, I was still pleasantly surprised with the caliber of the summit's presentations.  It is very difficult to pick out a favorite among these talks; however, I can say that I particularly enjoyed the presentations by Jarlath O'Neil Dunn from the University of Vermont (pictured below), on ‘Success Stories and Progress' in image analysis and mapping, and by Jason Ur from Harvard University, regarding his work with ‘Drones and Archaeology Case Studies' in Iraq.  The innovative approaches that they and others at the summit presented were truly inspiring!

One take-away from this event is that ESRI is making huge strides to incorporate more remote sensing processing options into ArcGIS Pro's ‘Image Analysis' toolbox.  Speaking for myself, as both a remote sensing and GIS practitioner, I am excited about the prospect of being able to do more of my work within just one application environment, as opposed to doing my image stitching in Pix4D, image analysis in ENVI, and then finally my spatial analysis and mapping in ArcGIS.  For the sake of efficiency, I look very forward to the day that I can do all of this in just one app. 

For you drone enthusiasts out there, one neat new feature in the ArcGIS Pro Image Analysis tools is basic image stitching for producing color balanced orthomosaics and digital surface model outputs.  This new function is not at the level of what Pix4D or Drone-to-Map can do yet, but for basic RGB image processing it may be good enough for many people's needs.  Plus, it is a brand new tool that is bound to improve over time.

A couple more neat news items that were mentioned at the summit include:

The future for spatial science has never looked brighter!

Posted on Friday, November 10, 2017 at 6:28 PM
Tags: Education (1), ESRI (11), Imagery (5), Summit (1)

GIS and Remote Sensing Workshop with CERES Imaging

Last week IGIS was very pleased to partner with CERES Imaging Inc. (http://www.ceresimaging.net/) to provide a workshop on GIS and Remote Sensing for Crop Agriculture, in Davis CA.  This particular event represents an example of how UC ANR's IGIS Program is working with private industries to better deliver valuable services (in this case training) to public audiences who are eager to put the services/information into action; reinforcing UC ANR's priorities of public service and Cooperative Extension. 

This event was partially funded by a Department of Water Resources (DWR) grant to CERES, which among other things has helped CERES to provide very affordable, high resolution, multi-spectral, thermal and NDVI data to agriculturalists around California.  The objective of the DWR funding was to help facilitate more agile farming practices for water conservation, through the adoption of newly available aerial image products, while introducing farmers to contemporary image processing and mapping methods.  By partnering with CERES in this effort, IGIS is helping farmers to better utilize CERES's image products, to ideally make their operations more efficient and profitable.

Posted on Friday, November 10, 2017 at 6:09 PM

GO FAIR Data Stewardship Initiative Launched at UCSD. So What Does It Mean For Ag?

GO FAIR is an initiative to promote and support data stewardship that allows data to be Findable, Accessible, Interoperable, and Reusable. I was pleased to attend the launch of the first North American FAIR network last week at the UC San Diego Supercomputing Center.

Coping with a Data Tsunami

To say that we live in a data rich world is an understatement. We live a data drenched world (a fact I'm constantly reminded of by the 'hard drive full' warnings that pop-up on my computer on a weekly basis). Thanks to simultaneous, order-of-magnitude, advances in our ability to produce, disseminate, and store all manner of data, people working in fields from economics to physics to agriculture are struggling to benefit from, rather than be paralyzed by, the volume and diversity of data we produce. And this is by no means a problem only affecting academics, as more and more individuals, private companies and organizations are collecting and working with large volumes of data, from personal health sensors to drones. 

Adding to the challenge, there are often major barriers to get data to talk to each other. They may be stored in different formats, use different scales or units of analysis, or be under different restrictions. If you've ever carried personal health data from one doctors office to another by hand, you know what I mean.

FAIR Data Stewardship Principles

These are not new problems, but have taken on increased sense of urgency as the challenge gets worse and the demand for integrated analyses of complex problems grows. GO (Global Open) FAIR is a European based initiative that has two faces: i) a set a principles for data stewardship, and ii) a growing network of institutions and programs that are taking tangible steps toward a world in which data are Findable, Accessible, Interoperable, and Reusable. FAIR certainly doesn't mean that collected data have to be free or open access, but data stewardship should have a way to share information about the existence of data, and a means for access when appropriate. 

The FAIR principles mirror what open science advocates have argued for many years. As a program, GO FAIR has gained more traction than many of its predecessors. Following endorsements from the European Commission and other international bodies, the EU has already committed €2 billion to the first phase of implementation. Starting in 2018, the major EU funding agencies will require applicants to submit data stewardship plans that align with the FAIR principles. The initiative is also investing a lot in training people to use metadata standards and tools, many of which already exist. 

How is This Relevant for ANR?

ANR academics are impacted by the data psunami in at least two ways (neither for good). Like all practicing scientists, we have to deal with the usual challenges of managing large volumes of data, the frustrations of not being able to find or use data that others have collected, and the burden of all the gymnastics one must do to combine data from different sources into a robust, repeatable analysis. On top of that, as public servants whose work is funded by taxpayers, we have an additional moral and legal responsibility to be good stewards of all data collected for our public mission, which means ensuring the data we collect remains discoverable and accessible for other studies. Similarly, our extension mission also requires us to help California growers and land stewards get the most value from the data they collect, with tools that address their requirements for privacy and security.

While this may all seem like a lot to think about and additional work, the rewards are pretty exciting as the following video shows:

How Close are Your Data to Being FAIR?

For many us, putting the principles of FAIR data stewardship into practice will require a step or two we're not accustomed to, such as i) generating metadata in a format that can be read by both people and machines, and ii) storing our data (and metadata) for the long-term. The table below from a recent Nature article breaks down the gold standard a little further. 

Findable
F1. (meta)data are assigned a globally unique and persistent identifier
F2. data are described with rich metadata (defined by R1 below)
F3. metadata clearly and explicitly include the identifier of the data it describes
F4. (meta)data are registered or indexed in a searchable resource
Interoperable
I1. (meta)data use a formal, accessible, shared, and broadly applicable language for knowledge representation.
I2. (meta)data use vocabularies that follow FAIR principles
I3. (meta)data include qualified references to other (meta)data
Accessible
A1. (meta)data are retrievable by their identifier using a standardized communications protocol
A1.1 the protocol is open, free, and universally implementable
A1.2 the protocol allows for an authentication and authorization procedure, where necessary
A2. metadata are accessible, even when the data are no longer available
Reusable
R1. meta(data) are richly described with a plurality of accurate and relevant attributes
R1.1. (meta)data are released with a clear and accessible data usage license
R1.2. (meta)data are associated with detailed provenance
R1.3. (meta)data meet domain-relevant community standards

Wilkinson, Mark D., et al. "The FAIR Guiding Principles for scientific data management and stewardship." Scientific Data 3 (2016): 160018.

Baby Steps

As a research technology unit, I think we're doing fairly well in terms of keeping our data organized and accessible for the long-term. However after looking at our data management practices through the FAIR lens, I now see our metadata misses some important characteristics, a lot of the quality metrics aren't machine readable, and need to learn more about metadata repositories and discoverability, particularly for our drone data. These are challenges common to many new sources of geospatial data, and we look forward to engaging with the new arm of the GO FAIR network to develop solutions.

 

Posted on Saturday, November 4, 2017 at 6:12 PM
Tags: Data Stewardship (1), GoFair (1), IGIS (46)

#Fly4Fall! Add your drone 360 images to a global project

Super fun. We at IGIS and the Kellylab are working with Drone Scholars on the #Fly4Fall project. Fly4Fall: A citizen science experiment for crowd sourcing UAV data

Consider being a participant! Here are the contributions so far: 

Posted on Saturday, November 4, 2017 at 6:11 PM
Tags: citizen (11), drones (13), ecology (12)

Fly4Fall Campaign Seeks Aerial 360 Photos from Citizen Scientist Drone Users

Our colleague Greg Crutsinger at Drone Scholars recently launched an initiative to mobilize the large network of amateur UAV enthusiasts for an exciting campaign called Fly4Fall.

Under the Fly4Fall campaign, amateur drone hobbyists across the globe are invited to take aerial 360 photos with their drone and contribute them to a collection of fall landscapes that will grow over time.

Never taken an aerial 360 photo before? Me either, but fortunately it recently got a whole lot easier with a free iOS app called Hangar 360. The Hangar app flies your DJI drone for you, climbing to the height you program and then taking about 25 photos in a circle at three different angles to the horizon. The whole thing takes about 2 minutes, and you can collect multiple panos per flight. You then land the drone (but don't turn it off just yet!), transfer the photos from the drone to your phone over the WiFi, and then upload the photos to Hangar. Hangar stitches the photos for you in the cloud (also free!), and sends you a link. The results are stunning! See the panoramic photo below of Kearny REC made by IGIS's Robert Johnson earlier this week.

Inspired by citizen science initiatives like the Christmas Bird Count and Project BudBurst, where large numbers of naturalists record observations in a coordinated way, Fly4Fall is part non-professional science project, part art, part community building, and a whole lot of fun. Crutsinger discussed some of the potential science angles in a recent LinkedIn post.

Full instructions can be found at Fly4Fall.com. Currently, the Hangar app only works on iOS, unfortunately, and only with DJI drones (but the list includes most of the popular ones). Android enthusiasts can check out Litchi, which includes similar functionality but costs $25 and you have to process the images on your own (look for tutorials online).

Of course like any drone flight you have the follow the rules - only fly in permitted areas, don't fly directly over people, and be safe!

We look forward to seeing the Fly4Fall panoramas coming in. Feel free to use the comment box below to share your experiences and thoughts!

Posted on Friday, October 27, 2017 at 10:37 AM
Tags: Drones (13), Fly4Fall (1), Hangar (1), IGIS (46), UAVs (3)

Next 5 stories | Last story

 
E-mail
 
Webmaster Email: sdhogan@ucanr.edu