Home

Awesome

Agriculture-Vision

Agriculture-Vision Dataset, Challenge and Workshop (CVPR 2020)

A joint effort with many great collaborators to bring Agriculture and Computer Vision / AI communities together to benefit humanity!

Update:

The 2nd Agriculture-Vision Prize Challenge has started and new deadline is June 5, 2021. Total Prize: 20,000$.

Content

Papers:

CVPR paper on Agriculture-Vision <br/> ArXiv, CVPR 2020 open access

@article{chiu2020agriculture,
  title={Agriculture-Vision: A Large Aerial Image Database for Agricultural Pattern Analysis},
  author={Chiu, Mang Tik and Xu, Xingqian and Wei, Yunchao and Huang, Zilong and Schwing, Alexander and Brunner, Robert and Khachatrian, Hrant and Karapetyan, Hovnatan and Dozier, Ivan and Rose, Greg and others},
  journal={arXiv preprint arXiv:2001.01306},
  year={2020}
}

@InProceedings{Chiu_2020_CVPR,
author = {Chiu, Mang Tik and Xu, Xingqian and Wei, Yunchao and Huang, Zilong and Schwing, Alexander G. and Brunner, Robert and Khachatrian, Hrant and Karapetyan, Hovnatan and Dozier, Ivan and Rose, Greg and Wilson, David and Tudor, Adrian and Hovakimyan, Naira and Huang, Thomas S. and Shi, Honghui},
title = {Agriculture-Vision: A Large Aerial Image Database for Agricultural Pattern Analysis},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2020}
}

CVPR Workshop paper on the 1st Agriculture-Vision Challenge (challenge dataset, methods and results) <br/> ArXiv, CVPR 2020 open access

@article{chiu20201st,
  title={The 1st Agriculture-Vision Challenge: Methods and Results},
  author={Chiu, Mang Tik and Xu, Xingqian and Wang, Kai and Hobbs, Jennifer and Hovakimyan, Naira and Huang, Thomas S. and Shi, Honghui and others},
  journal={arXiv preprint arXiv:2004.09754},
  year={2020}
}

@InProceedings{Chiu_2020_CVPR_Workshops,
author = {Chiu, Mang Tik and Xu, Xingqian and Wang, Kai and Hobbs, Jennifer and Hovakimyan, Naira and Huang, Thomas S. and Shi, Honghui},
title = {The 1st Agriculture-Vision Challenge: Methods and Results},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2020}
}

Dataset

Agriculture-Vision Challenge Dataset

<img src="images/cs.gif" width="300" title="Cloud shadow"> <img src="images/dp.gif" width="300" title="Double plant"> <img src="images/ps.gif" width="300" title="Planter skip"> <img src="images/sw.gif" width="300" title="Standing water"> <img src="images/ww.gif" width="300" title="Waterway"> <img src="images/wc.gif" width="300" title="Weed cluster">

The dataset used in this challenge is a subset of the Agriculture-Vision dataset. The challenge dataset contains 21,061 aerial farmland images captured throughout 2019 across the US. Each image consists of four 512x512 color channels, which are RGB and Near Infra-red (NIR). Each image also has a boundary map and a mask. The boundary map indicates the region of the farmland, and the mask indicates valid pixels in the image. Regions outside of either the boundary map or the mask are not evaluated.

This dataset contains six types of annotations: Cloud shadow, Double plant, Planter skip, Standing Water, Waterway and Weed cluster. These types of field anomalies have great impacts on the potential yield of farmlands, therefore it is extremely important to accurately locate them. In the Agriculture-Vision dataset, these six patterns are stored separately as binary masks due to potential overlaps between patterns. Users are free to decide how to use these annotations.

Each field image has a file name in the format of (field id)_(x1)-(y1)-(x2)-(y2).(jpg/png). Each field id uniquely identifies the farmland that the image is cropped from, and (x1, y1, x2, y2) is a 4-tuple indicating the position in which the image is cropped. Please refer to our paper for more details regarding how we construct the dataset.

Download

The challenge dataset contains images, boundaries and masks for train, val and test set. It also contains labels for the train and val set only. The 2021 new dataset .tar.gz file is around 20 GB. Please visit our website or our huggingface page to get access. By downloading or using the dataset, user signifies its agreement to Agriculture-Vision Workshop Terms and Conditions.

Evaluation

Evaluation metric

We use mean Intersection-over-Union (mIoU) as our main quantitative evaluation metric, which is one of the most commonly used measures in semantic segmentation datasets. The mIoU is computed as:

<img src="images/metric.png">

Where c is the number of annotation types (c = 7 in our dataset, with 6 patterns + background), P<sub>c</sub> and T<sub>c</sub> are the predicted mask and ground truth mask of class c respectively.

Since our annotations may overlap, we modify the canonical mIoU metric to accommodate this property. For pixels with multiple labels, a prediction of either label will be counted as a correct pixel classification for that label, and a prediction that does not contain any ground truth labels will be counted as an incorrect classification for all ground truth labels.

Concretely, we construct the confusion matrix M<sup>c×c</sup> with the following rules:

For each prediction x and label set Y:

  1. If xY, then M<sub>y,y</sub> = M<sub>y,y</sub> + 1 for each y in Y<br/>
  2. Otherwise, M<sub>x,y</sub> = M<sub>x,y</sub> + 1 for each y in Y<br/>

The mIoU is finally computed by (true_positive) / (prediction + target - true_positive), averaged across all classes.

Submission

Update:

Registration

We are now hosting our evaluation/competition on Codalab. The competition page can be found here. Each participating team is required to register for the challenge. To register your team, fill out the registration form here, then register on the competition page.

*Make sure your Codalab account email matches one of the member emails in the registration form. Each team can only register once.

Codalab submission

All registered teams can evaluate their results on Codalab and publish their results on the leaderboard. The submission file should be a compressed .zip file that contains all prediction images. All prediction images should be in png format and the file names and image sizes should match the input images exactly. The prediction images will be converted to a 2D numpy array with the following code:

numpy.array(PIL.Image.open(‘field-id_x1-y1-x2-y2.png’))

In the loaded numpy array, only 0-6 integer labels are allowed, and they represent the annotations in the following way:

0 - background<br/> 1 - cloud_shadow<br/> 2 - double_plant<br/> 3 - planter_skip<br/> 4 - standing_water<br/> 5 - waterway<br/> 6 - weed_cluster<br/>

This label order will be strictly followed during evaluation.

All teams can have 2 submissions per day and 20 submissions in total.

Challenge

We hosted the 1st Agriculture-Vision Prize Challenge with the following terms:

The prize award will be granted to the top 3 teams on the prize challenge leaderboard (closed on 4/20/2020) that provide a valid final submission.

Teams and Methods

Please refer to our workshop summary paper for more details regarding notable methods. Selected teams & methods:

Team Hyunseong: Hyunseong Park, Junhee Kim, Sungho Kim (Agency for Defense Development, South Korea)

Team SCG Vision: Qinghui Liu, Michael C. Kampffmeyer, Robert Jenssen, Arnt B. Salberg (Norwegian Computing Center, UiT The Arctic University of Norway)

Team AGR: Alexandre Barbosa, Rodrigo Trevisan (University of Illinois at Urbana Champaign)

Team TJU: Bingchen Zhao, Shaozuo Yu, Siwei Yang, Yin Wang (Tongji University)

Team Haossr: Hao Sheng, Xiao Chen, Jingyi Su, Ram Rajagopal, Andrew Ng (Stanford University, Chegg, Inc.)

Team CNUPR_TH2L: Van Thong Huynh, Soo-Hyung Kim, In-Seop Na (Chonnam National University, Chosun University)

Team TeamTiger: Ujjwal Baid, Shubham Innani, Prasad Dutande, Bhakti Baheti, Sanjay Talbar (SGGS Institute of Engineering and Technology)

Challenge leaderboard

TeammIoU (%)BackgroundCloud shadowDouble plantPlanter skipStanding waterWaterwayWeed cluster
Hyunseong63.980.656.057.957.575.063.756.9
seungjae62.279.344.460.465.976.955.453.2
yjl912.261.580.153.746.148.676.871.553.6
ddcm60.880.551.058.649.872.059.853.8
RodrigoTrevisan60.580.243.857.551.675.366.249.2
SYDu59.581.341.650.343.473.271.755.2
agri59.278.255.842.942.077.564.753.2
Tennant57.479.936.654.841.469.866.952.0
celery03.055.479.138.943.341.273.061.550.5
stevenwudi55.077.442.054.420.169.567.753.8
PAII55.079.938.647.626.274.662.155.7
agrichallenge1.254.680.950.939.329.273.457.850.5
hui54.080.241.646.420.872.864.851.4
shenchen61.653.779.436.756.321.667.061.852.8
NTU53.679.841.449.413.573.361.856.0
tpys53.081.150.537.125.967.458.750.1
Simple52.780.240.045.224.670.957.650.4
Ursus52.378.936.337.834.469.357.152.3
liepieshov52.177.240.246.016.071.362.951.1
Lunhao49.479.540.438.810.569.458.349.1
tetelias-mipt49.280.437.834.804.670.662.553.8
Dataloader48.979.142.035.809.168.756.751.3
Hakjin46.478.632.038.301.866.258.049.9
JianyuTANG44.678.137.931.815.447.354.846.9
haossr43.979.221.428.102.767.556.452.3
rpartsey41.572.521.636.209.159.740.750.6
baidujjwal40.875.226.140.109.948.037.149.5
Chaturlal40.777.723.020.405.055.051.052.9
Sciforce40.280.529.624.40.041.255.950.0
MustafaA40.176.534.425.611.146.036.550.3
HaotianYan36.877.121.925.113.757.524.337.9
gro36.376.437.508.40.060.329.741.8
oscmansan35.571.629.603.00.052.446.245.9
ThorstenC33.672.322.310.002.040.840.147.8
ZHwang33.576.532.412.90.057.215.939.9
fayzur2.022.165.421.802.200.223.313.428.7
gaslen.221.571.003.317.900.810.206.940.1
dvkhandelwal16.371.50.00.00.042.60.00.0
ajeetsinghiitd10.356.900.200.40.00.000.114.5

Workshop

Schedule

(Sunday, Jun 14 PDT)


08:30 – 08:40 Opening Remarks and Welcome

Humphrey Shi (University of Oregon)

Naira Hovakimyan (UIUC)


08:40 – 09:00 Invited Talk 1: Scaling Spatio-Temporal Analytics: A Case in Agricultural Insights (Talk video here)

Sharath Pankanti (IBM T.J. Watson Research Center)

09:00 – 09:20 Invited Talk 2: Multi-modality Remote Sensing in High Throughput Phenotyping: Opportunities for Machine Learning (Talk video here)

Melba M. Crawford, Edward J. Delp (Purdue)

09:20 – 09:40 Invited Talk 3: Bridging Application and Research. Is agriculture just another application? (Talk video here)

Jennifer Hobbs (IntelinAir Inc.)

09:40 – 10:00 Invited Talk 4: Improving Visual and Speech Recognition on Out-Domain Data (Talk video here)

Liangliang Cao (Google Inc. and Umass)

10:00 – 10:20 Invited Talk 5: Learning to Anticipate (Talk video here)

Alex Schwing (UIUC)


10:20 – 11:10 Poster Session 1


11:10 – 11:20 Oral 1: MSCG-Net with Adaptive Class Weighting Loss for Semantic Segmentation (Talk video here)<br/> 11:20 – 11:30 Oral 2: Finding Berries: Segmentation and Counting of Cranberries using Point Supervision and Shape Priors (Talk video here)<br/> 11:30 – 11:40 Oral 3: Visual 3D Reconstruction and Dynamic Simulation of Fruit Trees for Robotic Manipulation (Talk video here)<br/> 11:40 – 11:50 Oral 4: Cross-Regional Oil Palm Tree Detection (Talk video here)<br/> 11:50 – 11:00 Oral 5: Effective Data Fusion with Generalized Vegetation Index (Talk video here)<br/> 11:00 – 12:10 Oral 6: Weakly Supervised Learning Guided by Activation Mapping Applied to a Novel Citrus Pest Benchmark (Talk video here)<br/> 12:10 – 12:20 Oral 7: Fine-Grained Recognition in High-throughput Phenotyping (Talk video here)<br/> 12:20 – 12:30 Oral 8: Climate Adaptation: Reliably Predicting from Imbalanced Satellite Data (Talk video here)<br/>


12:30 – 13:30 Lunch Break


13:30 – 13:50 Invited Talk 6: Towards Understanding Agricultural Systems at Scale using Machine Learning (Talk video here)

Stefano Ermon (Stanford)

13:50 – 14:10 Invited Talk 7: Interactive Object Segmentation (Live*) (Talk video here)

Yunchao Wei (University of Technology Sydney)

14:10 – 14:30 Invited Talk 8: Project FarmBeats: Aerial Mapping forAgricultural Farms and Beyond (Live*) (Talk video here)

Ranveer Chandra, Sudipta Sinha (Microsoft)

14:30 – 14:50 Invited Talk 9: Enabling the African Farmer (Live*) (Talk video here)

Munther Dahleh (MIT)

14:50 – 15:10 Invited Talk 10: (Live*)

Zsolt Kira (Georgia Tech)


15:10 – 15:55 Poster Session 2


15:55 – 16:10 Agriculture-Vision Prize Challenge Results Summary (Talk video here)

16:10 – 16:20 Challenge Presentation 1: Residual DenseNet with Expert Network for Semantic Segmentation

16:20 – 16:30 Challenge Presentation 2: MSCG-Net Models for The 1st Agriculture-Vision Challenge


16:30 – 17:20 Panel Discussion: Challenge & Opportunities for Computer Vision in Agriculture (video here)

Host:<br/> Naira Hovakimyan (UIUC)

Panelists:<br/> Sharath Pankanti (IBM T. J. Watson Research Center)

Liangliang Cao (Google Research)

Ranveer Chandra (Microsoft Azure Global)

Sudipta Sinha (Microsoft)

Munther Dahleh (MIT)

Ed Delp (Purdue)

Melba Crawford (Purdue)

Jennifer Hobbs (Intelinair Inc.)

Yunchao Wei (University of Technology Sydney)

Alexander Schwing (UIUC)

Zsolt Kira (Georgia Tech)

David Wilson (Intelinair Inc.)

Jim Yuan (Google X)

Humphrey Shi (U of Oregon)


17:20 – 17:30 Closing Remarks

Humphrey Shi (University of Oregon)

Naira Hovakimyan (UIUC)


Paper Track

Topic Description

The Workshop will be open to the broader community addressing various technical and application aspects of challenges and opportunities in computer vision research for agriculture. We aim to provide a venue to both show current relevant efforts in interdisciplinary areas between computer vision and agriculture, and to encourage further research and conversations within the computer vision community to tackle impactful agriculture-vision problems. We will solicit papers from the following wide range of topics:

Accepted Papers

Accepted Posters