More Research
What Do You See in Common? Learning Hierarchical Prototypes over Tree-of-Life to Discover Evolutionary Traits
1Harish Babu Manogaran, 1M. Maruf, 6Arka Daw, 1Kazi Sajeed Mehrab, 1Caleb Charpentier, 1Josef Uyeda, 2Wasila Dahdul, 3Matthew J Thompson, 3Elizabeth G Campolongo, 3Kaiya L Provost, 4Paula Mabee, 5Hilmar Lapp, 1Anuj Karpatne
1Virginia Tech, 2University of California, Irvine, 3The Ohio State University, 4Battelle, 5Duke University, 6Oak Ridge National Lab,
harishbabu@vt.edu, karpatne@vt.edu
![](images/new_teaser_5/new_teaser_5-1.png)
HComP-Net
We introduce an hierarchical interpretable image classification framework called HComP-Net that can be applied to discover potential evolutionary traits from images by making use of the Phylogenetic tree also called as Tree-Of-Life. HComP-Net idenitfies evolutionary traits by means of prototypes that are representative of the visual trait that is unique to a species or a group of species that share a common ancestor in the phylogeny. HComPNet can generate hypothesis for potential evolutionary traits by learning semantically meaningful non-over-specific prototypes at each internal node of the hierarchy. In order to discover evolutionary traits through prototypes, our approach features the following,
- Hierarchical Prototypes: We learn prototypes at each internal node of the phylogeny such that each prototype captures a evolutionary trait of the species that are leaf descendants of the node.
- Avoiding Over-specificity: Over-specific prototypes represents traits that are specific to one or few of the leaf descendant species, rather than representing a common shared trait. In order to avoid over-specificity.
- We introduce a novel over-specificity loss that enforces each prototype to represent a common trait of all leaf descendants.
- We introduce masking module that helps to idenitfy and discard over-specific prototypes that are learned by the model.
- Learning Discriminative Traits: We introduce discriminative loss to ensure the prototypes represent traits that are absent in the contrasting set of species that have evolved from a different ancestor.
![](images/architecture/architecture-1.png)
Results
Classification Performance of HComP-Net
The comparison of the fine-grained accuracy calculated for HComP-Net and the baselines on Bird, Fish, and Butterfly datasets are given below
Model | Hierarchy | Bird | Butterfly | Fish |
---|---|---|---|---|
ResNet-50 | No | 74.18 | 95.76 | 86.63 |
INTR | 69.22 | 95.53 | 86.73 | |
HPnet | Yes | 36.18 | 94.69 | 77.51 |
HComP-Net | 70.01 | 97.35 | 90.80 |
Semantic Quality of Prototypes
We calculate the semantic quality of each prototype by means of part purity on Bird dataset. Prototypes that represent the same part (eye, beak, tail, etc.) in Top-10 closest images have part purity close to 1. We report the mean and standard deviation of the part purity of all the prototypes learned. % masked denotes the percentage of prototypes that were identified as over-specific and removed using masking module.
Model | Lovsp | Masking | Part purity | % masked |
---|---|---|---|---|
HPnet | - | - | 0.14 ± 0.09 | - |
HComP-Net | - | - | 0.68 ± 0.22 | - |
HComP-Net | - | ✓ | 0.75 ± 0.17 | 21.42% |
HComP-Net | ✓ | - | 0.72 ± 0.19 | - |
HComP-Net | ✓ | ✓ | 0.77 ± 0.16 | 16.53% |
Prototype Visualizations
![](images/Birds_and_Fish-Tree_proto_viz_horizontal_2.png)
![](images/Shorter Proto Timeline 2 - Western Grebe without INTR (1)-1 cropped.png)
![](images/Butterfly30-Tree_proto_viz_3/Butterfly30-Tree_proto_viz_3-1.png)
References
Please cite our paper if you use our code, data, model or results.
@article{manogaran2024you, title={What Do You See in Common? Learning Hierarchical Prototypes over Tree-of-Life to Discover Evolutionary Traits}, author={Manogaran, Harish Babu and Maruf, M and Daw, Arka and Mehrab, Kazi Sajeed and Charpentier, Caleb Patrick and Uyeda, Josef C and Dahdul, Wasila and Thompson, Matthew J and Campolongo, Elizabeth G and Provost, Kaiya L and others}, journal={arXiv preprint arXiv:2409.02335}, year={2024} }
Acknowledgements
This research is supported by National Science Foundation (NSF) awards for the HDR Imageomics Institute (OAC-Award #2118240). We are thankful for the support of computational resources provided by the Advanced Research Computing (ARC) Center at Virginia Tech. This manuscript has been authored by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The US government retains, and the publisher, by accepting the article for publication, acknowledges that the US government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this manuscript or allow others to do so for US government purposes. DOE will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (https://www.energy.gov/doe-public-access-plan).