Publications

Synthetic Hyperspectral Image Generation to Aid RS Applications

Synthetic Hyperspectral Image Generation to Aid RS Applications Link to this paper

Learning from hyperspectral images (HSI) requires large annotated training datasets, which is laborious. The low spatial resolution of the images exponentially increases the challenge further and requires skilled domain experts. Prior arts using generative techniques and self-supervised learning also rely on large datasets for learning pretext tasks. We propose a novel Hyper-SynGen framework to synthetically generate HSI data without manual labeling. We extract the spatial geometry of the semantic classes using the image-to-image translation method from Google Maps and add the spectral information from the Tarang spectral library. To replicate the complexity of the real-world landscapes, we incorporate stochasticity in the class distribution at i) pixel level, ii) region level, and iii) global level, depending on the class at hand. This process relies on crowd-sourced labeling provided by Google Maps and does not require further manual labeling. Using this synthetic dataset to learn a pretext task on a self-supervised learning setup helps achieve better performance than the existing state-of-the-art. Hence, when replaced with real annotated data, the synthetically generated dataset is experimentally found to be as capable as real data, as it boosts the overall SOTA by a considerable margin.
Ushasi Chaudhuri, Shailesh Deshpande, Arpan Pal
IEEE IGARSS 2025
[Paper] [BibTex]

Determining spatial distribution of disaster intensities from news article using LLMs

Determining spatial distribution of disaster intensities from news article using LLMs Link to this paper

In recent years, the ability to automatically extract and analyze disaster-related information from unstructured news articles has become important for disaster response and management. In this research , the approach leverages language models, including a fine-tuned DistilBERT model for news article classification and the Google Gemini Flash 1.5 language model for disaster knowledge extraction, combined with gaussian distribution fitting for event projection as heatmap. The results demonstrate the model’s high accuracy in identifying disaster types, locations, and severity, and its ability to generate real-time disaster heatmap for effective decision-making support. This work is the first to integrate these methods in such a comprehensive way, providing a powerful tool for disaster response. By addressing the research question of how to automate disaster information extraction and visualization from news sources, this study offers a novel contribution to the field of disaster management.
Aniket Zope, Ushasi Chaudhuri, Shailesh Deshpande
IEEE IGARSS 2025
[Paper] [arXiv] [BibTex]

Machine learning-enabled computer vision for plant phenotyping: a primer on AI/ML and a case study on stomatal patterning

Machine learning-enabled computer vision for plant phenotyping: a primer on AI/ML and a case study on stomatal patterning Link to this paper

Artificial intelligence and machine learning (AI/ML) can be used to automatically analyze large image datasets. One valuable application of this approach is estimation of plant trait data contained within images. Here we review 39 papers that describe the development and/or application of such models for estimation of stomatal traits from epidermal micrographs. In doing so, we hope to provide plant biologists with a foundational understanding of AI/ML and summarize the current capabilities and limitations of published tools. While most models show human-level performance for stomatal density (SD) quantification at superhuman speed, they are often likely to be limited in how broadly they can be applied across phenotypic diversity associated with genetic, environmental, or developmental variation. Other models can make predictions across greater phenotypic diversity and/or additional stomatal/epidermal traits, but require significantly greater time investment to generate ground-truth data. We discuss the challenges and opportunities presented by AI/ML-enabled computer vision analysis, and make recommendations for future work to advance accelerated stomatal phenotyping.
Grace D. Tan, Ushasi Chaudhuri, Sebastian Varela, Narendra Ahuja, Andrew D.B. Leakey
Journal of Experimental Botany, Vol. 75, No. 21 pp. 6683–6703, 2024
[Paper] [BibTex]

Zero-shot sketch based image retrieval using graph transformer

Zero-shot sketch based image retrieval using graph transformer Link to this paper

The performance of a zero-shot sketch-based image retrieval (ZS-SBIR) task is primarily affected by two challenges. The substantial domain gap between image and sketch features needs to be bridged, while at the same time the side information has to be chosen tactfully. Existing literature has shown that varying the semantic side information greatly affects the performance of ZS-SBIR. To this end, we propose a novel graph transformer based zero-shot sketch-based image retrieval (GTZSR) framework for solving ZS-SBIR tasks which uses a novel graph transformer to preserve the topology of the classes in the semantic space and propagates the context-graph of the classes within the embedding features of the visual space. To bridge the domain gap between the visual features, we propose minimizing the Wasserstein distance between images and sketches in a learned domain-shared space. We also propose a novel compatibility loss that further aligns the two visual domains by bridging the domain gap of one class with respect to the domain gap of all other classes in the training set. Experimental results obtained on the extended Sketchy, TU-Berlin, and QuickDraw datasets exhibit sharp improvements over the existing state-ofthe-art methods in both ZS-SBIR and generalized ZS-SBIR.
Sumrit Gupta, Ushasi Chaudhuri, Biplab Banerjee, Saurabh Kumar
IEEE ICPR, 2022
[Paper] [arXiv] [BibTex]

Zero-Shot Cross-Modal Retrieval for Remote Sensing Images with Minimal Supervision

Zero-Shot Cross-Modal Retrieval for Remote Sensing Images with Minimal Supervision Link to this paper

The performance of a deep-learning-based model primarily relies on the diversity and size of the training dataset. However, obtaining such a large amount of labeled data for practical remote sensing (RS) applications is expensive and labor-intensive. Training protocols have been previously proposed for few-shot learning (FSL) and zero-shot learning (ZSL). However, FSL is not compatible with handling unobserved class data at the inference phase, while ZSL requires many training samples of the seen classes. In this work, we propose a novel training protocol for image retrieval and name it as label-deficit zero-shot learning (LDZSL). We use this novel LDZSL training protocol for the challenging task of cross-sensor data retrieval in RS. This protocol uses very few labeled data samples of the seen classes during training and interprets unobserved class data samples at the inference phase. This strategy is critical as some data modalities are hard to annotate without domain experts. This work proposes a novel bilevel Siamese network to perform the LDZSL cross-sensor retrieval of multispectral and synthetic aperture radar (SAR) images. We use the available georeferenced SAR and multispectral data to domain align the embedding features of the two modalities. We experimentally demonstrate the proposed model’s efficacy using the So2Sat dataset compared with the existing state-of-the-art models of the ZSL framework trained under a reduced training set. We also show the generalizability of the proposed model using a sketch-based image retrieval task. Experimental results on the Earth on the Canvas dataset exhibit comparative performance over the literature.
Ushasi Chaudhuri, Rupak Bose, Biplab Banerjee, Avik Bhattacharya, Mihai Datcu
IEEE Transactions on Geoscience and Remote Sensing, volume 60
[Paper] [BibTex]

Multi-Stage Semantic Graph Embeddings for Compositional Zero-Shot Learning

Multi-Stage Semantic Graph Embeddings for Compositional Zero-Shot Learning Link to this paper

We tackle the problem of compositional zero-shot learning (CZSL) where the task is to recognize novel composite semantic concepts (like red-tomato, wet-dog) consisting of states (red, wet) and objects (tomato, dog) characterizing the visual primitives. CZSL is a complex task given the fact that similar states may appear to be visually distinct, thus hampering the recognition performance. Recently, graph convolution networks (GCN) have been used to learn a compatibility function among the visual, state, and object embeddings. However, we postulate that these techniques do not fully explore the compositional nature of the semantic space where states, objects, and pairs follow different neighborhood topologies. To this end, we introduce the concept of multi-stage graph embeddings where separate GCNs are used to initially model the pairwise interactions for the states, objects, and the composition label embeddings, respectively. A composite GCN subsequently combines this information to learn a discriminative and neighborhood preserving latent semantic space ensuring the strong coupling of a composition label with its respective state and object. Further, we use a vision transformer for projecting the images onto the same latent space where the cross-modal information can be compared. An adaptive margin based cross-entropy loss is introduced to train the model end-to-end while ensuring enriched discrimination among the categories. Results on the MIT-States, UT-Zappos and C-GQA datasets confirm the superiority of the proposed approach.
Hitesh Kandala, Ruchika Chavhan, Ushasi Chaudhuri, Biplab Banerjee
Pattern Recognition Letters, 2022
[Paper]

BDA-SketRet: Bi-Level Domain Adaptation for Zero-Shot SBIR

BDA-SketRet: Bi-Level Domain Adaptation for Zero-Shot SBIR Link to this paper

The efficacy of zero-shot sketch-based image retrieval (ZS-SBIR) models is governed by two challenges. The immense distributions-gap between the sketches and the images requires a proper domain alignment. Moreover, the fine-grained nature of the task and the high intra-class variance of many categories necessitates a class-wise discriminative mapping among the sketch, image, and the semantic spaces. Under this premise, we propose BDA-SketRet, a novel ZS-SBIR framework performing a bi-level domain adaptation for aligning the spatial and semantic features of the visual data pairs progressively. In order to highlight the shared features and reduce the effects of any sketch or image-specific artifacts, we propose a novel symmetric loss function based on the notion of information bottleneck for aligning the semantic features while a cross-entropy-based adversarial loss is introduced to align the spatial feature maps. Finally, our CNN-based model confirms the discriminativeness of the shared latent space through a novel topology-preserving semantic projection network. Experimental results on the extended Sketchy, TU-Berlin, and QuickDraw datasets exhibit sharp improvements over the literature.
Ushasi Chaudhuri, Ruchika Chavan, Biplab Banerjee, Anjan Dutta, Zeynep Akata
Neurocomputing, Volume 514, pp 245-255, 2022
[arXiv] [Paper] [BibTex]

Inter-band Retrieval and Classification Using the Multi-labeled Sentinel-2 BigEarthNet Archive

Inter-band Retrieval and Classification Using the Multi-labeled Sentinel-2 BigEarthNet Archive Link to this paper

Conventional remote sensing data analysistechniques have a significant bottleneck of operating on a selectively chosen small-scale dataset. Availability of an enormous volume of data demands handling large-scale, diverse data, which have been made possible with neural network-based architectures. This article exploits the contextual information capturing ability of deep neural networks, particularly investigating multispectral band properties from Sentinel-2 image patches. Besides, an increase in the spatial resolution often leads to nonlinear mixing of land-cover types within a target resolution cell. We recognize this fact and group the bands according to their spatial resolutions, and propose a classification and retrieval framework. We design a representation learning framework for classifying the multispectral data by first utilizing all the bands and then using the grouped bands according to their spatial resolutions. We also propose a novel triplet-loss function for multilabeled images and use it to design an inter-band group retrieval framework. We demonstrate its effectiveness over the conventional triplet-loss function. Finally, we present a comprehensive discussion of the obtained results. We thoroughly analyze the performance of the band groups on various land-cover and land-use areas from agro-forestry regions, water bodies, and human-made structures. Experimental results for the classification and retrieval framework on the benchmarked BigEarthNet dataset exhibit marked improvements over existing studies.
Ushasi Chaudhuri, Subhadip Dey, Biplab Banerjee, Avik Bhattacharya, Mihai Datcu
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing (JSTARS), 2021
[Paper] [BibTex]

Inter-band Retrieval and Classification Using the Multi-labeled Sentinel-2 BigEarthNet Archive

Dual-Path Morph-UNet for Road and Building Segmentation from Satellite Images Link to this paper

Building footprints and road network detection have gained significant attention for map preparation, humanitarian aid dissemination, disaster management, to name a few. Traditionally, morphological filters excel at extracting shape features from remotely sensed images and have been widely used in the literature. However, the structural element (SE) dimension selection impedes these classical and learning-based methods utilizing any morphological operators. To overcome this aspect, we propose a novel framework to extract road and building from remote sensing (RS) images by exploiting morphological networks. The method predominantly aims at learning an optimized SE to capture variably-sized building and road footprints. We substitute convolutions with 2-D morphological operations in the basic building blocks of the network architecture (Dual-path Morph-UNet) to manage the intricate task of optimizing the SE in addition to the actual segmentation task. The dual-path framework incorporates parallel residual and dense paths in an encoder-decoder architecture, which permits learning of higher-level feature representations with fewer parameters. Finally, we implement the proposed framework on the benchmarked Massachusetts roads and buildings dataset and demonstrate superior results than the state-of-the-art (SOTA). In addition, the proposed network consists of 10× less learnable parameters than the SOTA methods.
Moni Shankar Dey, Ushasi Chaudhuri, Biplab Banerjee, Avik Bhattacharya
IEEE Geoscience and Remote Sensing Letters (GRSL), 2021
[Paper] [BibTex]

Attention Driven Graph Convolution Network for Remote Sensing Image Retrieval

Attention Driven Graph Convolution Network for Remote Sensing Image Retrieval Link to this paper

Graph convolution networks (GCNs) are useful in remote sensing (RS) image retrieval. It is found to be effective because, in a graph representation, the relative geometrical interactions between different regions (or segments) are appropriately captured, along with their region-wise features in their region adjacency graphs. Also, the attention mechanism has often been applied to the nodes to highlight the essential features in each node. In this regard, a significant amount of high-frequency information is missed since each image segment is effectively summarized within a single node. To account for this and increase the learning capacity, we propose to attend over the edge/adjacency matrix to highlight the interactions among meaningful regions that contribute to supervised learning from images. We exploit this novel edge attention mechanism together with node attention to highlight essential image context by allowing more importance to the meaningful neighboring regions that highlight a relevant node. We implement the proposed context-attended GCN framework for image retrieval on the benchmarked UC-Merced and the PatternNet datasets. We observe a notable improvement in the results compared to the state of the art.
Ushasi Chaudhuri, Biplab Banerjee, Avik Bhattacharya, Mihai Datcu
IEEE Geoscience and Remote Sensing Letters (GRSL), 2021
[Paper] [BibTex]

Synergistic Use of TanDEM-X and Landsat-8 Data for Crop-type Classification and Monitoring

Synergistic Use of TanDEM-X and Landsat-8 Data for Crop-type Classification and Monitoring Link to this paper

Classification of crop types using Earth Observation (EO) data is a challenging task. The challenge increases many folds when we have diverse crops within a resolution cell. In this regard, optical and Synthetic Aperture Radar (SAR) data provide complementary information to characterize a target. Therefore, we propose to leverage the synergy between multispectral and Synthetic Aperture Radar (SAR) data for crop classification. We aim to use the newly developed model-free three-component scattering power components to quantify changes in scattering mechanisms at different phenological stages. By incorporating interferometric coherence information, we consider the morphological characteristics of the crops that are not available with only polarimetric information. We also utilize the reflectance values from Landsat-8 spectral bands as complementary biochemical information of crops. The classification accuracy is enhanced by using these two pieces of information combined using a neural network-based architecture with an attention mechanism. We utilize the time series dual co-polarimetric (i.e., HH–VV) TanDEM-X SAR data and the multispectral Landsat-8 data acquired over an agricultural area in Seville, Spain. The use of the proposed attention mechanism for fusing SAR and optical data shows a significant improvement in classification accuracy by 6.0% to 9.0% as compared to the sole use of either the optical or SAR data. Besides, we also demonstrate that the utilization of single-pass interferometric coherence maps in the fusion framework enhances the overall classification accuracy by ≈3.0%. Therefore, the proposed synergistic approach will facilitate accurate and robust crop mapping with high-resolution EO data at larger scales.
Subhadip Dey, Ushasi Chaudhuri, Narayana Bhogapurapu, Juan Lopez Sanchez, Biplab Banerjee, Avik Bhattacharya, Dipankar Mandal, Yalamanchili S. Rao
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing (JSTARS), 2021
[Paper] [BibTex]

Contact me

Email

[first_name].c@tcs.com

Post

Quadra II, S.No. 238/239, Magarpatta Rd,
Keshav Nagar, Hadapsar, Pune,
Maharashtra 411028.