Modular Transformer Architecture for Precision Agriculture Imaging
By: Brian Gopalan, Nathalia Nascimento, Vishal Monga
Potential Business Impact:
Helps farmers find weeds in drone pictures.
This paper addresses the critical need for efficient and accurate weed segmentation from drone video in precision agriculture. A quality-aware modular deep-learning framework is proposed that addresses common image degradation by analyzing quality conditions-such as blur and noise-and routing inputs through specialized pre-processing and transformer models optimized for each degradation type. The system first analyzes drone images for noise and blur using Mean Absolute Deviation and the Laplacian. Data is then dynamically routed to one of three vision transformer models: a baseline for clean images, a modified transformer with Fisher Vector encoding for noise reduction, or another with an unrolled Lucy-Richardson decoder to correct blur. This novel routing strategy allows the system to outperform existing CNN-based methods in both segmentation quality and computational efficiency, demonstrating a significant advancement in deep-learning applications for agriculture.
Similar Papers
A Hybrid CNN-ViT-GNN Framework with GAN-Based Augmentation for Intelligent Weed Detection in Precision Agriculture
CV and Pattern Recognition
Helps robots find weeds to save crops.
Lightweight Multispectral Crop-Weed Segmentation for Precision Agriculture
CV and Pattern Recognition
Helps farmers tell weeds from crops better.
A Multimodal Transformer Approach for UAV Detection and Aerial Object Recognition Using Radar, Audio, and Video Data
CV and Pattern Recognition
Spots drones using many senses at once.