Score: 0

Privacy-Preserving Offloading for Large Language Models in 6G Vehicular Networks

Published: August 30, 2025 | arXiv ID: 2509.05320v1

By: Ikhlasse Badidi , Nouhaila El Khiyaoui , Aya Riany and more

Potential Business Impact:

Keeps car data private when using smart driving AI.

Business Areas:
Autonomous Vehicles Transportation

The integration of Large Language Models (LLMs) in 6G vehicular networks promises unprecedented advancements in intelligent transportation systems. However, offloading LLM computations from vehicles to edge infrastructure poses significant privacy risks, potentially exposing sensitive user data. This paper presents a novel privacy-preserving offloading framework for LLM-integrated vehicular networks. We introduce a hybrid approach combining federated learning (FL) and differential privacy (DP) techniques to protect user data while maintaining LLM performance. Our framework includes a privacy-aware task partitioning algorithm that optimizes the trade-off between local and edge computation, considering both privacy constraints and system efficiency. We also propose a secure communication protocol for transmitting model updates and aggregating results across the network. Experimental results demonstrate that our approach achieves 75\% global accuracy with only a 2-3\% reduction compared to non-privacy-preserving methods, while maintaining DP guarantees with an optimal privacy budget of $\varepsilon = 0.8$. The framework shows stable communication overhead of approximately 2.1MB per round with computation comprising over 90\% of total processing time, validating its efficiency for resource-constrained vehicular environments.

Country of Origin
🇲🇦 Morocco

Page Count
7 pages

Category
Computer Science:
Cryptography and Security