Federated large language models (FedLLMs) enable powerful generative
capabilities within wireless networks while preserving data privacy.
Nonetheless, FedLLMs remain vulnerable to model poisoning attacks. This article
first reviews recent advancements in model poisoning techniques and existing
defense mechanisms for FedLLMs, underscoring critical limitations, especially
when dealing with non-IID textual data distributions. Current defense
strategies predominantly employ distance or similarity-based outlier detection
mechanisms, relying on the assumption that malicious updates markedly differ
from benign statistical patterns. However, this assumption becomes inadequate
against adaptive adversaries targeting billion-parameter LLMs. The article
further investigates graph representation-based model poisoning (GRMP), an
emerging attack paradigm that exploits higher-order correlations among benign
client gradients to craft malicious updates indistinguishable from legitimate
ones. GRMP can effectively circumvent advanced defense systems, causing
substantial degradation in model accuracy and overall performance. Moreover,
the article outlines a forward-looking research roadmap that emphasizes the
necessity of graph-aware secure aggregation methods, specialized vulnerability
metrics tailored for FedLLMs, and evaluation frameworks to enhance the
robustness of federated language model deployments.