Tuesday, May 7, 2024

Large Language Models for UAVs: Current State and Pathways to the Future


Large Language Models for UAVs: Current State and Pathways to the Future

Computer Science > Artificial Intelligence

Summary 

The paper provides a comprehensive review of integrating Large Language Models (LLMs) with Unmanned Aerial Vehicles (UAVs) to enhance their capabilities and applications. Key points:

1. LLMs like BERT, GPT, T5, XLNet, ERNIE, and BART can significantly improve UAV performance in various tasks such as command interpretation, situation awareness, and decision-making.

2. LLM integration enables UAVs to process and analyze large amounts of data in real-time, enhancing their autonomy and adaptability in dynamic environments.

3. LLMs can optimize UAV communication by improving spectrum sensing and sharing capabilities, leading to more efficient and reliable data transmission.

4. LLM-integrated UAVs have wide applications in surveillance, emergency response, delivery services, environmental monitoring, and satellite communications.

5. Challenges in implementing LLMs in UAVs include computational resource constraints, latency issues, model robustness, integration with existing systems, and data security.

6. Future research directions focus on advancements in LLM algorithms, integration with emerging technologies, computational efficiency optimization, latency reduction, reliability enhancement, and regulatory considerations.

The paper highlights the transformative potential of LLM-UAV integration in creating more sophisticated, intelligent, and responsive autonomous systems across various domains.

Challenges

The paper discusses several key challenges and considerations in implementing LLMs in UAVs:

1. Computational Resources and Power Consumption:
  •    - LLMs require significant computational power and energy, which can be limited on UAVs due to their lightweight design and power supply constraints.
  •    - The high power consumption of LLMs can quickly drain UAV batteries, reducing flight duration and operational efficiency.
  •    - Addressing this challenge involves model simplification, quantization techniques, and the use of advanced AI hardware like GPUs, FPGAs, and ASICs.
2. Communication Latency:
  •    - Real-time data processing and decision-making in UAV operations can be hindered by the latency introduced when offloading LLM computations to cloud servers.
  •    - This latency can be detrimental in scenarios requiring immediate responses, such as navigation, surveillance, and tactical operations.
  •    - Mitigating latency involves enhancing onboard processing capabilities, implementing edge computing solutions, and optimizing data transmission protocols.
3. Model Robustness and Reliability:
  •    - LLMs may produce unpredictable or incorrect outputs in novel or edge-case scenarios, as they rely on patterns learned from training data that may not cover all real-world situations.
  •    - The risk is particularly high in dynamic UAV environments where quick and accurate decisions are crucial.
  •    - Ensuring robustness requires continuous model updating with new data, extensive simulation-based testing, fail-safe mechanisms, and redundant systems.
4. Integration with Existing Systems:
  •    - Integrating LLMs with UAVs' diverse hardware and software components, each with unique specifications and requirements, can be complex and time-consuming.
  •    - Seamless interaction between LLMs and flight control, navigation, communication, and data processing units is essential for optimal performance.
  •    - Addressing integration challenges involves modular system design, standardized data formats, incremental testing, and systematic maintenance frameworks.
5. Data Security and Privacy:
  •    - UAVs equipped with LLMs often process sensitive data, including personal information collected during missions, which can be vulnerable to breaches and unauthorized access.
  •    - Ensuring data security is crucial to prevent privacy violations and maintain the confidentiality and integrity of the information.
  •    - Robust data encryption, access control mechanisms, and compliance with data protection regulations are essential for safeguarding sensitive data processed by LLMs in UAVs.
Addressing these challenges requires a multifaceted approach involving advancements in hardware, software, and system design, as well as the development of robust security measures and regulatory frameworks. Ongoing research and innovation in these areas are crucial for realizing the full potential of LLM-integrated UAVs while ensuring their safe, reliable, and responsible deployment.

LLMs used on UAVs

Several LLMs have been implemented on UAVs to enhance their capabilities and performance. Some notable examples mentioned in the paper include:

1. BERT (Bidirectional Encoder Representations from Transformers):
  •    - BERT has been used to interpret complex commands, extract relevant information from mission data, and enhance contextual understanding in UAV operations.
  •    - It has been particularly useful in surveillance and monitoring missions where precise understanding of sensor data and operational directives is critical.

2. GPT (Generative Pre-trained Transformer) Series:
  •    - GPT models have been employed to generate detailed mission reports, conduct dialogues, and enable natural language interaction with UAVs.
  •    - They have been used in UAV training simulations for generating realistic scenarios and interactive communications.

3. T5 (Text-to-Text Transfer Transformer):
  •    - T5 has been utilized for tasks such as translating communications between different languages or protocols, summarizing extensive exploration data, and transforming raw sensor outputs into actionable text formats.
  •    - Its versatility in handling diverse text-based tasks has made it valuable in various UAV communication scenarios.

4. XLNet (eXtreme Learning NETwork):
  •    - XLNet has been applied in complex, dynamic operational environments such as search, rescue, and disaster response, where interpreting and responding to context-heavy instructions in real-time is essential.
  •    - Its flexible and comprehensive language understanding capabilities have been beneficial in these challenging scenarios.

5. ERNIE (Enhanced Representation through kNowledge Integration):
  •    - ERNIE has been used in UAV missions that require a deep understanding of specific terminologies or concepts, such as environmental monitoring applications involving ecological data.
  •    - Its ability to integrate external knowledge through knowledge graphs has been valuable in these domain-specific contexts.
6. BART (Bidirectional and Auto-Regressive Transformers):
  •    - BART has been employed for tasks such as generating precise and contextually accurate mission reports or instructions, particularly in scenarios involving the summarization of detailed surveillance data.
  •    - Its dual capabilities in understanding and generating text have made it useful in UAV communication tasks that require both comprehension and generation.

These LLMs have been adapted and integrated into various UAV systems to enhance their natural language processing, decision-making, and communication capabilities. The choice of LLM depends on the specific requirements and challenges of the UAV application, such as the need for contextual understanding, text generation, or domain-specific knowledge integration. As research in this field progresses, it is likely that more LLMs will be developed and optimized specifically for UAV applications, further expanding the possibilities for intelligent and autonomous UAV operations.

Abstract: Unmanned Aerial Vehicles (UAVs) have emerged as a transformative technology across diverse sectors, offering adaptable solutions to complex challenges in both military and civilian domains. Their expanding capabilities present a platform for further advancement by integrating cutting-edge computational tools like Artificial Intelligence (AI) and Machine Learning (ML) algorithms. These advancements have significantly impacted various facets of human life, fostering an era of unparalleled efficiency and convenience.

Large Language Models (LLMs), a key component of AI, exhibit remarkable learning and adaptation capabilities within deployed environments, demonstrating an evolving form of intelligence with the potential to approach human-level proficiency. This work explores the significant potential of integrating UAVs and LLMs to propel the development of autonomous systems.

We comprehensively review LLM architectures, evaluating their suitability for UAV integration. Additionally, we summarize the state-of-the-art LLM-based UAV architectures and identify novel opportunities for LLM embedding within UAV frameworks. Notably, we focus on leveraging LLMs to refine data analysis and decision-making processes, specifically for enhanced spectral sensing and sharing in UAV applications. Furthermore, we investigate how LLM integration expands the scope of existing UAV applications, enabling autonomous data processing, improved decision-making, and faster response times in emergency scenarios like disaster response and network restoration. Finally, we highlight crucial areas for future research that are critical for facilitating the effective integration of LLMs and UAVs.
Subjects: Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Robotics (cs.RO)
Cite as: arXiv:2405.01745 [cs.AI]
  (or arXiv:2405.01745v1 [cs.AI] for this version)

Submission history

From: Shumaila Javaid [view email]
[v1] Thu, 2 May 2024 21:30:10 UTC (3,455 KB)

arxiv.org

Large Language Models for UAVs: Current State and Pathways to the Future


Shumaila Javaid , Nasir Saeed , Bin He S. Javaid and B. He are with the Department of Control Science and Engineering, College of Electronics and Information Engineering, Tongji University, Shanghai 201804, China, and also with Frontiers Science Center for Intelligent Autonomous Systems, Shanghai 201210. Email: {shumaila, hebin}@tongji.edu.cn.
N. Saeed is with the Department of Electrical and Communication Engineering, United Arab Emirates University (UAEU), Al Ain, 15551, UAE. Email: mr.nasir.saeed@ieee.org.

Index Terms:

UAVs, Large Language Models, spectral sensing, autonomous systems, decision-making

I Introduction

Unmanned Aerial Vehicles (UAVs) have been a focus of attention for over fifty years due to their remarkable autonomy, mobility, and adaptability, enhancing a wide array of applications, including surveillance [1, 2], monitoring [3, 4], search and rescue [5], healthcare [6], maritime communications [7], and wireless network provisioning [8]. These foundational achievements prompted the integration of Artificial Intelligence (AI) with UAVs. Particularly in the 2010s, advances in both UAV technology and AI reached a critical juncture, leading to substantial benefits across various applications. For example, AI-enabled UAVs employ facial recognition and real-time video analysis techniques to enhance security and monitoring of remote areas [9, 10, 11]. In agriculture, UAVs with AI models analyze crop health for precise farming, improving resource efficiency and yields [12, 13]. Meanwhile, AI-driven UAVs optimize logistics route planning and inventory management, streamlining warehouse operations and improving delivery efficiency [14, 15, 16].

Among these advancements, Large Language Models (LLMs) have recently gained significant attention as they enable systems to learn from application behavior and optimize existing systems [17, 18]. Various LLMs employing transformer architectures, such as the Generative Pre-trained Transformer (GPT) series [19], Bidirectional Encoder Representations from Transformers (BERT) [20], and Text-to-Text Transfer Transformer (T5) [21], exhibit fundamental capabilities. Due to extensive training on large datasets, they excel at understanding, generating, and translating human-like text, making them valuable for robotics, healthcare, finance, education, customer service, and content creation applications. Furthermore, the proficiency of these models in real-time data processing, natural language understanding and generation, content recommendation, sentiment analysis, automated response, language translation, and content summarization creates opportunities in the UAV domain. For instance, they enable UAVs to respond swiftly to dynamic environmental changes and communication demands [22, 23]. Their adaptive learning capabilities facilitate continuous improvement in operational strategies based on incoming data, enhancing decision-making processes [24]. Additionally, their ability to support multiple languages broadens their applicability in global operations, particularly valuable for UAV communications in diverse applications such as smart cities, healthcare, rescue operations, emergency response, media, and entertainment [25, 26, 27].

Recent literature [28, 29, 30] has explored incorporating LLMs into UAV communication systems to enhance interaction with human operators and among UAVs. Traditionally, UAVs operate on pre-programmed commands with limited dynamic interaction capabilities. However, integrating LLMs enables support for natural and intuitive communication methods. For example, LLMs can interpret and respond to commands in natural language, simplifying UAV control and allowing the handling of complex, real-time mission adjustments. This transforms UAVs into more adaptable and practical tools across various applications [31]. LLMs can enhance UAVs’ autonomous decision-making based on communication context or environmental data [32, 33]. For example, without human input, LLMs can analyze messages and environmental data in search and rescue operations to determine priorities and actions. In multi-UAV operations, LLMs can facilitate better communication and coordination, managing and optimizing information flow between UAVs and improving overall efficiency and effectiveness. LLMs can also enhance data processing and reporting capabilities by generating summaries, insights, and actionable recommendations from vast amounts of collected data. Furthermore, LLMs can be trained to recognize patterns and anomalies in communication data, crucial for preempting and resolving potential issues [34, 35]. For instance, if a UAV sends inconsistent data, LLMs could quickly detect anomalies and alert operators. LLMs can enhance scalability and adaptation in communication protocols, automatically learning and adapting to new protocols based on new data or operational changes, ensuring seamless communication. Pre-training LLMs with simulated data helps understand mission conditions and requirements, enabling real-time adaptation during missions for optimal performance.

This work is motivated by the potential of integrating LLMs into UAV communication systems. We comprehensively analyze existing LLM methods focusing on UAV integration to highlight advantages and limitations in expanding current UAV communication systems’ capabilities. The review summarizes state-of-the-art LLM-integrated architectures, explores opportunities for LLM incorporation into UAV architecture, and addresses spectrum sensing and sharing concerning LLM integration. We aim to showcase how LLMs can optimize communication, adapt to new missions dynamically, and process complex data streams, enhancing UAV efficiency and versatility across various domains, including emergency response, environmental monitoring, urban planning, and satellite communication. Additionally, we address the legal, ethical, and technical challenges of deploying AI-driven UAVs, emphasizing responsible and effective integration, laying the groundwork for advancing UAV technology to meet future demands, and exploring innovative AI applications in UAV systems.

I-A Contributions

The contributions of this paper are summarized as follows:

  • First, we present an in-depth analysis of various LLM architectures, assessing their suitability and potential for integration within UAV systems. This evaluation helps to understand the capabilities and effectiveness of state-of-the-art LLM models for different UAV applications.

  • Then, we explore various LLM-based UAV architectures, providing a consolidated view of how UAV technologies evolve by integrating sophisticated AI models. Furthermore, areas that can further benefit from LLM integration are highlighted.

  • After that, we discuss enhancing spectral sensing and sharing capabilities in UAVs through LLM integration. The impact of LLM integration on UAVs for optimized spectrum sensing, data processing, and decision-making is presented.

  • Finally, we demonstrate how LLM integration with the UAV framework can extend the capabilities of UAVs in various sectors, including surveillance and reconnaissance, emergency response, delivery, and enhanced network connectivity during emergencies. Moreover, critical areas for future research essential for the successful and effective integration of LLMs with UAV systems are identified and highlighted.

I-B Related Surveys

TABLE I: Summary of the existing surveys.
Ref. Main Focus Key Findings
[36] LLM-based autonomous agents
Comprehensive analysis of construction, application, and
evaluation of LLM-based agents in diverse fields.
[37] Development of LLM-based AI agents
Role of LLMs in artificial general intelligence and introduced
a novel framework based on brain, perception, and action
components to enhance the performance of LLM in complex
environments.
[38] Aligning LLMs with human expectations
Focuses on the challenges such as misunderstanding, biased
outputs and factual inaccuracies in model-generated content.
It also provides an in-depth analysis of technologies that enhance
LLM alignments according to human expectations.
[39] Deployment challenges of LLMs
Explores model compression techniques such as quantization,
pruning, and distillation to improve LLM efficiency.
[40] Challenges of LLM
Discusses the challenges of LLM related to dataset management,
tokenizer reliance, and pre-training costs.
[41] Limitations of LLM
Focuses on operational challenges of LLM such as design flaws,
and inefficiencies of LLMs.
[30] GAI in UAV systems
Reviews GAI technologies for improving UAV communication,
networking, and security.
[22] GAI in UAV swarms
Surveys applications, challenges, and opportunities of GAI in
enhancing UAV swarm coordination and functionality.
[42]
Large-GenAI models
for future wireless networks
Investigated the potential of Large-GenAI models
to enhance future wireless networks by improving wireless
sensing and transmission.
[43] LLMs in wireless networks
Investigated the application of LLMs for developing advanced
signal-processing algorithms in wireless communication and
networking.
This paper LLM-integrated UAV systems
A comprehensive analysis of various LLM architectures for UAV
systems, exploring their integration potential and impact across
different applications, including enhanced spectrum sensing and
sharing for autonomous decision-making. It also discusses how
LLM integration can extend UAV capabilities in various sectors,
including surveillance, emergency response, and delivery, while
identifying critical areas for future research to optimize their
integration and effectiveness.

 

 

 

































The future holds promise for revolutionizing various domains; therefore, several recent review articles exist on the topic. For instance, [44, 45, 46] investigate LLMs architectures, [47, 48, 49, 50] present the overview of the training process, fine-tuning, logical reasoning, and related challenges to address their limitations for broad adoption of LLM-based systems across domains. In [36], a comprehensive analysis is provided on LLM-based autonomous agents, focusing on their construction, application, and evaluation. These agents, equipped with sophisticated natural language understanding and generation capabilities, operate without human intervention. They interact with environments and users in complex ways, necessitating the integration of advanced AI techniques for tasks like communication and problem-solving across various domains such as social science, natural science, and engineering. Another work [37] delves into the development and use of LLM-based AI agents, emphasizing their role in advancing artificial general intelligence. LLMs are recognized as foundational for creating versatile AI agents due to their language capabilities, which are crucial for various autonomous tasks. The authors propose a framework based on brain, perception, and action components to enhance agents’ performance in complex environments. In [38], challenges and advancements in aligning LLMs with human expectations are critically examined. Concerns like misunderstanding instructions and biased outputs are addressed through technologies enhancing LLM alignment. Data collection strategies, training methodologies, and model evaluation techniques are explored to improve performance in understanding and generating human-like responses. Another study [39] investigates challenges in deploying LLMs, particularly in resource-constraint settings. Model compression techniques like quantization, pruning, and knowledge distillation are discussed to improve efficiency and applicability. While [40, 41] investigate challenges of LLMs, including vast dataset management and high costs, they point out limitations that cannot be overcome merely by increasing model size. [30] explores Generative Artificial Intelligence (GAI) applications in improving UAV communication, networking, and security performances. A GAI framework is introduced to advance UAV networking capabilities. [22] surveys the challenges of UAV swarms in dynamic environments, discussing various GAI techniques for enhancing coordination and functionality. In [42], the authors explore the potential of Large-GenAI models to enhance future wireless networks by improving wireless sensing and transmission. They highlight the benefits of these models, including enhanced efficiency, reduced training requirements, and improved network management. In another study [43], the authors investigated the application of LLMs for developing advanced signal-processing algorithms in wireless communication and networking. They explored the potential and challenges of using LLMs to generate hardware description language code for complex tasks, focusing on code refactoring, reuse, and validation through software-defined radios. This approach led to significant productivity improvements and reduced computational challenges. Although [30, 22] focus broadly on GAI, the specific application of LLMs in UAV communication systems still needs to be explored. This gap highlights an area poised for investigation. Table I summarizes existing surveys’ primary focus and critical findings.

I-C Organization

The rest of the paper is organized as follows. In Section II, we present an overview of LLMs, introducing foundational concepts and developments in this area. Section III is dedicated to exploring LLMs for UAVs, where we discuss the integration and adaptation of LLM technologies within UAV systems. Section IV focuses on network architectures for LLMs in UAV communications, examining the structural designs that support LLM functionalities in UAV networks. Section V addresses spectrum management and regulation for LLMs in UAVs. Section VI explores applications and use cases of LLMs in UAV communications, outlining practical implementations and the benefits derived from these technologies. Section VII examines the challenges and considerations in implementing LLM-Integrated UAVs, discussing potential obstacles and operational considerations. Section VIII is dedicated to future directions and research opportunities, presenting potential areas for further exploration and development in the LLMs in UAVs. Finally, Section IX concludes the paper with a summary of our findings and reflections on the broader implications of our research.

II Overview of Large Language Models (LLMs)

LLMs undergo extensive data training, demanding high computational power and the integration of ML algorithms and deep neural network architectures to enable sophisticated language processing abilities [51]. This training encompasses diverse and extensive datasets, including text from various sources such as books, websites, and articles, aiding the model in learning language structure, vocabulary, grammar, and contextual nuances [51]. Integrating ML algorithms, LLMs interpret data, adjusting their learning process to language nuances and refining prediction and response generation [52]. Typically based on transformer architecture, LLMs utilize self-attention mechanisms, simultaneously enabling parallel processing of every word in a sentence. This capability allows a comprehensive understanding of context, significantly enhancing the model’s ability to manage long-range dependencies in text, thereby improving context awareness [53]. LLMs consist of multiple neural network layers, each performing complex computations. These layers process inputs sequentially, refining information progressively. With parameters ranging from millions to billions, they dictate how input data is transformed into outputs. During training, these parameters are adjusted to minimize prediction errors, progressively improving the model’s ability to recognize complex patterns and relationships within the data [54].

The specific architecture of LLMs, based on transformer architecture, incorporates attention mechanisms to weigh the importance of different parts of input data, essential for tasks like language understanding [55]. Positional encodings maintain word order, which is crucial for sequential natural language. Training involves adjusting parameters through gradient descent and backpropagation. Backpropagation calculates gradients of the loss function for each parameter, guiding parameter updates to enhance model performance. Gradient descent iteratively adjusts parameters to minimize prediction error. Through this process, LLMs continually learn from mistakes, generating coherent and contextually appropriate responses and improving their ability to handle complex language tasks. Fig. 1 illustrates the typical layers involved, showing how input text is transformed through multiple processing stages to produce output.

Refer to caption
Figure 1: A general architecture of LLM incorporating input layer, embedding layer, transformer block, and output layer.

The broad landscape of LLMs in based on several essential models having unique characteristics tailored to specific linguistic tasks. For example, BERT [20, 56] revolutionized the field by using a bidirectional training of transformers to improve context understanding, making it exceptionally practical for tasks such as question answering and language inference. Since its introduction, BERT has been widely adopted for improving search engine results, enabling more nuanced and context-aware responses to queries. Google has incorporated BERT into its search algorithms, significantly enhancing search accuracy by better understanding the intent behind users’ queries [57]. Moreover, Enhanced Representation through Knowledge Integration (ERNIE) [58] extends BERT by incorporating structured knowledge, such as entity concepts, to enhance language comprehension. ERNIE is used for enhancing language understanding tasks in Chinese, showcasing superior performance in language inference and named entity recognition. It demonstrates the importance of integrating structured knowledge into pre-trained language models, especially for languages with complex structures like Chinese.

While the GPT series [59, 60] focuses on generating human-like text by predicting the next word in a sequence, showcasing remarkable proficiency in various generative tasks. The GPT models, particularly the later versions such as GPT-3, have been instrumental in creating advanced chatbots and virtual assistants. These models have been used to generate creative content, from writing articles to composing poetry, demonstrating their versatility in handling various innovative and conversational tasks.

On the other hand, T5 [21] adopts a unified framework that converts all text-based language problems into a text-to-text format, simplifying the process of training and applying transformers. T5 has been effectively used in summarization, translation, and classification tasks. Its flexible text-to-text approach can be fine-tuned for various language tasks without significantly changing the underlying model architecture [61].

While eXtreme Learning NETwork (XLNet) introduces permutation-based training, which better captures the bidirectional context by predicting all tokens in a sequence rather than one at a time. XLNet’s permutation-based training strategy allows it to outperform models like BERT on tasks requiring a deep understanding of context order, such as question answering and document ranking, making it particularly useful in academic and professional settings where the precise interpretation of information is crucial [62].

Lastly, Bidirectional and Auto-Regressive Transformers (BART) combine an autoencoder and autoregressive approach, making it highly effective for sequence-to-sequence tasks and text generation. It has been used extensively in text summarization applications as it generates concise summaries of long documents without losing critical information. It is particularly useful in law and medicine files where extracting accurate information quickly from large texts is vital. The next section details these models’ working and highlights their opportunities for UAVs [63].

Some less common LLM models have also been integrated into UAVs for innovative applications. In [64], the authors employed UAVs to handle complex tasks involving semantically rich scene understanding by integrating LLMs and Visual Language Models (VLMs). The approach allows UAVs to provide zero-shot literary text descriptions of scenes, which are both instant and data-rich, suitable for applications ranging from the film industry to theme park experiences and advertising. The descriptions generated by this system achieve a high readability score, highlighting the system’s ability to develop highly readable and detailed descriptions, demonstrating the practical use of microdrones in challenging environments for efficient and cost-effective scene interpretation. The authors in [65] integrated LLMs in UAV control by developing a multimodal evaluation dataset that combines text commands with associated utterances and relevant images, enhancing UAV command interpretation. The presented study evaluates the effectiveness of generic versus domain-specific speech recognition systems, adapted with varying data volumes, to optimize command accuracy. It also innovatively integrates visual information into the language model, using semi-automatic methods to link commands with images, providing a richer context for command execution.

In another work [66], the authors focused on enhancing human-drone interaction by implementing a multilingual speech recognition system to control UAVs. This work addresses the complexities and training challenges associated with traditional RF remote controls and ground control stations using natural, user-intuitive interfaces that recognize speech in English, Arabic, and Amazigh. The study developed a two-stage approach by initially creating a deep learning-based model for multilingual speech recognition, which was then implemented in a real-world setting using a quadrotor UAV. The model was trained on extensive records, including commands and unknown words mixed with background noise, to enhance its robustness for controlling drones across linguistic backgrounds.

In [67], the authors explore the application of LLMs to enhance the resilience and adaptability of autonomous multirotors through REsilience and Adaptation (REAL) using the LLMs approach. REAL integrates LLMs into robots’ mission planning and control frameworks, leveraging their extensive capabilities in long-term reasoning, natural language comprehension, and prior knowledge extraction. REAL utilizes LLMs to enhance robot resilience in facing novel or challenging scenarios, interpret natural language and log data for better mission planning, and adjust control inputs based on minimal user-provided information about robot dynamics. Experimental results demonstrate that this integration effectively reduces position-tracking errors under scenarios of controller parameter errors and unmodeled dynamics. Furthermore, it enables the multirotor to make decisions that avoid potentially dangerous situations not previously considered in the design phase, showcasing LLMs’ practical benefits and potential in improving autonomous UAV operations.

III LLMs for UAVs

Due to the growing interest in LLM-integrated UAV systems across various applications, several recent work exists. For instance, in [68], the authors introduced a vision-based autonomous planning system for quadrotor UAVs to enhance safety. The system predicts trajectories of dynamic obstacles and generates safer flight paths using NanoDet for precise obstacle detection and Kalman Filtering for accurate motion estimation. Additionally, the system incorporates LLMs such as GPT-3 and ChatGPT to facilitate more intuitive human-UAV interactions. These LLMs enable Natural Language Processing (NLP), allowing users to control UAVs through simple language commands without requiring complex programming knowledge. They translate user instructions into executable code, enabling UAVs to execute tasks and provide feedback in natural language, simplifying the control process. UAVs can operate in ad-hoc and mesh fashion to form dynamic networks without relying on pre-existing infrastructure. This makes them particularly valuable when establishing permanent network infrastructure is impractical, such as disaster response, military operations, or environmental monitoring. Both ad-hoc and mesh networks enhance UAVs’ ability to configure and maintain connectivity as they move automatically. They continuously discover new neighbors and can adjust routes based on the network’s topology and traffic conditions, improving scalability and flexibility [69]. Integrating LLMs into UAV communications enhances their ability to understand network conditions and generate insights based on the networks’ characteristics, thereby highlighting their adaptability and responsiveness to quickly adapt to changing environmental conditions and operational demands. LLMs also help UAVs understand network traffic patterns for recommending adaptive protocols that reduce latency and increase throughput, particularly in the variable conditions common in these networks. They also assist in simulating or modeling the behavior of the networks under various scenarios, helping in planning and decision-making processes for UAV deployments. Therefore, the LLM incorporation can enhance data analysis, improving data exchange efficiency between UAVs. LLMs, with their capability to process and learn from vast amounts of data, enable UAVs to make informed decisions about route planning, data forwarding, and network configuration. For instance, in response to UAV failures or environmental obstacles, LLMs can swiftly calculate alternative routes or reconfigure the network to sustain connectivity and performance. Moreover, LLMs foster higher levels of autonomy in UAVs by equipping them with advanced cognitive capabilities, allowing UAVs to understand and execute complex commands and interact more naturally with human operators or other autonomous systems.

Furthermore, LLMs can analyze data from UAVs (such as operational logs and telemetry data) to predict potential failures or maintenance needs before they occur. This predictive capability can significantly increase the reliability and lifespan of UAVs, reducing downtime and maintenance costs. Security is also a paramount concern in decentralized ad-hoc networks; LLMs can enhance security protocols by identifying potential threats through pattern recognition and anomaly detection and simulating attack scenarios to develop more robust security measures. LLMs can also optimize allocating critical resources such as bandwidth and power within the UAV network. LLMs dynamically allocate resources by understanding and predicting network demands to maximize efficiency and extend UAV operational times. They improve the interface between human operators and the UAV network, providing more intuitive control and feedback systems, including generating natural language reports on network status or translating complex network data into actionable insights for decision-makers. In addition, LLMs address scalability challenges inherent in ad-hoc networks. They dynamically adjust network protocols and configurations as the number of UAVs changes, ensuring the network remains stable and efficient regardless of size. By integrating LLM capabilities, UAV ad-hoc networks can become more intelligent, responsive, and efficient, significantly enhancing their effectiveness across various applications.

This section provides a detailed overview of different LLMs and discusses the opportunities they bring for UAV-based communication systems.

III-A BERT-enabled UAVs

As discussed in the previous section, BERT is an influential model in NLP, developed by researchers at Google and released in 2018 [20]. BERT’s development represented a turning point in NLP, providing a more nuanced and effective way for machines to process and understand human language by fully leveraging the context surrounding each word. BERT employed pre-training and fine-tuning stages. In pre-training, the model is trained on a large corpus of text with tasks designed to help it learn general language patterns. These tasks include predicting masked words in a sentence (i.e., Masked Language Model (MLM)) and predicting whether two sentences logically follow each other (i.e., Next Sentence Prediction (NSP)). After pre-training, BERT is fine-tuned with additional data tailored to specific tasks (i.e., question answering or sentiment analysis) [70, 71].

The introduction of BERT significantly advanced the state of the art in a wide range of NLP tasks. It showed remarkable performance improvements on leaderboards for tasks such as named entity recognition [72, 73], sentiment analysis [74, 75], and especially question answering and natural language inference, where the full-sentence context from both directions can be crucial for understanding subtleties. In addition, BERT has inspired numerous variations and improvements, leading to the development of different models such as Robustly optimized BERT approach (RoBERTa) [76], Distilled Bidirectional Encoder Representations from Transformers (DistilBERT) [77], and A Lite BERT (ALBERT) [78] that use the original architecture and training procedures of BERT to optimize other factors such as training speed, model size, or enhanced performance.

Integrating BERT can significantly enhance UAV performance across various domains. For instance, in emergency response scenarios, BERT can help UAVs understand complex natural language commands from disaster management teams. Moreover, BERT can interpret and summarize information from UAV sensors and reports, making it particularly valuable in surveillance missions where quick summarization of extensive video data is essential. Moreover, BERT’s ability to rapidly analyze and interpret data from multiple sources enables timely, informed decisions, crucial in environmental monitoring for assessing conditions like forest fires or pollution. Furthermore, its proficiency in parsing and understanding commands ensures precise coordination among multiple UAVs, which is critical for complex logistics operations involving supply delivery in challenging environments. Recently, in [79], the authors introduced an innovative end-to-end Language Model-based fine-grained Address Resolution framework (LMAR) explicitly designed to enhance UAV delivery systems. Traditional address resolution systems rely primarily on user-provided Point of Interest (POI) information, often lacking the necessary precision for accurate deliveries. To address this, LMAR employs a language model to process and refine user-input text data, improving data handling and regularization with enhanced accuracy and efficiency in UAV deliveries. In another work [80, 81], the authors develop enhanced security and forensic analysis protocols for UAVs to support increased drone usage across various sectors, including those vulnerable to criminal misuse. They introduce a named entity recognition system to extract information from drone flight logs. This system utilizes fine-tuned BERT and DistilBERT models with annotated data, significantly improving the identification of relevant entities crucial for forensic investigations of drone-related incidents. The authors in [82] focused on enhancing the target recognition capabilities of UAVs in intelligent warfare by constructing a standardized knowledge graph from large-scale, unstructured UAV data. The authors introduced a two-stage knowledge extraction model with an integrated BERT pre-trained language model to generate character feature encoding, which enhances the efficiency and accuracy of information extraction for future UAV systems.

III-B GPT-enabled UAVs

GPT series developed by OpenAI represents a significant evolution in the design and capabilities of LLMs that enhance various natural language processing tasks such as text generation, translation, summarization, and question answering [83]. The first architecture, GPT-1, was introduced in June 2018, and it was based on the transformer model that uses a stack of decoder blocks from transformer architecture. GPT-1 was pre-trained on a language modeling task (predicting the next word in a sentence) using the BooksCorpus dataset, which comprises over 7,000 unique unpublished books (totaling around 800 million words). After this initial pre-training, supervised learning was fine-tuned for specific tasks [19, 83].

The GPT-2 was released in February 2019 and expanded significantly on its predecessor, featuring up to 48 layers in its largest version, with 1,600 hidden units, 48 attention heads, and 1.5 billion parameters [84]. GPT-2 used a WebText dataset created by scraping web pages linked from Reddit posts with at least three upvotes. This resulted in a diverse dataset of around 40GB of text data. GPT-2 continued using the unsupervised learning approach, leveraging only language modeling for pre-training without task-specific fine-tuning. This demonstrated the model’s ability to generalize from language understanding to specific tasks [85]. While GPT-3 released in June 2020 is one of the largest AI language models ever created, with 175 billion parameters. It includes 96 layers, with 12,288 hidden units and 96 attention heads [86]. It was trained on an even more extensive and diverse dataset, including a mixture of licensed data, data created by human trainers, and publicly available data, significantly larger than GPT-2. GPT-2 and GPT-3 used an unsupervised learning model, demonstrating exceptional capabilities in learning from large datasets [87]. GPT-4, built on an advanced transformer-style architecture, significantly expands in size and complexity compared to its predecessors, GPT-2 and GPT-3. This model has been fine-tuned using Reinforcement Learning from Human Feedback and employs publicly available internet data and data licensed from third-party providers. However, specific details related to the architecture, such as the model size, hardware specifications, computational resources used for training, dataset construction, and training methodology, have not been publicly disclosed [88].

GPT series in UAVs represents an innovative intersection of AI and drone technology that can enhance UAVs’ functionality, autonomy, and interaction capabilities in a broad range spanning from enhanced control systems to fully autonomous task execution [89, 90]. For example, GPT series integration allows UAVs to execute instructions provided in plain language with a high level of proficiency. For instance, the operator command to inspect the condition of a bridge at specific coordinates will instruct the UAV to devise a flight path and carry out all the necessary steps for bridge inspection without requiring manual inputs for each step. Similarly, it can generate detailed reports based on data collected during flights, and integration of these models with the UAV’s sensors and data collection systems can automatically generate textual descriptions highlighting various aspects, such as mission outcomes and anomalies detected [91]. Accordingly, it can make it easier for human operators to understand what the UAV has observed without reviewing extensive raw data. For instance, Tazir et al. in [89] integrated LLM system OpenAI’s GPT-3.5-Turbo with UAV simulation systems (i.e., PX4/Gazebo simulator) to create a natural language-based drone control system. The system architecture is designed to allow seamless interaction between the user and the UAV simulator through a chatbot interface facilitated by a Python-based middleware. The Python middleware is the core component, establishing a communication channel between the chatbot (GPT-3.5-Turbo) and the PX4/Gazebo simulator. It processes natural language inputs from the user, forwards these inputs to the ChatGPT model using the OpenAI API, retrieves the generated responses, and converts them into commands that the simulator understands. ChatGPT provides guidance and support through PX4 commands and explanations, thus enhancing the interactivity and accessibility of UAV simulation systems. It also facilitates the control and management of UAVs through sophisticated AI-driven interfaces. In another work [92], the authors integrated advanced GPT models and dense captioning technologies into autonomous UAVs to enhance their functionality in indoor inspection environments. The proposed system enables UAVs to understand and respond to natural language commands like humans, increasing their accessibility and ease of use for operators without advanced technical skills. The UAVs’ dense captioning models facilitate this human-like interaction by analyzing images captured during flight to generate detailed object dictionaries. These dictionaries allow the UAVs to recognize and understand various elements within their environment, dynamically adapting their behavior in response to both expected and unexpected conditions, thus improving the efficiency and accuracy of UAVs for indoor inspections across various environmental conditions and applications.

Furthermore, in dynamic or complex environments requiring fast decisions, GPT series can assist by processing real-time data and communications, providing suggestions or automated decisions based on the data. For example, a search and rescue operation can analyze live video feeds and text reports from multiple UAVs, synthesize the information, and recommend areas to focus on or adjust search patterns [29]. It can also play a crucial role in enhanced collaborative UAV-to-UAV communication by establishing a decentralized swarm intelligence system where UAVs can share information and make group decisions. For instance, UAVs can use natural language to report their status and findings to each other, coordinate their actions based on shared goals, and optimize task distribution among the group without constant human intervention [93]. GPT series can also simulate various communication scenarios for UAV’s training by generating realistic mission scenarios and responses, providing operators with robust training on handling different situations to enhance their response for real-world operations [91].

III-C Text-to-Text Transfer Transformer (T5) for UAVs

Google introduced the T5 model in October 2019 and adopted a novel and streamlined approach to handling various NLP tasks by reframing them as text-to-text problems [94]. Unlike traditional models that require different architectures for different tasks and produce varied outputs, T5 standardizes the input and output across all tasks [95]. Each NLP task, such as translation, summarization, question answering, or text classification, is treated as generating new text from a given text. Consequently, T5 employs a single consistent model architecture for all tasks. This simplification streamlines both the model training and deployment pipeline, as the same model can be trained across multiple tasks with minimal architectural modifications [21]. For example, in a translation task where the input is English text, and the output is French text, both are treated merely as sequences of words. T5 is pre-trained on a large corpus of text in a self-supervised manner, primarily using a variation of the masked language model task, similar to BERT. This pre-training equips the model to understand and generate natural language effectively. Following this, T5 is fine-tuned on specific tasks by adjusting its training data to suit the text-to-text format. The versatility of T5 makes it suitable for a broad range of applications, including language translation, document summarization, and sentiment analysis, where it interprets text sentiment by producing descriptive tags. It also excels in question answering by generating appropriate textual answers [96].

UAVs can integrate the T5 framework to enhance the efficiency of UAV operations. Similar to GPT and BERT, T5 also improves command interpretation and response generation for UAVs, where complex commands issued by operators in natural language are interpreted and converted into executable instructions for the UAV. T5 is also proficient in generating comprehensive mission reports based on the data collected by the UAV by summarizing key findings, highlighting anomalies, and describing the surveyed area for environmental monitoring or disaster response applications. Moreover, T5 can perform real-time operations by processing data streams from UAV sensors and cameras, providing immediate, actionable insights. For instance, during a search and rescue operation, T5 could quickly summarize visual and sensor data to describe potential areas of interest or hazards, helping guide rescue efforts more effectively. At the same time, T5 can significantly enhance the performance of coordinated UAV missions by interpreting messages from one UAV and generating appropriate responses or commands for others, facilitating seamless teamwork for various applications ranging from managing flight patterns to avoiding collisions or coordinating timings for area surveillance.

T5-assisted UAV communication also enables automated troubleshooting and feedback, such as if a UAV encounters issues or anomalies during its operation, it can help by interpreting error messages or sensor data and generating troubleshooting steps or advice in natural language. This can also extend to providing real-time feedback to operators on mission progress or suggesting adjustments to improve operational efficiency. In addition, T5 can generate simulated mission scenarios and dialogues based on historical data or potential future situations for training purposes.

III-D eXtreme Learning NETwork (XLNet) for UAVs

XLNet is an advanced NLP model developed jointly by researchers from Google and Carnegie Mellon University [62]. Unlike BERT, which employs an MLM approach (i.e., where some words in a sentence are randomly masked and predicted), XLNet uses a permutation-based training strategy. This approach considers all possible permutations of the words in a sentence during training, enabling the model to predict a target word based on all potential contexts provided by other words before and after it. This method significantly enhances the flexibility and depth of contextual understanding. In addition, the permutation-based training allows XLNet to capture a richer understanding of language context, unlike BERT, which focuses only on predicting masked words and might miss contextual nuances [97, 98].

Furthermore, XLNet avoids the discrepancies between pre-training and fine-tuning phases seen in BERT by not relying on word masking during training, leading to more consistent behavior across different operational phases. XLNet also merges strategies from autoregressive language modeling (e.g., GPT series) and autoencoding (e.g., BERT) by training autoregressively without adhering to a fixed sequence order. Instead, it predicts words based on varied permutations, enhancing its comprehension and generation capabilities [99]. Consequently, XLNet has demonstrated superior performance on several NLP tasks, including question answering, natural language inference, and document ranking, by effectively utilizing complete sentence structures for a deeper and more accurate contextual understanding [100, 101, 102].

Integrating XLNet to UAVs can offer unique advantages due to its sophisticated approach to language processing [103, 104]. XLNet’s permutation-based training enables a more nuanced and comprehensive understanding of context, making it particularly effective for interpreting complex instructions or environmental data where the context can vary significantly. For instance, during search and rescue missions, where the operational environment is complex and dynamic, XLNet can provide more reliable interpretations of context-heavy commands in real time. Similarly, since XLNet considers all permutations of input data, it can be more robust against noisy or incomplete inputs, which are common in real-world UAV tasks. This feature is especially beneficial in combat or disaster response scenarios where communications may be disrupted or incomplete. XLNet’s ability to contextually predict missing information can maintain the effectiveness of UAV operations.

III-E Enhanced Representation through kNowledge Integration (ERNIE) for UAVs

Baidu Research introduced ERNIE in June 2019 to integrate world knowledge into pre-trained language models [58]. It represents a significant evolution in language understanding due to its novel approach of integrating structured world knowledge into the training of language models. Unlike conventional models that rely on vast amounts of textual data to learn language patterns, ERNIE enhances these models by incorporating knowledge graphs into the training process. Knowledge graphs are structured databases that store information about the world so that machines can understand and process by including entities (such as people, places, and things) and their relationships.

ERNIE is trained on both traditional text corpora and knowledge graphs. Including knowledge graphs allow ERNIE to understand and represent complex relationships and attributes associated with various entities [105]. This training involves two key components: textual data and knowledge integration. Textual data is similar to other models like BERT or GPT, and ERNIE processes this vast amount of text to learn syntactic and semantic patterns of language. At the same time, the knowledge Integration component enables ERNIE to simultaneously learn from knowledge graphs, where it absorbs structured information about real-world entities and their interrelations. Thus, this process enables ERNIE to understand context from the linear text and a multi-dimensional perspective involving real-world facts and relationships. Integrating knowledge graphs offers ERNIE a deeper understanding of language semantics, as it can relate words and phrases to real-world entities and their properties. This capability allows it to perform better on tasks that require nuanced understanding, such as question answering and named entity recognition [106, 107].

In addition, ERNIE’s ability to draw upon external knowledge helps it provide contextually appropriate responses or analysis, especially when background knowledge about specific topics is crucial. It can also better handle ambiguity in language, as the additional data from knowledge graphs provides clarity on potentially confusing or unclear text based on the broader context of the entities involved [108]. The applications of ERNIE are broad and impactful, particularly in areas where deeper understanding and contextual awareness are necessary. For example, ERNIE can leverage its integrated knowledge base to answer complex questions that require understanding beyond the text, such as historical facts or specific details about people or places. It also improves the performance of semantic search engines by understanding the deeper meanings of queries regarding the knowledge it has learned, offering more relevant and precise answers.

ERNIE’s unique capability to integrate structured world knowledge from knowledge graphs with textual data can substantially benefit UAV communication. For example, ERNIE can interpret complex, context-dependent commands operators issue more effectively than traditional language models. For instance, if an operator gives a command involving geographic or operational terms, ERNIE’s integration of knowledge graphs allows it to understand and execute the command more accurately. This is crucial during complex missions in unfamiliar territories where a precise understanding of local geography and terms is necessary. ERNIE also demonstrates effective autonomous decision-making capabilities based on environmental data and mission objectives, as it can process both the current mission data and integrated knowledge to make informed decisions. For example, in environmental monitoring, ERNIE could identify specific features or anomalies in the landscape based on its broader understanding of environmental science, aiding in more effective data collection and analysis.

ERNIE also exhibits high real-time situational awareness attributes during critical missions such as search and rescue or disaster response, where ERNIE can apply its semantic understanding to interpret real-time data inputs (e.g., visual or sensor data) against its knowledge graph. This helps identify relevant entities or situations quickly, such as recognizing areas historically known to be hazardous or interpreting signs of human activities in remote sensing data. In multiple UAV scenarios, ERNIE can facilitate better communication and coordination by understanding and managing the information exchange between UAVs. It can interpret and prioritize communication-based on the relevance and urgency related to the mission’s objectives, using its semantic understanding to ensure that UAVs operate harmoniously.

Furthermore, in the context of training purposes, ERNIE can generate context-rich simulation scenarios that incorporate real-world knowledge into training exercises that assist in developing a better understanding of how to interact with UAVs in complex scenarios, enhancing their preparedness for real-world operations. Similarly to other LLMs, after mission completion, ERNIE can assist in generating detailed incident reports and debriefings that include observational data and contextual insights based on the integrated knowledge to provide a semantic analysis of the mission outcomes. Accordingly, by leveraging its ability to integrate and utilize extensive knowledge graphs alongside textual data, ERNIE can significantly enhance the capabilities of UAV communication systems, making them more intelligent, responsive, and effective in complex operational environments. This makes ERNIE particularly valuable for advanced UAV applications where conventional language models might fail to understand and process complex contextual information.

III-F Bidirectional and Auto-Regressive Transformers (BART) for UAVs

Facebook developed BART that combines the strengths of both auto-encoding and auto-regressive techniques within the transformer framework, making it exceptionally effective for sequence-to-sequence tasks [109]. Unlike BERT, which is primarily designed for understanding and predicting elements within the same input text, BART is optimized for tasks requiring text generation or transformation. It is trained by corrupting text with various noising functions, such as token masking, text infilling, and learning to reconstruct the original text [110, 70]. The BART training equips it to handle a wide range of applications, including text summarization, where it can generate concise versions of longer documents, and text generation, suitable for creating content or generating dialogue. In addition, BART’s capabilities extend to machine translation and data augmentation, making it a versatile tool for transforming input text into coherent and contextually appropriate output sequences [111].

BART integration into UAVs offers several advantages, particularly in tasks that involve complex text processing and generation. For example, BART can enhance the formulation and interpretation of mission reports, automatically generating concise summaries from extensive surveillance data or sensor readings, thus aiding in quicker decision-making and briefing. BART is also proficient in generating coherent text sequences for automated responses or instructions to UAV operators, specifically in scenarios requiring fast and accurate communication.

Furthermore, BART can improve real-time strategy adjustments during search and rescue operations to interpret incoming data and provide updated mission objectives or directions based on the evolving scenario. It also can transform noisy, incomplete textual data into intelligible information, making it particularly valuable in dynamic and challenging environments for UAV operations to ensure that communications remain clear and contextually relevant despite the complexities involved.

III-G Comparison of different LLMs for UAVs

The comparison of different LLMs (i.e., BERT, GPT, T5, XLNet, ERNIE, and BART) for UAVs reveals distinct capabilities tailored to various aspects of UAV operations, reflecting their unique architectures and training approaches. For instance, BERT excels at understanding context from both directions around a word, making it highly effective for interpreting complex commands and extracting relevant information from mission data. It is particularly suited for tasks where a precise understanding of sensor data or operational directives is critical, such as in surveillance or monitoring missions where deep contextual knowledge is crucial. In contrast, GPT specializes in generating coherent and extended text outputs, which are beneficial for creating detailed mission reports or conducting dialogues. This model is ideal for UAV training simulations that require narrative-style updates or interactive communications for generating operational logs or debriefing reports.

Whereas T5 exhibits high versatility and converts any text-based task into a text-to-text format, simplifying the processing of diverse types of communication. It proves effective in UAV communication tasks such as translating communications between different languages or protocols, summarizing extensive exploration data, and transforming raw sensor outputs into actionable text formats. On the other hand, XLNet employs permutation-based training and understands language context more flexibly and comprehensively than BERT. This model is helpful in complex, dynamic operational environments such as search, rescue, and disaster response, where interpreting and responding to context-heavy instructions in real-time is essential.

Similarly, ERNIE enhances the semantic understanding of language by integrating external knowledge through knowledge graphs, making it well-suited for missions that require a deep understanding of specific terminologies or concepts, such as environmental monitoring applications that involve specific ecological data. While BART compromises the benefits of auto-encoding and auto-regressive models, it excels at understanding and generating text. It is ideal for developing precise and contextually accurate mission reports or instructions for summarizing detailed surveillance data, where maintaining the integrity of information and its concise presentation are crucial.

Therefore, in conclusion, BERT and XLNet are highly effective at understanding commands due to their profound contextual understanding, with XLNet providing additional flexibility in dynamic contexts. Meanwhile, GPT and BART excel at creating coherent, extensive texts, with BART offering additional capabilities in text transformation tasks. T5 offers broad applicability across text transformation tasks, making it versatile for various communication needs. ERNIE stands out in scenarios where integrating specialized knowledge is essential for accurate operation and decision-making. Accordingly, each model can be incorporated based on the UAV mission’s specific requirements to ensure that communication remains effective and efficient, tailored to the complexities and challenges of UAV operations. Table II highlights various LLM models, including their key features, applications in the UAV domain, and challenges for integration in UAV systems.

TABLE II: Overview of different state-of-the-art LLM models for UAVs.
 

IV Network Architectures for Integrating LLMs in UAVs

Integrating LLMs with UAVs involves deploying advanced language processing capabilities to enable sophisticated decision-making and interaction abilities. The UAV platform consists of essential hardware, including the UAV itself equipped with flight control hardware, sensors like cameras and LiDAR, and communication modules such as Wi-Fi, LTE, and satellite. It also includes small-scale onboard computers for real-time data processing. In LLM integration, lightweight versions of LLMs are directly deployed on UAVs for rapid and autonomous decision-making through Edge AI. For more complex computations, UAV data is transmitted to cloud servers where more robust LLMs perform analyses, and the results are then sent back to the UAV. Ground control stations support these operations, allowing operators to monitor and control UAVs remotely via direct line-of-sight or satellite communication, using secure data links for data transmission. The operation of this system involves several key functions. UAVs collect data via onboard sensors, capturing visual imagery, environmental data, or specific readings relevant to their missions. This data is either processed locally or sent to a ground station or cloud server, depending on the task’s complexity and onboard processing unit capabilities. The embedded LLM on the UAV processes data for simple tasks to make real-time decisions. For more complex decisions, data is sent to the cloud, where powerful LLMs analyze it, make decisions, or generate insights, which are then transmitted back to the UAV. Based on this processed data and decisions made by the LLMs, the UAV executes actions such as optimizing its flight path, interacting with the environment, or performing specific tasks like delivery, surveillance, or data collection. Feedback and learning are integral to this system, where data from missions are used to retrain or refine LLMs, improving their accuracy and decision-making capabilities. This continuous feedback loop helps the model adapt to specific environments for optimal task performance. The integration of LLMs with UAVs thus offers significant enhancements to UAV operations, opening up vast possibilities for improved capabilities and effectiveness.

Fig. 2 illustrates the comprehensive architecture of a UAV system integrated with an LLM, where the UAV collects data from sensors. This data, encompassing various types such as text, audio, and video, is input into the integrated LLM architecture. The LLM processes this data and outputs results to the decision layer, which then issues commands to operational components, including the flight controller, sensor systems, energy systems, and payload management systems.

Refer to caption
Figure 2: Comprehensive architecture of LLM-integrated UAV systems.

Besides, ground control and base stations are key elements in the UAV operations infrastructure, serving as command and control centers that handle everything from flight authorization and monitoring to data processing and deployment management. Integrating LLMs with ground control and base stations significantly enhances UAV management and operation. For example, LLMs can significantly improve communication between UAVs and their control stations by interpreting and processing natural language commands or queries. It allows operators to interact with UAVs more intuitively, making complex commands simpler to execute and reducing the potential for human error.

LLMs can process real-time data received from UAVs at the ground control stations to make instant decisions regarding flight paths, mission adjustments, and responses to changing environmental conditions. LLMs can also analyze vast quantities of data much faster than humans, providing critical insights that enable quick decision-making to optimize UAV operations and ensure mission success. In addition, LLMs can utilize historical and real-time data to predict potential issues before they arise, such as mechanical failures, battery depletion, or adverse weather conditions. This predictive capability ensures that preventative measures can be taken in advance, enhancing the safety and reliability of UAV operations.

Furthermore, LLMs can automate routine tasks such as flight scheduling, monitoring UAV conditions, and managing data collection to improve efficiency for complex decision-making and operational strategy. LLMs can also signficantly contribute to improved data handling and analyzing by automatically categorizing data, extracting relevant information, and generating comprehensive reports. Moreover, they can analyze imagery and sensor data to identify patterns or anomalies, aiding in missions such as surveillance, environmental monitoring, and infrastructure inspection. LLMs can create detailed simulations and training scenarios based on accumulated data, providing operators with realistic and varied training experiences and improving the skills of UAV operators to ensure they are better prepared for complex operational scenarios.

Moreover, with their advanced pattern recognition capabilities, LLMs integrated into ground and base stations can enhance security protocols. They can detect potential cyber threats and unauthorized access attempts, ensuring UAV operations are protected against digital intrusions. In addition, LLMs can optimize resource allocation by predicting the best use of available UAVs and support equipment based on mission requirements. LLMs can also facilitate better interoperability between systems and software used in UAV operations to ensure seamless integration and communication across diverse platforms by acting as a bridge that understands and translates between various data formats and protocols. Thus, leading to efficient management, superior decision-making support, enhanced safety, and improved effectiveness of UAV missions. This broad application of LLMs lays the groundwork for their targeted use in enhancing spectrum sensing capabilities.

Moreover, given the pivotal role of spectrum sensing in ensuring effective Radio Frequency (RF) communication for UAVs, especially in complex or congested environments, the integration of LLMs proves immensely beneficial and can significantly enhance UAVs’ spectrum sensing capabilities through sophisticated data processing techniques. This integration deepens the understanding of dynamic RF conditions, which are prevalent in areas with shared frequencies or high interference levels and enables UAV systems to identify and utilize optimal frequency bands intelligently. Such capabilities drastically improve the reliability and efficiency of their communication networks, which are crucial for maintaining robust links and ensuring the successful execution of UAV operations in RF-dense environments where traditional methods might fail. Consequently, this survey highlights the critical need for LLM integration in spectrum sensing and thoroughly explores its opportunities and challenges in the subsequent section.

V Spectrum Management and Regulation in LLMs-assisted UAVs

UAVs depend on RF communication for various tasks, including remote control, telemetry, data transmission, and connectivity with ground stations. Spectrum sensing is a key technology that enhances UAVs’ RF communication capabilities by enabling them to identify and utilize appropriate frequency ranges crucial for their missions. Moreover, it is particularly critical in environments where UAVs share frequency bands or encounter rapidly changing RF conditions [112, 113]. Accordingly, by accurately sensing the spectrum, UAVs can dynamically adjust their communication parameters, such as channel selection and power control, to prevent interference with primary users and optimize their communication performance [114]. In addition, spectrum sensing enhances the operational efficiency of UAVs by enabling them to make informed decisions about frequency band selection, thereby ensuring efficient utilization of available spectrum resources and minimizing the risk of interference with existing wireless systems [115, 116].

Furthermore, spectrum sensing also plays an essential role in enabling cognitive radio capabilities [117, 112], dynamic spectrum access [118], interference avoidance [117], and ensuring regulatory compliance for UAV communication systems [119]. For example, cognitive radio allows UAV systems to adaptively select and switch between different frequency channels or bands based on real-time spectrum sensing results, enabling UAVs to locate and utilize the most suitable, least congested, and interference-free frequency bands for reliable and efficient communication [120]. Dynamic spectrum access allows UAVs to access available spectrum resources, dynamically ensuring that UAVs can avoid interference with existing users while optimizing their communication links. In addition, spectrum sensing facilitates coexistence and interference avoidance by enhancing UAVs’ capabilities in detecting the presence of other RF devices or systems in their vicinity. If interference or potential conflicts are detected, UAVs can autonomously or semi-autonomously change their operating frequency or adjust their communication protocols to avoid interference [121].

V-A Regulatory Frameworks and Compliance Considerations

Regulatory bodies worldwide, such as the Federal Communications Commission (FCC) in the United States, establish guidelines for spectrum use to ensure fair access and prevent conflicts among various technologies and services, including UAVs. These guidelines designate specific frequency bands for UAV use to avoid conflicts with commercial, residential, and emergency communications to balance the growing demand for UAV services with the needs of traditional spectrum users. These authorities have established rules for dynamic spectrum access, particularly in bands where UAVs share the spectrum with other devices. This framework involves protocols and technologies that enable UAVs to detect and utilize vacant frequencies without interfering with incumbent users. Compliance with these frameworks is essential for legal and efficient UAV operations.

To ensure compliance, UAV operators must consider several key aspects. For instance, UAVs must be equipped with advanced spectrum sensing technologies that reliably identify available and occupied channels to prevent unauthorized use of occupied frequencies. UAVs must also operate to minimize interference with other spectrum users, adhering to power limits, frequency boundaries, and operational protocols designed to mitigate the risk of signal interference [122]. Moreover, it is necessary to implement software solutions that help manage spectrum usage and ensure adherence to local and international regulations to automate many aspects of spectrum management, reducing the burden on UAV operators and decreasing the risk of non-compliance.

V-B Integrating LLMs in UAVs Spectrum Management

Recent research has significantly advanced the application of spectrum sensing and sharing in UAV operations, focusing on several key directions to enhance communication efficiency and mitigate interference. Shen et al. [123] introduced a 3D spatial-temporal sensing approach that leverages UAV mobility for dynamic spectrum opportunity detection in heterogeneous environments. The authors in [124] and [125, 126] developed methods to optimize spectrum sensing and sharing in cognitive radio systems, improving UAV communication performance by managing interference with ground links. Chen et al. [127] focused on spectrum access management among UAV clusters to reduce interference, while Xu et al. [128] focused on transmit power allocation and trajectory planning in UAV relay systems for effective data relay between devices.

In another work [129], Qiu et al. utilized blockchain technology to ensure privacy and efficiency in spectrum transactions between terrestrial and aerial systems. Hu et al. [130] focused on the strategies for spectrum allocation using contract theory to balance the interests of macro base stations and UAV operators. Azari et al. [131] compared underlay and overlay spectrum sharing mechanisms in densely populated urban scenarios, emphasizing the effectiveness of overlay strategies for maintaining service quality for both UAVs and ground users. While significant advancements have been made in spectrum sensing and sharing technologies for UAVs, the integration of LLMs has not been widely explored in the existing research. Integrating LLMs can revolutionize the UAV domain by enhancing spectrum sensing capabilities, enabling more dynamic and efficient use of communication frequencies [132]. LLMs can interpret and analyze the vast amounts of data generated by spectrum sensors on UAVs. With their advanced natural language processing capabilities, they can extract meaningful insights from unstructured data, facilitating intelligent decision-making in real time. LLMs can also predict spectrum availability and potential interference by analyzing historical data and current communication patterns. Thus, UAVs can proactively adjust their communication parameters, such as channel selection and power levels, to maintain optimal performance.

Furthermore, LLMs can process sensor data and identify patterns indicating potential frequency conflicts or areas of congestion. UAVs can then autonomously make adjustments to avoid these issues, enhancing operational efficiency and reducing the risk of communication failures. In addition, LLMs can contribute to cognitive radio enhancements by assisting in making more informed choices about frequency selection by providing a deeper analysis of spectrum conditions and user behavior. This integration enhances UAVs’ ability to select the least congested and most efficient channels. LLM’s continuous learning and adaptation ability can also optimize UAVs’ spectrum access strategies, ensuring they utilize the best available frequencies based on real-time data and sophisticated algorithms. LLMs can also ensure that UAVs operate within the legal spectrum allocations by continuously monitoring compliance parameters and adapting to regulatory changes. LLMs can also contribute significantly to interference management and adherence to regulatory frameworks by analyzing communication patterns and environmental data. LLMs can detect potential interference sources more accurately and suggest immediate corrective actions to avoid them.

VI Applications and Use Cases of LLMs in UAVs

VI-A Surveillance and Reconnaissance Applications

LLMs offer advanced cognitive and analytical abilities that can significantly enhance the efficiency, accuracy, and effectiveness of UAV surveillance systems [2]. With LLM integration, UAVs can process and analyze large volumes of visual data more efficiently, enabling real-time image recognition, object detection, and situation awareness. LLMs are exceptional at identifying specific objects, individuals, vehicles, or activities in video streams or images, providing detailed insights crucial for military and civilian surveillance operations. It also enables UAVs to operate more autonomously by interpreting and reacting to their surroundings without constant human oversight, significantly benefiting in complex or hostile environments with critical response times.

Moreover, UAVs equipped with LLMs can make real-time decisions about flight paths, areas to focus on, and when to capture critical footage based on the mission’s objectives and evolving ground realities. NLP allows UAVs to understand and process human language, enabling them to receive and interpret more complex commands and queries. In addition, LLMs can predict potential security threats or points of interest by analyzing patterns and historical data. This predictive capability allows for proactive surveillance measures, where UAVs can monitor suspected areas more closely or alert human operators about unusual activities or anomalies detected based on learned patterns. It can also enhance real-time decision support by processing and analyzing data on the fly by summarizing vast amounts of collected data into actionable intelligence. It enables quick and informed decisions crucial in surveillance and reconnaissance missions where conditions can change rapidly [133].

Refer to caption
Figure 3: Applications of LLM-integrated UAVs.

VI-B Emergency Response and Disaster Management

LLMs incorporation with UAVs for emergency response and disaster management can significantly enhance the efficiency, accuracy, and effectiveness of emergency operations. LLMs can quickly analyze images and sensor data collected by UAVs to assess the extent of damage immediately following a disaster, including identifying blocked roads, damaged buildings, and flood areas [134]. Due to real-time situational awareness, LLM equipped UAVs help emergency responders prioritize areas that require urgent attention and plan the most effective response [135].

While in search and rescue missions, time is critical, so UAVs with LLM capabilities can autonomously scan large areas, using object recognition and pattern detection to locate survivors. They can navigate challenging terrains without direct human guidance, speeding up search operations and improving the chances of rescuing distressed people. In addition, as disasters disrupt the communication networks, UAVs integrated with LLMs can establish temporary communication networks, acting as airborne relays to facilitate communication between responders and affected people. LLMs optimize the placement and routing of UAVs to ensure maximum coverage and network efficiency.

Furthermore, they can also improve disaster preparedness due to their predictive analytics abilities by analyzing historical data and current weather reports to predict potential disasters before they occur. This predictive capability allows authorities to deploy UAVs proactively, monitoring at-risk areas and initiating preemptive evacuations or other mitigation measures. Moreover, LLMs can manage logistical aspects by analyzing needs assessments and resource availabilities. They ensure that supplies such as food, water, and medical equipment are optimally distributed and delivered using UAVs, especially to areas that are hard to reach by traditional means due to the disaster [136].

Moreover, they can also play a crucial role in damage analysis and recovery planning by performing detailed assessments of the damage, tracking recovery progress, and analyzing data over time to guide rebuilding efforts. LLMs can model different recovery scenarios, helping planners make data-driven decisions to rebuild infrastructure more resiliently. Given the sensitivity of data involved in emergency response, LLMs ensure that all information collected and transmitted by UAVs is securely encrypted and protected from unauthorized access. They monitor data flows for anomalies that indicate threats, safeguarding critical information during chaotic situations. LLMs enhance interactions between UAVs and human operators by translating complex data into actionable insights and intuitive reports. This allows emergency responders to make informed decisions quickly and effectively based on comprehensive analyzes provided by UAVs in understandable formats.

VI-C Delivery Services and Logistics

LLMs-integrated UAVs can transform delivery services and logistics by optimizing routes, enhancing customer interactions, and improving operational efficiency [79]. LLMs can process complex datasets, including traffic patterns, weather conditions, and geographical data, to dynamically optimize delivery routes. This ensures faster delivery times and helps reduce operational costs. LLMs can adapt these routes in real-time to account for changing conditions, ensuring deliveries are made as efficiently as possible [137].

UAVs can interact with customers using LMMs to update the status of delivery, answer queries, and even handle complaints or special instructions in real time. Enhanced interaction increases customer satisfaction and streamlines the delivery process, reducing the need for human intervention in customer service. LLMs enable UAVs to make autonomous decisions during delivery missions. For example, when faced with unexpected obstacles or emergencies, UAVs can decide the best course of action, be it rerouting, waiting for clearance, or returning to the base. This level of autonomy ensures reliability and consistency in delivery services, even in unpredictable circumstances. In addition, their proactive approach prevents downtime, extends the lifespan of the UAV fleet, and ensures that technical issues do not disrupt delivery schedules.

Furthermore, LLMs can assist in weight distributions, package sizes, and delivery priorities to ensure that each UAV is loaded efficiently, maximizing delivery capacity and minimizing the number of trips required. LLMs continuously analyze traffic and weather data to adjust UAV flight plans in real-time to maintain delivery schedules, especially in adverse weather conditions or congested airspaces, ensuring that deliveries are made safely and on time.

VI-D Environmental Monitoring and Wildlife Conservation

LLMs can process and analyze vast amounts of environmental data collected by UAVs, such as images, temperature readings, and pollution levels. The collected data enable identifying environmental trends and anomalies, such as changes in vegetation cover, water quality, or the presence of pollutants [138, 139, 140, 141]. LLM can quickly analyze this data and provide actionable insights for conservationists and environmental scientists. LLMs can also help track and study wildlife by analyzing video and audio recordings captured by UAVs to identify individual animals, track their movements, and observe their behaviors over time without human presence, which can reduce stress and behavioral changes in animals caused by human interaction [142, 143].

Moreover, UAVs integrated with LLMs can map large and inaccessible areas more efficiently than traditional methods. LLMs can analyze the collected geographical data to create detailed maps of habitats, including changes over time. This information is vital for managing natural reserves, planning reforestation projects, or assessing the impacts of human activities on natural habitats. LLMs can also use historical and ongoing monitoring data to predict future environmental conditions and wildlife trends. These predictions can inform conservation efforts, such as predicting the best times and places to implement species protection measures or anticipating ecological changes impacting biodiversity.

LLMs can automate generating reports and presentations for stakeholders involved in environmental conservation projects. By synthesizing complex data into comprehensible formats, LLMs facilitate more transparent communication of findings and recommendations, making it easier for decision-makers to understand the issues and take action. Moreover, LLMs can optimize UAVs and other resources in conservation projects where resources are often limited to ensure maximum coverage and data collection efficiency by scheduling UAV flights at optimal times, planning routes to cover critical areas, and ensuring data is collected cost-effectively.

VI-E Satellite and High Altitude Platform Communications with LLM-Integrated UAVs

Integrating LLMs with UAVs for enhancing satellite and High Altitude Platform (HAP) communications involves leveraging advanced analytics and cognitive capabilities to improve data relay, processing, and autonomous decision-making [25]. Since UAVs serve as mobile nodes or relay points in satellite and HAP communications networks, particularly in areas where direct communication is hindered by geographical barriers or where additional bandwidth is needed temporarily. LLMs can manage these connections dynamically, optimizing data flow between ground stations, satellites, HAPs, and end users. They can make real-time decisions on routing data through UAVs to improve network resilience and reduce latency. LLMs enable UAVs to autonomously navigate to positions where they can most effectively bridge communication gaps between satellites, HAPs, and terrestrial networks. This is particularly useful in disaster areas or during large public events requiring temporary communication boosts. UAVs, equipped with LLM capabilities, can analyze environmental data, satellite paths, and network demand to determine optimal positions without human intervention [144].

LLMs can predict potential system failures or sub-optimal performance before they become critical issues by analyzing telemetry and operational data from UAVs used in satellite and HAP communications. This predictive maintenance capability ensures higher uptime and reliability of the UAVs serving in these critical roles. LLMs can process and compress data onboard UAVs in real-time before relaying it to satellites or HAPs. This reduces the bandwidth needed for data transmission and speeds up communication. The LLMs can employ advanced algorithms to determine the most efficient ways to encode and transmit data based on current network conditions and data priorities.

LLM-integrated UAVs can adapt their communication protocols to maintain effective links with satellites and HAPs in response to changing environmental conditions, interference, or shifting network loads. LLMs can learn from past communications, predict optimal communication windows, and adjust frequencies or modulation schemes to enhance the quality of the connection. In addition, for UAVs operating in complex environments, LLMs provide enhanced situational awareness by processing data from multiple sources, including satellite and HAP sensors. This helps in making informed decisions about UAV positioning, communication strategies, and even emergency maneuvers to avoid conflicts or hazards. Integrating LLMs allows for scalable and flexible communication networks that can adapt to varying demands without extensive reconfiguration. UAVs can be deployed rapidly to scale up network capabilities in response to increased communication needs or to cover temporary satellite or HAP coverage gaps.

VII Challenges and Considerations in Implementing LLMs-integarted UAVs

Implementing LLMs for UAV communication is a novel direction with a series of challenges and considerations that must be addressed to ensure the effectiveness and safety of UAV applications. This section highlights a few fundamental challenges that must be considered to effectively incorporate LLMs in the UAV domain.

VII-A Computational Resources and Power Consumption

LLMs require significant computational power and energy to run effectively [145, 146]. However, UAVs have limited onboard computing capabilities and power supplies, constrained by the need for lightweight design to ensure longer flight duration and operational efficiency. The power consumption needed to process large models can quickly drain the UAV’s batteries, reducing the time available for mission-critical tasks [147, 8]. In addition, adding extra resources can significantly affect UAV efficiency, thus complicating the integration of LLMs. To address these issues, it is crucial to simplify LLMs by pruning unnecessary parameters and using quantization techniques to reduce model size and enhance processing speed with less power consumption [148]. Edge computing can further alleviate the need for high-bandwidth connectivity by processing data locally [149].

Furthermore, advanced AI hardware such as Graphics Processing Units (GPUs) [150], Field-Programmable Gate Arrays (FPGAs) [151], and model distillation techniques [152] can help optimize computational demands. Implementing adaptive systems that adjust resource usage based on current needs can also help manage power efficiently, ensuring operational effectiveness without compromising on performance.

VII-B Communication Latency

The latency challenge in communication is particularly critical when using LLMs for UAV operations involving real-time data processing and decision-making. For example, navigation, surveillance, and tactical responses demand minimal data processing and decision-making delays. However, when LLMs require significant computational resources, a standard solution is to offload this processing to cloud-based servers. While this approach leverages powerful computing capabilities, it inherently introduces latency due to the communication delays from the UAV to the cloud server and back. This delay can be detrimental when crucial immediate responses compromise mission effectiveness and safety [25].

To mitigate these issues, UAVs can enhance onboard processing capabilities by integrating advanced computational resources like microprocessors, GPUs, or custom Application-Specific Integrated Circuits (ASICs) to handle complex algorithms more efficiently. Balancing computational power with latency needs is crucial and can be optimized by adopting a hybrid processing approach. This involves handling urgent, real-time processing tasks directly onboard UAV, while delegating more complex, less time-sensitive tasks to the cloud. Such a strategy helps balance the computational load and tailor response times to the urgency and complexity of specific tasks. Additionally, establishing robust near-field communication networks and utilizing edge computing solutions can further reduce latency. By situating processing power closer to the UAV, either through local servers or other nearby UAVs equipped with edge servers, the communication distance and time are significantly decreased, enhancing the overall responsiveness of UAV operations [153, 154].

VII-C Model Robustness and Reliability

The challenge of model robustness and reliability is critical in deploying UAV communications, as the decisions based on model outputs lead to significant consequences [46]. For example, the models may produce unpredictable or incorrect outputs in novel or edge-case scenarios due to their reliance on patterns learned from training data, which may not adequately cover all possible real-world situations [155]. The risk is exceptionally high in dynamic environments where decisions must be made quickly and accurately, as is often the case with UAV operations [64, 103]. Continuously updating and retraining the model with new data can help it learn from recent experiences and adapt to changes or new scenarios it might encounter. This adaptation involves incorporating data from novel situations that UAVs have encountered, expanding the model’s understanding and range of responses. For instance, establishing a system where data from UAV missions is regularly fed back into the model’s training routine refines and updates its algorithms.

Furthermore, simulation-based testing and validation are essential when relying on LLMs for critical operations. Testing these models under various simulated conditions is crucial to identify potential failures or weaknesses in their responses during complex scenarios such as adverse weather conditions, communication interruptions, or unusual mission parameters. Robust fail-safe mechanisms can also be implemented to prevent harmful actions due to incorrect model outputs by establishing thresholds or conditions requiring human intervention if the model’s output is uncertain or falls outside expected parameters. Implementing redundant systems can also double-check critical decisions before execution, and enhanced error handling can address unexpected outputs from LLMs without disrupting the UAV’s operations [156].

VII-D Integration with Existing Systems

Advanced LLMs need to interact seamlessly with the UAV’s existing hardware and software modules (such as flight control, navigation systems, communication protocols, and data processing units) each with its unique specifications and operational requirements to enhance decision-making and communication within UAV operations. Integrating LLMs into these diverse frameworks is complex and time-consuming, potentially leading to extensive development and testing periods to ensure full compatibility and functionality. Therefore, adopting a modular approach to system design can significantly ease the integration of LLMs by allowing for the integration, removal, or updating of individual LLM components without disrupting the system. Modular designs offer flexibility and scalability, accommodating the specific needs of different missions or operational adjustments [26].

Furthermore, ensuring the interoperability of new LLM components with existing systems is crucial [157]. Interoperability allows different systems and software applications to communicate and work together effectively despite being developed independently. The incremental integration of LLMs into UAV systems through phased testing and deployment can also reduce the complexity of integration. It also allows for identifying and resolving specific issues without the risk of widespread system failures. In addition, developing a systematic approach for regular updates and maintenance to ensure that the integrated LLMs remain effective and that the overall system adapts to new technological advancements or changes in operational requirements is necessary [155].

VII-E Data Security and Privacy

Integrating LLMs in UAV operations raises significant concerns regarding data security and privacy, primarily since these models frequently process sensitive data, which may include personal information collected during surveillance missions. This data type is highly vulnerable and, if compromised, can lead to serious privacy violations and other security issues. Implementing robust data security measures is crucial to mitigate these risks. Thus, strong data encryption is fundamental to ensure that data remains inaccessible to unauthorized users while being transmitted and stored [158].

Furthermore, robust access control mechanisms must be established to restrict data access to only authorized personnel, thus preventing any unauthorized data manipulation or leakage. Compliance with data protection regulations is also essential. These regulations are designed to protect the privacy and integrity of data and require organizations to take stringent measures to safeguard all personal information. By adhering to these guidelines, UAV operators can help secure sensitive data processed by LLMs, minimizing the risk of breaches and maintaining the confidentiality and integrity of the information [159].

VIII Future Research Directions

This section considers the challenges and considerations previously discussed to outline future research directions. It highlights research areas that require immediate attention to enhance UAVs’ intelligence, efficiency, and adaptability through LLM integration. Such exploration is essential for overcoming current limitations and unlocking the full potential of UAV applications across various sectors.

VIII-A Advancements in LLM Algorithms for UAVs

The future work directions and opportunities for LLM technology advancements for UAV communications are rich and diverse, driven by UAV operations’ increasing complexity and demands. Novel schemes should focus on developing LLM algorithms that enable UAVs to dynamically adjust communication protocols and strategies based on real-time data about weather, terrain, and electromagnetic interference. This adaptive capability can significantly improve the effectiveness of UAVs in disaster response and military operations, where conditions can change rapidly and unpredictably [37, 160].

Future work should integrate LLMs to enhance UAV swarm intelligence, allowing for sophisticated group behaviors that mimic biological systems. In addition, future research needs to focus on algorithms that enable individual UAVs to make decisions based on the collective input of the swarm, optimizing flight paths and task allocation for efficiency and reduced energy consumption [155]. This technology holds promise for applications ranging from large-scale agricultural monitoring to search and rescue missions, where coordinated multi-UAV operations are critical.

In addition, improvement in error correction and signal processing is essential to maintain communication integrity in challenging environments. Future research needs to explore deep learning models to predict and compensate for signal degradation and develop new modulation and coding forms more resistant to interference. This technology can be particularly beneficial in crowded urban areas or during severe weather conditions, where signal loss can critically impair UAV operations [161, 68].

Future efforts should also expand the range of applications of LLM-enhanced UAV communications into new fields, such as humanitarian aid, environmental monitoring, and logistics. Future research needs to explore how UAVs equipped with advanced LLM and communication technologies can be deployed in emergency zones to provide real-time updates and aid distribution, monitor wildlife or environmental changes with minimal human involvement, and streamline supply chains with autonomous delivery services.

VIII-B Integration of LLMs with Emerging Technologies

Integrating LLMs with emerging technologies offers promising advancements for UAV communication systems. For example, incorporating Reconfigurable Intelligent Surfaces (RIS) can optimize signal processing algorithms and significantly enhance the efficiency and reliability of UAV communications by dynamically configuring RIS based on real-time environmental and traffic data [162]. It can also improve various environments, from smart cities to enhanced healthcare opportunities, by optimizing remote patient monitoring and telemedicine through optimized data transmission. In addition, LLMs can boost the performance of Augmented Reality (AR) and Virtual Reality (VR) applications by supporting high-bandwidth and low-latency communications crucial for immersive experiences [163, 164].

Furthermore, integrating LLMs with 5G/6G technology can significantly enhance UAV communication capabilities due to the higher bandwidth and lower latency offered by these networks [165]. It enables UAVs to stream high-definition video for surveillance or inspection tasks, receive updates in real-time for dynamic mission adjustment, and participate in swarm operations with improved coordination. Connecting LLM-equipped UAVs with IoT devices will lead to more interactive and responsive UAV operations within smart cities and industrial environments. UAVs can act as mobile nodes in an IoT network, collecting and processing data from various sources and making decisions on the fly. This integration can be particularly useful in disaster response scenarios, where UAVs can assess damages, detect anomalies, and communicate with other IoT devices to manage emergency services efficiently [166, 167].

In addition, integrating LLMs with edge computing platforms can decentralize data processing, reducing the latency involved in cloud computing scenarios and allowing UAVs to perform real-time data analysis at the network’s edge. This capability enables UAVs to make quicker decisions during critical missions, such as tracking moving targets or navigating complex terrains without waiting for data to be processed remotely [168]. Similarly, enhancing LLMs with specialized neural networks that can process visual and sensory data can improve UAVs’ ability to understand and interact with their environment. It allows UAVs to perform more complex recognition tasks, such as identifying specific individuals in search and rescue operations, detecting structural issues in infrastructure inspection, or monitoring agricultural lands for pest and disease patterns.

Moreover, quantum computing integration can also exponentially increase the processing capabilities of LLMs, enabling them to handle vast datasets more efficiently. Quantum-enhanced LLMs can optimize flight paths and communication protocols far beyond the current capabilities, reducing operational costs and increasing the efficiency of data-heavy tasks [169].

VIII-C Computational Efficiency and Power Management Optimization for LLM Integrated-UAV Systems

To effectively implement LLM-integrated UAV operations, novel schemes should focus on reducing the computational complexity of LLM by removing parameters that do not contribute significantly to the improved performance of UAV communication. Future work can implement a pruning scheme to reduce the model size and computational load, making it more feasible for devices with limited resources. Future schemes should also emphasize on employing quantization techniques that can lower the precision of the model’s parameters (e.g., from floating point to integers) to significantly decrease the model size and speed up inference times while consuming less power. In addition, UAVs can also benefit from edge computing services, which allow local data processing without the need to transmit data back to a central server. This reduces the necessity for continuous high-bandwidth connectivity and helps execute complex models by distributing the computational load between the UAV and the edge device.

Furthermore, future hardware design should be explicitly tailored for AI tasks. Employing GPUs. FPGAs or ASICs, optimized for AI interference can significantly enhance power and computational efficiency, delivering superior performance per watt compared to general-purpose processors. Model distillation is another effective strategy that can focused on in the future, and it involves training a smaller ’student’ model to replicate the performance of a larger ’teacher’ model. The distilled model maintains high accuracy but requires only a fraction of the computational resources, making it suitable for deployment on devices with limited capabilities. Implementing systems that dynamically adjust computing resources based on current needs and available power can optimize power usage. For example, the UAV could deploy a simplified version of the model when battery levels are low or detailed processing is unnecessary [170]. Thus, focusing on these strategies in the future can significantly enhance the feasibility of integrating sophisticated LLMs into UAV systems. These approaches help balance the trade-offs between model performance and the practical limitations of UAV platforms, ensuring that the benefits of advanced NLP capabilities can be harnessed without compromising the operational effectiveness of the UAVs.

VIII-D Latency Reduction Techniques

In order to address the latency issues, UAVs can enhance their onboard processing power by utilizing advanced computational resources such as microprocessors, GPUs, or custom ASICs, which can efficiently execute complex machine-learning algorithms. Future schemes should consider the trade-off between computational power and latency based on the specific requirements of each UAV mission to address these challenges effectively. A hybrid approach can be particularly effective, where UAVs perform critical real-time processing onboard while more complex but less time-sensitive tasks are offloaded to the cloud. Accordingly, it can balance the computational load and optimize response times according to the urgency and complexity of the tasks. For instance, integrating smart routing algorithms can dynamically determine the best location for processing data, considering current network conditions, task complexity, and the urgency of processing requirements.

Furthermore, future schemes should explore robust near-field communication networks and edge server deployment possibilities to perform computation-intensive tasks at the edge of the network with faster processing speed and lower delay.

VIII-E Standardization Efforts and Protocol Enhancements

Integrating advanced LLMs into UAV operations presents a complex challenge, as these models must seamlessly interact with a variety of existing UAV hardware and software systems [23]. UAV’s components, including flight control modules, navigation systems, communication protocols, and data processing units have their own unique specifications and operational demands. This diversity can lead to prolonged development and testing periods needed to ensure full compatibility and functionality.

Future work should focus on adopting a modular system design so individual components can easily be added, removed, or updated without disrupting the overall system integrity [171]. Moreover, future efforts must ensure that different systems and software applications can communicate and operate effectively together, even if developed independently. Thus, they can adopt standardized data formats and communication protocols widely accepted within the UAV industry. This helps LLMs understand and adhere to established standards, thereby smoothing the integration process.

In addition, future work should focus on gradually integrating LLMs into UAV systems using a phased approach to ensure compatibility and performance and establishing a systematic framework for regular updates, maintenance, and training by dedicated teams to adapt to technological advancements and maintain effective integration.

VIII-F Reliability Enhancement Strategies

To improve the reliability of LLM-integrated UAV communication systems, future work should focus on implementing advanced error correction techniques employing robust algorithms to ensure that communication remains reliable even under adverse conditions. The initial testing of LLM based UAV-system should also consider redundancy in communication channels by using multiple communication channels and backup systems to safeguard against the failure of any single channel.

Future work should also focus on AI-driven predictive maintenance by integrating AI tools to predict and schedule maintenance to prevent failures before they occur. It can help minimize downtime and extend the life span of communication components [68]. Furthermore, future schemes should adopt dynamic routing and spectrum management techniques to implement AI-driven dynamic routing algorithms, and spectrum management approaches to optimize available frequencies and paths for data transmission. This approach is beneficial for adapting to changing environmental conditions and communication traffic, enhancing overall system resilience. In addition, AI-based training and simulation of LLM-integrated UAV systems must also be carried out extensively to ensure that they can handle various operational contexts and unexpected situations for enhanced reliability.

Future work should also emphasize the establishment of real-time monitoring and decision support systems [29]. These systems are vital as they provide ongoing assessments of UAV health and communication status and can suggest or automate corrective actions.

VIII-G Interference Mitigation Schemes

Interference mitigation schemes are necessary for LLM-based UAV communication in every domain, ranging from commercial delivery services to essential emergency response operations. To meet these demands, future research must develop advanced signal-processing algorithms that can dynamically identify and mitigate interference in real-time [172]. This involves employing machine learning models, particularly deep learning techniques that predict and counteract interference patterns based on historical data and real-time inputs [173, 174].

In addition, novel schemes should explore beamforming techniques to improve signal clarity and strength. This can be achieved by implementing smart antenna technologies that adaptively focus and steer beams away from interference sources or use multiple antennas to send and receive signals, thus reducing interference effects [175, 176]. Enhancing spectrum management strategies is also vital to optimize frequency usage without causing or suffering from interference. This includes developing LLM-driven models that dynamically allocate bandwidth and adjust frequencies based on the UAV’s mission requirements and the spectral environment.

Future efforts should also focus on integrating cognitive radio capabilities that allow UAV communication systems to automatically change their frequency to avoid interference. Exploring the development of LLM algorithms can enable UAVs to sense their operational environment and make intelligent decisions about frequency hopping or modulation adjustments in essential.

Moreover, improving network coordination among UAVs to manage and mitigate interference collectively is essential. It requires future research into decentralized decision-making models where LLMs enable UAVs to share information about interference sources and collaboratively decide on the best communication paths and protocols. Moreover, strengthening UAV communications against adversarial attacks that can cause interference or disrupt communications is critical. Another key area of focus is utilizing LLMs to develop detection and response systems that identify and neutralize sophisticated signal jamming and spoofing techniques.

VIII-H Regulatory Advocacy and Policy Recommendations

Future directions and research opportunities for regulatory advocacy and policy recommendations regarding LLM-based UAV communication systems are increasingly pertinent as these technologies become more integrated into various sectors [132]. The primary focus in the future should be developing comprehensive policies addressing safety, privacy, and ethical standards while promoting innovation and integration in UAV operations. This involves collaborating with regulatory bodies to establish clear guidelines that adapt to the rapid advancements in LLM and UAV technologies.

Future work must ensure secure data communication as UAVs handle and transmit substantial amounts of potentially sensitive data [177]. Thus, measures must be implemented to protect this data against breaches and unauthorized access, safeguarding both the integrity of the data and the privacy of individuals [178]. In addition, future work should keep focusing on setting airspace usage regulations to prevent conflicts and accidents by determining how UAVs integrate with existing air traffic and defining specific zones or altitudes for UAV operations. At the same time, defining accountability measures for AI decisions is essential as UAVs increasingly make autonomous decisions based on AI; it’s crucial to establish who is responsible if those decisions lead to undesirable outcomes. Therefore, creating standards for AI behavior, ensuring AI systems are transparent and their actions are traceable, and establishing legal frameworks to address liability and compliance. Moreover, continuous monitoring and revising of these policies are crucial as technology evolves to sustain an environment that supports technological advancements and protects public interests [15].

IX Conclusion

This paper presents the transformative potential of integrating LLMs with UAVs, ushering in a new era of autonomous systems. We comprehensively analyze LLM architectures, evaluating their suitability for enhancing UAV capabilities. Our key contributions encompass a detailed assessment of LLM architectures for UAV integration and an exploration of cutting-edge LLM-based UAV architectures. This paves the way for developing significantly more sophisticated, intelligent, and responsive UAV operations. Furthermore, the focus on improved spectral sensing and sharing through LLM integration unlocks novel avenues for data processing advancements, crucial for robust decision-making within UAV systems. We demonstrate the expanded scope of existing UAV applications by integrating LLMs. We highlight how this empowers them with greater autonomy and more effective responses in various applications, ultimately leading to increased reliability and functionality across diverse sectors. The paper concludes by outlining critical areas demanding future research to harness the benefits of LLM-UAV integration fully. The advancements discussed lay the groundwork for a future where UAVs transcend their traditional roles, evolving into pivotal components of sophisticated, integrated systems that unlock the full potential of AI. This work can serve as a cornerstone in ongoing technological progress, propelling us toward a future where the synergy between LLMs and UAV technology can revolutionize various domains by achieving unprecedented levels of automation and efficiency.

References

  • [1] X. Li and A. V. Savkin, “Networked unmanned aerial vehicles for surveillance and monitoring: A survey,” Future Internet, vol. 13, no. 7, p. 174, 2021.
  • [2] N. Thakur, P. Nagrath, R. Jain, D. Saini, N. Sharma, and D. J. Hemanth, “Artificial intelligence techniques in smart cities surveillance using uavs: A survey,” Machine Intelligence and Data Analytics for Sustainable Future Smart Cities, pp. 329–353, 2021.
  • [3] H. Ren, Y. Zhao, W. Xiao, and Z. Hu, “A review of uav monitoring in mining areas: Current status and future perspectives,” International Journal of Coal Science and Technology, vol. 6, pp. 320–333, 2019.
  • [4] D. Popescu, F. Stoican, G. Stamatescu, O. Chenaru, and L. Ichim, “A survey of collaborative uav–wsn systems for efficient monitoring,” Sensors, vol. 19, no. 21, p. 4690, 2019.
  • [5] R. A. Khalil, N. Saeed, and M. Almutiry, “UAVs-assisted passive source localization using robust TDOA ranging for search and rescue,” ICT Express, vol. 9, no. 4, pp. 677–682, 2023.
  • [6] S. Ullah, K.-I. Kim, K. H. Kim, M. Imran, P. Khan, E. Tovar, and F. Ali, “Uav-enabled healthcare architecture: Issues and challenges,” Future Generation Computer Systems, vol. 97, pp. 425–432, 2019.
  • [7] M. W. Akhtar and N. Saeed, “UAVs-Enabled maritime communications: UAVs-Enabled maritime communications: Opportunities and challenges,” IEEE Systems, Man, and Cybernetics Magazine, vol. 9, no. 3, pp. 2–8, 2023.
  • [8] M. Mozaffari, W. Saad, M. Bennis, Y.-H. Nam, and M. Debbah, “A tutorial on uavs for wireless networks: Applications, challenges, and open problems,” IEEE communications surveys and tutorials, vol. 21, no. 3, pp. 2334–2360, 2019.
  • [9] S. P. H. Boroujeni, A. Razi, S. Khoshdel, F. Afghah, J. L. Coen, L. O’Neill, P. Fule, A. Watts, N.-M. T. Kokolakis, and K. G. Vamvoudakis, “A comprehensive survey of research towards ai-enabled unmanned aerial systems in pre-, active-, and post-wildfire management,” Information Fusion, p. 102369, 2024.
  • [10] A. Koubaa, A. Ammar, M. Abdelkader, Y. Alhabashi, and L. Ghouti, “Aero: Ai-enabled remote sensing observation with onboard edge computing in uavs,” Remote Sensing, vol. 15, no. 7, p. 1873, 2023.
  • [11] N. Cheng, S. Wu, X. Wang, Z. Yin, C. Li, W. Chen, and F. Chen, “Ai for uav-assisted iot applications: A comprehensive review,” IEEE Internet of Things Journal, 2023.
  • [12] S. Qazi, B. A. Khawaja, and Q. U. Farooq, “Iot-equipped and ai-enabled next generation smart agriculture: A critical review, current challenges and future trends,” Ieee Access, vol. 10, pp. 21 219–21 235, 2022.
  • [13] D. R. Vincent, N. Deepa, D. Elavarasan, K. Srinivasan, S. H. Chauhdary, and C. Iwendi, “Sensors driven ai-based agriculture recommendation model for assessing land suitability,” Sensors, vol. 19, no. 17, p. 3667, 2019.
  • [14] L. S. Iyer, “Ai enabled applications towards intelligent transportation,” Transportation Engineering, vol. 5, p. 100083, 2021.
  • [15] F. Al-Turjman and H. Zahmatkesh, “A comprehensive review on the use of ai in uav communications: Enabling technologies, applications, and challenges,” Unmanned Aerial Vehicles in Smart Cities, pp. 1–26, 2020.
  • [16] R. A. Khalil, Z. Safelnasr, N. Yemane, M. Kedir, A. Shafiqurrahman, and N. Saeed, “Advanced learning technologies for intelligent transportation systems: Prospects and challenges,” IEEE Open Journal of Vehicular Technology, vol. 5, pp. 397–427, 2024.
  • [17] M. Mohammadi and A. Al-Fuqaha, “Enabling cognitive smart cities using big data and machine learning: Approaches and challenges,” IEEE Communications Magazine, vol. 56, no. 2, pp. 94–101, 2018.
  • [18] J. Rasley, S. Rajbhandari, O. Ruwase, and Y. He, “Deepspeed: System optimizations enable training deep learning models with over 100 billion parameters,” in Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2020, pp. 3505–3506.
  • [19] J. Ye, X. Chen, N. Xu, C. Zu, Z. Shao, S. Liu, Y. Cui, Z. Zhou, C. Gong, Y. Shen et al., “A comprehensive capability analysis of gpt-3 and gpt-3.5 series models,” arXiv preprint arXiv:2303.10420, 2023.
  • [20] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” arXiv preprint arXiv:1810.04805, 2018.
  • [21] A. Mastropaolo, S. Scalabrino, N. Cooper, D. N. Palacio, D. Poshyvanyk, R. Oliveto, and G. Bavota, “Studying the usage of text-to-text transfer transformer to support code-related tasks,” in 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE).   IEEE, 2021, pp. 336–347.
  • [22] G. Liu, N. Van Huynh, H. Du, D. T. Hoang, D. Niyato, K. Zhu, J. Kang, Z. Xiong, A. Jamalipour, and D. I. Kim, “Generative ai for unmanned vehicle swarms: Challenges, applications and opportunities,” arXiv preprint arXiv:2402.18062, 2024.
  • [23] A. Agapiou and V. Lysandrou, “Interacting with the artificial intelligence (ai) language model chatgpt: a synopsis of earth observation and remote sensing in archaeology,” Heritage, vol. 6, no. 5, pp. 4072–4085, 2023.
  • [24] H. Kurunathan, H. Huang, K. Li, W. Ni, and E. Hossain, “Machine learning-aided operations and communications of unmanned aerial vehicles: A contemporary survey,” IEEE Communications Surveys and Tutorials, 2023.
  • [25] B. Rong and H. Rutagemwa, “Leveraging large language models for intelligent control of 6g integrated tn-ntn with iot service,” IEEE Network, 2024.
  • [26] A. Ullah, G. Qi, S. Hussain, I. Ullah, and Z. Ali, “The role of llms in sustainable smart cities: Applications, challenges, and future directions,” arXiv preprint arXiv:2402.14596, 2024.
  • [27] S. Liu, H. Zhang, Y. Qi, P. Wang, Y. Zhang, and Q. Wu, “Aerialvln: Vision-and-language navigation for uavs,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 15 384–15 394.
  • [28] B. Piggott, S. Patil, G. Feng, I. Odat, R. Mukherjee, B. Dharmalingam, and A. Liu, “Net-gpt: A llm-empowered man-in-the-middle chatbot for unmanned aerial vehicle,” in 2023 IEEE/ACM Symposium on Edge Computing (SEC).   IEEE, 2023, pp. 287–293.
  • [29] I. De Zarzà, J. de Curtò, and C. T. Calafate, “Socratic video understanding on unmanned aerial vehicles,” Procedia Computer Science, vol. 225, pp. 144–154, 2023.
  • [30] G. Sun, W. Xie, D. Niyato, H. Du, J. Kang, J. Wu, S. Sun, and P. Zhang, “Generative ai for advanced uav networking,” arXiv preprint arXiv:2404.10556, 2024.
  • [31] W. J. Yun, B. Lim, S. Jung, Y.-C. Ko, J. Park, J. Kim, and M. Bennis, “Attention-based reinforcement learning for real-time uav semantic communication,” in 2021 17th International Symposium on Wireless Communication Systems (ISWCS).   IEEE, 2021, pp. 1–6.
  • [32] A. Zhao, D. Huang, Q. Xu, M. Lin, Y.-J. Liu, and G. Huang, “Expel: Llm agents are experiential learners,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 17, 2024, pp. 19 632–19 642.
  • [33] E. Eigner and T. Händler, “Determinants of llm-assisted decision-making,” arXiv preprint arXiv:2402.17385, 2024.
  • [34] M. Jin, S. Wang, L. Ma, Z. Chu, J. Y. Zhang, X. Shi, P.-Y. Chen, Y. Liang, Y.-F. Li, S. Pan et al., “Time-llm: Time series forecasting by reprogramming large language models,” arXiv preprint arXiv:2310.01728, 2023.
  • [35] J. Zhu, S. Cai, F. Deng, and J. Wu, “Do llms understand visual anomalies? uncovering llm capabilities in zero-shot anomaly detection,” arXiv preprint arXiv:2404.09654, 2024.
  • [36] L. Wang, C. Ma, X. Feng, Z. Zhang, H. Yang, J. Zhang, Z. Chen, J. Tang, X. Chen, Y. Lin et al., “A survey on large language model based autonomous agents,” Frontiers of Computer Science, vol. 18, no. 6, pp. 1–26, 2024.
  • [37] Z. Xi, W. Chen, X. Guo, W. He, Y. Ding, B. Hong, M. Zhang, J. Wang, S. Jin, E. Zhou et al., “The rise and potential of large language model based agents: A survey,” arXiv preprint arXiv:2309.07864, 2023.
  • [38] Y. Wang, W. Zhong, L. Li, F. Mi, X. Zeng, W. Huang, L. Shang, X. Jiang, and Q. Liu, “Aligning large language models with human: A survey,” arXiv preprint arXiv:2307.12966, 2023.
  • [39] X. Zhu, J. Li, Y. Liu, C. Ma, and W. Wang, “A survey on model compression for large language models,” arXiv preprint arXiv:2308.07633, 2023.
  • [40] M. Gao, X. Hu, J. Ruan, X. Pu, and X. Wan, “Llm-based nlg evaluation: Current status and challenges,” arXiv preprint arXiv:2402.01383, 2024.
  • [41] J. Kaddour, J. Harris, M. Mozes, H. Bradley, R. Raileanu, and R. McHardy, “Challenges and applications of large language models,” arXiv preprint arXiv:2307.10169, 2023.
  • [42] L. Bariah, Q. Zhao, H. Zou, Y. Tian, F. Bader, and M. Debbah, “Large generative ai models for telecom: The next big thing?” IEEE Communications Magazine, 2024.
  • [43] Y. Du, S. C. Liew, K. Chen, and Y. Shao, “The power of large language models for wireless communication system development: A case study on fpga platforms,” arXiv preprint arXiv:2307.07319, 2023.
  • [44] Y. Gao, Y. Xiong, X. Gao, K. Jia, J. Pan, Y. Bi, Y. Dai, J. Sun, and H. Wang, “Retrieval-augmented generation for large language models: A survey,” arXiv preprint arXiv:2312.10997, 2023.
  • [45] G. Mialon, R. Dessì, M. Lomeli, C. Nalmpantis, R. Pasunuru, R. Raileanu, B. Rozière, T. Schick, J. Dwivedi-Yu, A. Celikyilmaz et al., “Augmented language models: a survey,” arXiv preprint arXiv:2302.07842, 2023.
  • [46] S. Schwartz, A. Yaeli, and S. Shlomov, “Enhancing trust in llm-based ai automation agents: New considerations and future challenges,” arXiv preprint arXiv:2308.05391, 2023.
  • [47] S. Yin, C. Fu, S. Zhao, K. Li, X. Sun, T. Xu, and E. Chen, “A survey on multimodal large language models,” arXiv preprint arXiv:2306.13549, 2023.
  • [48] S. Zhang, L. Dong, X. Li, S. Zhang, X. Sun, S. Wang, J. Li, R. Hu, T. Zhang, F. Wu et al., “Instruction tuning for large language models: A survey,” arXiv preprint arXiv:2308.10792, 2023.
  • [49] J. Huang and K. C.-C. Chang, “Towards reasoning in large language models: A survey,” arXiv preprint arXiv:2212.10403, 2022.
  • [50] X. Liu, P. Xu, J. Wu, J. Yuan, Y. Yang, Y. Zhou, F. Liu, T. Guan, H. Wang, T. Yu et al., “Large language models and causal inference in collaboration: A comprehensive survey,” arXiv preprint arXiv:2403.09606, 2024.
  • [51] R. Jozefowicz, O. Vinyals, M. Schuster, N. Shazeer, and Y. Wu, “Exploring the limits of language modeling,” arXiv preprint arXiv:1602.02410, 2016.
  • [52] S. Zhang, C. Gong, L. Wu, X. Liu, and M. Zhou, “Automl-gpt: Automatic machine learning with gpt,” arXiv preprint arXiv:2305.02499, 2023.
  • [53] B. Peng, S. Narayanan, and C. Papadimitriou, “On limitations of the transformer architecture,” arXiv preprint arXiv:2402.08164, 2024.
  • [54] H. Naveed, A. U. Khan, S. Qiu, M. Saqib, S. Anwar, M. Usman, N. Barnes, and A. Mian, “A comprehensive overview of large language models,” arXiv preprint arXiv:2307.06435, 2023.
  • [55] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
  • [56] S. Alaparthi and M. Mishra, “Bidirectional encoder representations from transformers (bert): A sentiment analysis odyssey,” arXiv preprint arXiv:2007.01127, 2020.
  • [57] F. Sun, J. Liu, J. Wu, C. Pei, X. Lin, W. Ou, and P. Jiang, “Bert4rec: Sequential recommendation with bidirectional encoder representations from transformer,” in Proceedings of the 28th ACM international conference on information and knowledge management, 2019, pp. 1441–1450.
  • [58] Y. Sun, S. Wang, Y. Li, S. Feng, X. Chen, H. Zhang, X. Tian, D. Zhu, H. Tian, and H. Wu, “Ernie: Enhanced representation through knowledge integration,” arXiv preprint arXiv:1904.09223, 2019.
  • [59] A. T. Rosário, “Generative ai and generative pre-trained transformer applications: Challenges and opportunities,” Making Art With Generative AI Tools, pp. 45–71, 2024.
  • [60] G. Yenduri, M. Ramalingam, G. C. Selvi, Y. Supriya, G. Srivastava, P. K. R. Maddikunta, G. D. Raj, R. H. Jhaveri, B. Prabadevi, W. Wang et al., “Gpt (generative pre-trained transformer)–a comprehensive review on enabling technologies, potential applications, emerging challenges, and future directions,” IEEE Access, 2024.
  • [61] R. Rodriguez-Torrealba, E. Garcia-Lopez, and A. Garcia-Cabot, “End-to-end generation of multiple-choice questions using text-to-text transfer transformer models,” Expert Systems with Applications, vol. 208, p. 118258, 2022.
  • [62] Z. Yang, Z. Dai, Y. Yang, J. Carbonell, R. R. Salakhutdinov, and Q. V. Le, “Xlnet: Generalized autoregressive pretraining for language understanding,” Advances in neural information processing systems, vol. 32, 2019.
  • [63] A. Sankar and R. Dhanalakshmi, “Comparative study of transformer models,” in Australasian Database Conference.   Springer, 2022, pp. 193–200.
  • [64] J. De Curtò, I. De Zarzà, and C. T. Calafate, “Semantic scene understanding with large language models on unmanned aerial vehicles,” Drones, vol. 7, no. 2, p. 114, 2023.
  • [65] D. Oneață and H. Cucu, “Multimodal speech recognition for unmanned aerial vehicles,” Computers and Electrical Engineering, vol. 90, p. 106943, 2021. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0045790620307904
  • [66] K. Choutri, M. Lagha, S. Meshoul, M. Batouche, Y. Kacel, and N. Mebarkia, “A multi-lingual speech recognition-based framework to human-drone interaction,” Electronics, vol. 11, no. 12, p. 1829, 2022.
  • [67] A. Tagliabue, K. Kondo, T. Zhao, M. Peterson, C. T. Tewari, and J. P. How, “Real: Resilience and adaptation using large language models on autonomous aerial robots,” arXiv preprint arXiv:2311.01403, 2023.
  • [68] J. Zhong, M. Li, Y. Chen, Z. Wei, F. Yang, and H. Shen, “A safer vision-based autonomous planning system for quadrotor uavs with dynamic obstacle trajectory prediction and its application with llms,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024, pp. 920–929.
  • [69] I. Jawhar, N. Mohamed, J. Al-Jaroodi, D. P. Agrawal, and S. Zhang, “Communication and networking of uav-based systems: Classification and associated architectures,” Journal of Network and Computer Applications, vol. 84, pp. 93–108, 2017.
  • [70] M. Koroteev, “Bert: a review of applications in natural language processing and understanding,” arXiv preprint arXiv:2103.11943, 2021.
  • [71] S. Ravichandiran, Getting Started with Google BERT: Build and train state-of-the-art natural language processing models using BERT.   Packt Publishing Ltd, 2021.
  • [72] M. Ehrmann, A. Hamdi, E. L. Pontes, M. Romanello, and A. Doucet, “Named entity recognition and classification in historical documents: A survey,” ACM Computing Surveys, vol. 56, no. 2, pp. 1–47, 2023.
  • [73] K. Hakala and S. Pyysalo, “Biomedical named entity recognition with multilingual bert,” in Proceedings of the 5th workshop on BioNLP open shared tasks, 2019, pp. 56–61.
  • [74] H. Xu, B. Liu, L. Shu, and P. S. Yu, “Bert post-training for review reading comprehension and aspect-based sentiment analysis,” arXiv preprint arXiv:1904.02232, 2019.
  • [75] S. Alaparthi and M. Mishra, “Bert: A sentiment analysis odyssey,” Journal of Marketing Analytics, vol. 9, no. 2, pp. 118–126, 2021.
  • [76] Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, and V. Stoyanov, “Roberta: A robustly optimized bert pretraining approach,” arXiv preprint arXiv:1907.11692, 2019.
  • [77] V. Sanh, L. Debut, J. Chaumond, and T. Wolf, “Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter,” arXiv preprint arXiv:1910.01108, 2019.
  • [78] Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, and R. Soricut, “Albert: A lite bert for self-supervised learning of language representations,” arXiv preprint arXiv:1909.11942, 2019.
  • [79] S. Luo, Y. Yao, H. Zhao, and L. Song, “A language model-based fine-grained address resolution framework in uav delivery system,” IEEE Journal of Selected Topics in Signal Processing, 2024.
  • [80] S. Silalahi, T. Ahmad, and H. Studiawan, “Named entity recognition for drone forensic using bert and distilbert,” in 2022 International Conference on Data Science and Its Applications (ICoDSA).   IEEE, 2022, pp. 53–58.
  • [81] ——, “Transformer-based named entity recognition on drone flight logs to support forensic investigation,” IEEE Access, vol. 11, pp. 3257–3274, 2023.
  • [82] Y. Fan, B. Mi, Y. Sun, and L. Yin, “Research on the intelligent construction of uav knowledge graph based on attentive semantic representation,” Drones, vol. 7, no. 6, p. 360, 2023.
  • [83] A. Radford, K. Narasimhan, T. Salimans, I. Sutskever et al., “Improving language understanding by generative pre-training,” 2018.
  • [84] A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, I. Sutskever et al., “Language models are unsupervised multitask learners,” OpenAI blog, vol. 1, no. 8, p. 9, 2019.
  • [85] K. S. Kalyan, “A survey of gpt-3 family large language models including chatgpt and gpt-4,” Natural Language Processing Journal, p. 100048, 2023.
  • [86] T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell et al., “Language models are few-shot learners,” Advances in neural information processing systems, vol. 33, pp. 1877–1901, 2020.
  • [87] A. Hendy, M. Abdelrehim, A. Sharaf, V. Raunak, M. Gabr, H. Matsushita, Y. J. Kim, M. Afify, and H. H. Awadalla, “How good are gpt models at machine translation? a comprehensive evaluation,” arXiv preprint arXiv:2302.09210, 2023.
  • [88] J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, F. L. Aleman, D. Almeida, J. Altenschmidt, S. Altman, S. Anadkat et al., “Gpt-4 technical report,” arXiv preprint arXiv:2303.08774, 2023.
  • [89] M. L. TAZIR, M. MANCAS, and T. DUTOIT, “From words to flight: Integrating openai chatgpt with px4/gazebo for natural language-based drone control,” in International Workshop on Computer Science and Engineering, 2023.
  • [90] S. Wang, J. Chen, Z. Zhang, G. Wang, Y. Tan, and Y. Zheng, “Construction of a virtual reality platform for uav deep learning,” in 2017 Chinese Automation Congress (CAC).   IEEE, 2017, pp. 3912–3916.
  • [91] S. Biswas, “Prospective role of chat gpt in the military: According to chatgpt,” Qeios, 2023.
  • [92] N. R. Choudhury, Y. Wen, and K. Chen, “Natural language navigation for robotic systems: Integrating gpt and dense captioning models with object detection in autonomous inspections,” in Construction Research Congress 2024, pp. 972–980.
  • [93] Y. Wang, Y. Pan, M. Yan, Z. Su, and T. H. Luan, “A survey on chatgpt: Ai-generated contents, challenges, and solutions,” IEEE Open Journal of the Computer Society, 2023.
  • [94] C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, and P. J. Liu, “Exploring the limits of transfer learning with a unified text-to-text transformer,” Journal of machine learning research, vol. 21, no. 140, pp. 1–67, 2020.
  • [95] J. Ni, G. H. Abrego, N. Constant, J. Ma, K. B. Hall, D. Cer, and Y. Yang, “Sentence-t5: Scalable sentence encoders from pre-trained text-to-text models,” arXiv preprint arXiv:2108.08877, 2021.
  • [96] A. Roberts, C. Raffel, K. Lee, M. Matena, N. Shazeer, P. J. Liu, S. Narang, W. Li, and Y. Zhou, “Exploring the limits of transfer learning with a unified text-to-text transformer,” Google, Tech. Rep., 2019.
  • [97] D. Cortiz, “Exploring transformers in emotion recognition: a comparison of bert, distillbert, roberta, xlnet and electra,” arXiv preprint arXiv:2104.02041, 2021.
  • [98] A. F. Adoma, N.-M. Henry, and W. Chen, “Comparative analyses of bert, roberta, distilbert, and xlnet for text-based emotion recognition,” in 2020 17th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP).   IEEE, 2020, pp. 117–121.
  • [99] P. Rajapaksha, R. Farahbakhsh, and N. Crespi, “Bert, xlnet or roberta: the best transfer learning model to detect clickbaits,” IEEE Access, vol. 9, pp. 154 704–154 716, 2021.
  • [100] H. Li, J. Choi, S. Lee, and J. H. Ahn, “Comparing bert and xlnet from the perspective of computational characteristics,” in 2020 International Conference on Electronics, Information, and Communication (ICEIC).   IEEE, 2020, pp. 1–4.
  • [101] M. O. Topal, A. Bas, and I. van Heerden, “Exploring transformers in natural language generation: Gpt, bert, and xlnet,” arXiv preprint arXiv:2102.08036, 2021.
  • [102] D. Oneata and H. Cucu, “Kite: Automatic speech recognition for unmanned aerial vehicles,” arXiv preprint arXiv:1907.01195, 2019.
  • [103] L. Li, R. Yang, M. Lv, A. Wu, and Z. Zhao, “From behavior to natural language: Generative approach for unmanned aerial vehicle intent recognition,” IEEE Transactions on Artificial Intelligence, 2024.
  • [104] Y. Yao, S. Luo, H. Zhao, G. Deng, and L. Song, “Can llm substitute human labeling? a case study of fine-grained chinese address entity recognition dataset for uav delivery,” arXiv preprint arXiv:2403.06097, 2024.
  • [105] Y. Sun, S. Wang, Y. Li, S. Feng, H. Tian, H. Wu, and H. Wang, “Ernie 2.0: A continual pre-training framework for language understanding,” in Proceedings of the AAAI conference on artificial intelligence, vol. 34, no. 05, 2020, pp. 8968–8975.
  • [106] Y. Sun, S. Wang, S. Feng, S. Ding, C. Pang, J. Shang, J. Liu, X. Chen, Y. Zhao, Y. Lu et al., “Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation,” arXiv preprint arXiv:2107.02137, 2021.
  • [107] D. Xiao, H. Zhang, Y. Li, Y. Sun, H. Tian, H. Wu, and H. Wang, “Ernie-gen: An enhanced multi-flow pre-training and fine-tuning framework for natural language generation,” arXiv preprint arXiv:2001.11314, 2020.
  • [108] F. Yu, J. Tang, W. Yin, Y. Sun, H. Tian, H. Wu, and H. Wang, “Ernie-vil: Knowledge enhanced vision-language representations through scene graphs,” in Proceedings of the AAAI conference on artificial intelligence, vol. 35, no. 4, 2021, pp. 3208–3216.
  • [109] M. Lewis, Y. Liu, N. Goyal, M. Ghazvininejad, A. Mohamed, O. Levy, V. Stoyanov, and L. Zettlemoyer, “Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension,” arXiv preprint arXiv:1910.13461, 2019.
  • [110] K. Bhattacharjee, M. Ballesteros, R. Anubhai, S. Muresan, J. Ma, F. Ladhak, and Y. Al-Onaizan, “To bert or not to bert: Comparing task-specific and task-agnostic semi-supervised approaches for sequence tagging,” arXiv preprint arXiv:2010.14042, 2020.
  • [111] D. Van Ravenzwaaij, G. Dutilh, and E.-J. Wagenmakers, “Cognitive model decomposition of the bart: Assessment and application,” Journal of Mathematical Psychology, vol. 55, no. 1, pp. 94–105, 2011.
  • [112] Z. Wei, J. Zhu, Z. Guo, and F. Ning, “The performance analysis of spectrum sharing between uav enabled wireless mesh networks and ground networks,” IEEE Sensors Journal, vol. 21, no. 5, pp. 7034–7045, 2020.
  • [113] M. A. Jasim, H. Shakhatreh, N. Siasi, A. H. Sawalmeh, A. Aldalbahi, and A. Al-Fuqaha, “A survey on spectrum management for unmanned aerial vehicles (uavs),” IEEE Access, vol. 10, pp. 11 443–11 499, 2021.
  • [114] L. Wang, H. Yang, J. Long, K. Wu, and J. Chen, “Enabling ultra-dense uav-aided network with overlapped spectrum sharing: Potential and approaches,” IEEE Network, vol. 32, no. 5, pp. 85–91, 2018.
  • [115] A. Shamsoshoara, M. Khaledi, F. Afghah, A. Razi, and J. Ashdown, “Distributed cooperative spectrum sharing in uav networks using multi-agent reinforcement learning,” in 2019 16th IEEE Annual Consumer Communications & Networking Conference (CCNC).   IEEE, 2019, pp. 1–6.
  • [116] B. Shang, V. Marojevic, Y. Yi, A. S. Abdalla, and L. Liu, “Spectrum sharing for uav communications: Spatial spectrum sensing and open issues,” IEEE Vehicular Technology Magazine, vol. 15, no. 2, pp. 104–112, 2020.
  • [117] C. Zhang and W. Zhang, “Spectrum sharing for drone networks,” IEEE Journal on Selected Areas in Communications, vol. 35, no. 1, pp. 136–144, 2016.
  • [118] Y. Lin, M. Wang, X. Zhou, G. Ding, and S. Mao, “Dynamic spectrum interaction of uav flight formation communication with priority: A deep reinforcement learning approach,” IEEE Transactions on Cognitive Communications and Networking, vol. 6, no. 3, pp. 892–903, 2020.
  • [119] M. Massaro, “Next generation of radio spectrum management: Licensed shared access for 5g,” Telecommunications Policy, vol. 41, no. 5-6, pp. 422–433, 2017.
  • [120] G. Santana, R. S. Cristo, C. Dezan, J.-P. Diguet, D. P. Osorio, and K. R. Branco, “Cognitive radio for uav communications: Opportunities and future challenges,” in 2018 International Conference on Unmanned Aircraft Systems (ICUAS).   IEEE, 2018, pp. 760–768.
  • [121] L. Zhang, Y.-C. Liang, and M. Xiao, “Spectrum sharing for internet of things: A survey,” IEEE Wireless Communications, vol. 26, no. 3, pp. 132–139, 2018.
  • [122] S. Bayhan, A. Zubow, P. Gawłowicz, and A. Wolisz, “Smart contracts for spectrum sensing as a service,” IEEE Transactions on Cognitive Communications and Networking, vol. 5, no. 3, pp. 648–660, 2019.
  • [123] F. Shen, G. Ding, Z. Wang, and Q. Wu, “Uav-based 3d spectrum sensing in spectrum-heterogeneous networks,” IEEE Transactions on Vehicular Technology, vol. 68, no. 6, pp. 5711–5722, 2019.
  • [124] X. Liu, M. Guan, X. Zhang, and H. Ding, “Spectrum sensing optimization in an uav-based cognitive radio,” IEEE Access, vol. 6, pp. 44 002–44 009, 2018.
  • [125] Y. Huang, J. Xu, L. Qiu, and R. Zhang, “Cognitive uav communication via joint trajectory and power control,” in 2018 IEEE 19th international workshop on signal processing advances in wireless communications (SPAWC).   IEEE, 2018, pp. 1–5.
  • [126] Y. Huang, W. Mei, J. Xu, L. Qiu, and R. Zhang, “Cognitive uav communication via joint maneuver and power control,” IEEE Transactions on Communications, vol. 67, no. 11, pp. 7872–7888, 2019.
  • [127] J. Chen, Y. Xu, Q. Wu, Y. Zhang, X. Chen, and N. Qi, “Interference-aware online distributed channel selection for multicluster fanet: A potential game approach,” IEEE Transactions on Vehicular Technology, vol. 68, no. 4, pp. 3792–3804, 2019.
  • [128] W. Xu, S. Wang, S. Yan, and J. He, “An efficient wideband spectrum sensing algorithm for unmanned aerial vehicle communication networks,” IEEE Internet of Things Journal, vol. 6, no. 2, pp. 1768–1780, 2018.
  • [129] J. Qiu, D. Grace, G. Ding, J. Yao, and Q. Wu, “Blockchain-based secure spectrum trading for unmanned-aerial-vehicle-assisted cellular networks: An operator’s perspective,” IEEE Internet of Things Journal, vol. 7, no. 1, pp. 451–466, 2019.
  • [130] Z. Hu, Z. Zheng, L. Song, T. Wang, and X. Li, “Uav offloading: Spectrum trading contract design for uav-assisted cellular networks,” IEEE Transactions on Wireless Communications, vol. 17, no. 9, pp. 6093–6107, 2018.
  • [131] M. M. Azari, G. Geraci, A. Garcia-Rodriguez, and S. Pollin, “Uav-to-uav communications in cellular networks,” IEEE Transactions on Wireless Communications, vol. 19, no. 9, pp. 6130–6144, 2020.
  • [132] A. Ghasemi and P. Guinand, “Accelerating radio spectrum regulation workflows with large language models (llms),” arXiv preprint arXiv:2403.17819, 2024.
  • [133] A. Kuwertz, D. Mühlenberg, J. Sander, and W. Müller, “Applying knowledge-based reasoning for information fusion in intelligence, surveillance, and reconnaissance,” in Multisensor Fusion and Integration in the Wake of Big Data, Deep Learning and Cyber Physical System: An Edition of the Selected Papers from the 2017 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI 2017).   Springer, 2018, pp. 119–139.
  • [134] W. Maharani, “Sentiment analysis during jakarta flood for emergency responses and situational awareness in disaster management using bert,” in 2020 8th International Conference on Information and Communication Technology (ICoICT).   IEEE, 2020, pp. 1–5.
  • [135] V. G. Goecks and N. R. Waytowich, “Disasterresponsegpt: Large language models for accelerated plan of action development in disaster response scenarios,” arXiv preprint arXiv:2306.17271, 2023.
  • [136] M. Lee, L. Mesicek, K. Bae, and H. Ko, “Ai advisor platform for disaster response based on big data,” Concurrency and Computation: Practice and Experience, vol. 35, no. 16, p. e6215, 2023.
  • [137] R. She and Y. Ouyang, “Efficiency of uav-based last-mile delivery under congestion in low-altitude air,” Transportation Research Part C: Emerging Technologies, vol. 122, p. 102878, 2021.
  • [138] S. Asadzadeh, W. J. de Oliveira, and C. R. de Souza Filho, “Uav-based remote sensing for the petroleum industry and environmental monitoring: State-of-the-art and perspectives,” Journal of Petroleum Science and Engineering, vol. 208, p. 109633, 2022.
  • [139] K. Nova, “Ai-enabled water management systems: an analysis of system components and interdependencies for water conservation,” Eigenpub Review of Science and Technology, vol. 7, no. 1, pp. 105–124, 2023.
  • [140] M. J. Mashala, T. Dube, B. T. Mudereri, K. K. Ayisi, and M. R. Ramudzuli, “A systematic review on advancements in remote sensing for assessing and monitoring land use and land cover changes impacts on surface water resources in semi-arid tropical environments,” Remote Sensing, vol. 15, no. 16, p. 3926, 2023.
  • [141] K. S. Adu-Manu, C. Tapparello, W. Heinzelman, F. A. Katsriku, and J.-D. Abdulai, “Water quality monitoring using wireless sensor networks: Current trends and future research directions,” ACM Transactions on Sensor Networks (TOSN), vol. 13, no. 1, pp. 1–41, 2017.
  • [142] P. J. Stephenson, “Integrating remote sensing into wildlife monitoring for conservation,” Environmental Conservation, vol. 46, no. 3, pp. 181–183, 2019.
  • [143] M. Chanev, N. Dolapchiev, I. Kamenova, and L. Filchev, “Application of remote sensing methods for monitoring wild life populations: a review,” in Ninth International Conference on Remote Sensing and Geoinformation of the Environment (RSCy2023), vol. 12786.   SPIE, 2023, pp. 731–737.
  • [144] F. Fourati and M.-S. Alouini, “Artificial intelligence for satellite communication: A review,” Intelligent and Converged Networks, vol. 2, no. 3, pp. 213–243, 2021.
  • [145] L. J. Wan, Y. Huang, Y. Li, H. Ye, J. Wang, X. Zhang, and D. Chen, “Invited paper: Software/hardware co-design for llm and its application for design verification,” in 2024 29th Asia and South Pacific Design Automation Conference (ASP-DAC), 2024, pp. 435–441.
  • [146] J. Yang, H. Jin, R. Tang, X. Han, Q. Feng, H. Jiang, S. Zhong, B. Yin, and X. Hu, “Harnessing the power of llms in practice: A survey on chatgpt and beyond,” ACM Transactions on Knowledge Discovery from Data, 2023.
  • [147] S. Javaid, N. Saeed, Z. Qadir, H. Fahim, B. He, H. Song, and M. Bilal, “Communication and control in collaborative uavs: Recent advances and future trends,” IEEE Transactions on Intelligent Transportation Systems, 2023.
  • [148] X. Ma, G. Fang, and X. Wang, “Llm-pruner: On the structural pruning of large language models,” Advances in neural information processing systems, vol. 36, pp. 21 702–21 720, 2023.
  • [149] M. Xu, D. Niyato, H. Zhang, J. Kang, Z. Xiong, S. Mao, and Z. Han, “Cached model-as-a-resource: Provisioning large language model agents for edge intelligence in space-air-ground integrated networks,” arXiv preprint arXiv:2403.05826, 2024.
  • [150] W. Kwon, Z. Li, S. Zhuang, Y. Sheng, L. Zheng, C. H. Yu, J. Gonzalez, H. Zhang, and I. Stoica, “Efficient memory management for large language model serving with pagedattention,” in Proceedings of the 29th Symposium on Operating Systems Principles, 2023, pp. 611–626.
  • [151] H. Chen, J. Zhang, Y. Du, S. Xiang, Z. Yue, N. Zhang, Y. Cai, and Z. Zhang, “A comprehensive evaluation of fpga-based spatial acceleration of llms,” in Proceedings of the 2024 ACM/SIGDA International Symposium on Field Programmable Gate Arrays, 2024, pp. 185–185.
  • [152] Y. Huang, “Leveraging large language models for enhanced nlp task performance through knowledge distillation and optimized training strategies,” arXiv preprint arXiv:2402.09282, 2024.
  • [153] S. S. Hassan, Y. M. Park, Y. K. Tun, W. Saad, Z. Han, and C. S. Hong, “Satellite-based its data offloading & computation in 6g networks: A cooperative multi-agent proximal policy optimization drl with attention approach,” IEEE Transactions on Mobile Computing, 2023.
  • [154] Q. Chen, Z. Guo, W. Meng, S. Han, C. Li, and T. Q. Quek, “A survey on resource management in joint communication and computing-embedded sagin,” arXiv preprint arXiv:2403.17400, 2024.
  • [155] K. Telli, O. Kraa, Y. Himeur, A. Ouamane, M. Boumehraz, S. Atalla, and W. Mansoor, “A comprehensive review of recent research trends on unmanned aerial vehicles (uavs),” Systems, vol. 11, no. 8, p. 400, 2023.
  • [156] S. Mishra and P. Palanisamy, “Autonomous advanced aerial mobility—an end-to-end autonomy framework for uavs and beyond,” IEEE Access, vol. 11, pp. 136 318–136 349, 2023.
  • [157] Z. Wang, J. Zhang, H. Du, R. Zhang, D. Niyato, B. Ai, and K. B. Letaief, “Generative ai agent for next-generation mimo design: Fundamentals, challenges, and vision,” arXiv preprint arXiv:2404.08878, 2024.
  • [158] Y. Yao, J. Duan, K. Xu, Y. Cai, Z. Sun, and Y. Zhang, “A survey on large language model (llm) security and privacy: The good, the bad, and the ugly,” High-Confidence Computing, p. 100211, 2024.
  • [159] F. Wu, N. Zhang, S. Jha, P. McDaniel, and C. Xiao, “A new era in llm security: Exploring security concerns in real-world llm-based systems,” arXiv preprint arXiv:2402.18649, 2024.
  • [160] Q. Wu, G. Bansal, J. Zhang, Y. Wu, S. Zhang, E. Zhu, B. Li, L. Jiang, X. Zhang, and C. Wang, “Autogen: Enabling next-gen llm applications via multi-agent conversation framework,” arXiv preprint arXiv:2308.08155, 2023.
  • [161] N. Chacko and V. Chacko, “Paradigm shift presented by large language models (llm) in deep learning,” ADVANCES IN EMERGING COMPUTING TECHNOLOGIES, vol. 40, 2023.
  • [162] M. A. ElMossallamy, H. Zhang, L. Song, K. G. Seddik, Z. Han, and G. Y. Li, “Reconfigurable intelligent surfaces for wireless communications: Principles, challenges, and opportunities,” IEEE Transactions on Cognitive Communications and Networking, vol. 6, no. 3, pp. 990–1002, 2020.
  • [163] K. Faisal and W. Choi, “Machine learning approaches for reconfigurable intelligent surfaces: A survey,” IEEE Access, vol. 10, pp. 27 343–27 367, 2022.
  • [164] X. Yuan, Y.-J. A. Zhang, Y. Shi, W. Yan, and H. Liu, “Reconfigurable-intelligent-surface empowered wireless communications: Challenges and opportunities,” IEEE wireless communications, vol. 28, no. 2, pp. 136–143, 2021.
  • [165] I. de Zarzà, J. de Curtò, G. Roig, and C. T. Calafate, “Llm adaptive pid control for b5g truck platooning systems,” Sensors, vol. 23, no. 13, p. 5899, 2023.
  • [166] R. Ali, Y. B. Zikria, A. K. Bashir, S. Garg, and H. S. Kim, “Urllc for 5g and beyond: Requirements, enabling incumbent technologies and network intelligence,” IEEE Access, vol. 9, pp. 67 064–67 095, 2021.
  • [167] Y. Hong, J. Wu, and R. Morello, “Llm-twin: Mini-giant model-driven beyond 5g digital twin networking framework with semantic secure communication and computation,” arXiv preprint arXiv:2312.10631, 2023.
  • [168] Z. Lin, G. Qu, Q. Chen, X. Chen, Z. Chen, and K. Huang, “Pushing large language models to the 6g edge: Vision, challenges, and opportunities,” arXiv preprint arXiv:2309.16739, 2023.
  • [169] Z. Liang, J. Cheng, R. Yang, H. Ren, Z. Song, D. Wu, X. Qian, T. Li, and Y. Shi, “Unleashing the potential of llms for quantum computing: A study in quantum architecture design,” arXiv preprint arXiv:2307.08191, 2023.
  • [170] J. Zhu, A. Pande, P. Mohapatra, and J. J. Han, “Using deep learning for energy expenditure estimation with wearable sensors,” in 2015 17th International Conference on E-health Networking, Application & Services (HealthCom).   IEEE, 2015, pp. 501–506.
  • [171] Y. Xia, M. Shenoy, N. Jazdi, and M. Weyrich, “Towards autonomous system: flexible modular production system enhanced with large language model agents,” in 2023 IEEE 28th International Conference on Emerging Technologies and Factory Automation (ETFA).   IEEE, 2023, pp. 1–8.
  • [172] H. Xu, L. Han, Q. Yang, M. Li, and M. Srivastava, “Penetrative ai: Making llms comprehend the physical world,” in Proceedings of the 25th International Workshop on Mobile Computing Systems and Applications, 2024, pp. 1–7.
  • [173] N. Formosa, M. Quddus, S. Ison, M. Abdel-Aty, and J. Yuan, “Predicting real-time traffic conflicts using deep learning,” Accident Analysis & Prevention, vol. 136, p. 105429, 2020.
  • [174] L. Shi, B. Li, C. Kim, P. Kellnhofer, and W. Matusik, “Towards real-time photorealistic 3d holography with deep neural networks,” Nature, vol. 591, no. 7849, pp. 234–239, 2021.
  • [175] J. A. Srar, K.-S. Chung, and A. Mansour, “Adaptive array beamforming using a combined lms-lms algorithm,” IEEE Transactions on Antennas and Propagation, vol. 58, no. 11, pp. 3545–3557, 2010.
  • [176] L. Bariah, Q. Zhao, H. Zou, Y. Tian, F. Bader, and M. Debbah, “Large language models for telecom: The next big thing?” arXiv preprint arXiv:2306.10249, 2023.
  • [177] D. Pathirana, Towards better regulation of unmanned aerial vehicles in national airspace: a comparative analysis of selected national regulations.   McGill University (Canada), 2018.
  • [178] W. Y. B. Lim, S. Garg, Z. Xiong, Y. Zhang, D. Niyato, C. Leung, and C. Miao, “Uav-assisted communication efficient federated learning in the era of the artificial intelligence of things,” IEEE Network, vol. 35, no. 5, pp. 188–195, 2021.

 




























No comments:

Post a Comment

Threats to GNSS receivers providing PNT services pose threat globally in 2024: Civil v. Military Users

Summary of GNSS Vulnerabilities Here's a concise summary of the webinar transcript on GNSS hacking and cybersecurity: Key Points: 1. GNS...