SkillAgentSearch skills...

RadioLLM

RadioLLM: Introducing Large Language Model into Cognitive Radio via Hybrid Prompt and Token Reprogrammings

Install / Use

/learn @Andrewchenxd/RadioLLM
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<div align="center"> <a href="https://www.python.org/"> <img src="./docs/images/logo.svg" width="200" alt="logo"/> </a> <h1>RadioLLM: Introducing Large Language Model into Cognitive Radio via Hybrid Prompt and Token Reprogrammings</h1>

<a href="https://www.python.org/"><img alt="Python" src="https://img.shields.io/badge/Python-3.8-blue"></a> <a href="https://pytorch.org/"><img alt="PyTorch" src="https://img.shields.io/badge/Pytorch-latest-orange"></a> <a href="https://arxiv.org/abs/2501.17888"><img alt="arXiv" src="https://img.shields.io/badge/Paper-arXiv-B31B1B"></a> <a href="https://huggingface.co/datasets/"><img alt="Dataset" src="https://img.shields.io/badge/Dataset-🤗-FFFDF5"></a> <a href="https://github.com/SparkZu/RadioLLM"><img alt="GitHub Repo stars" src="https://img.shields.io/github/stars/Andrewchenxd/RadioLLM"></a>

</div>

📚 Introduction

RadioLLM, a novel framework that incorporates Hybrid Prompt and Token Reprogramming (HPTR) and a Frequency Attuned Fusion (FAF) module to enhance LLMs for CRT tasks. HPTR enables the integration of radio signal features with expert knowledge, while FAF improves the modeling of high-frequency features critical for precise signal processing. These innovations allow RadioLLM to handle diverse CRT tasks, bridging the gap between LLMs and traditional signal processing methods. Extensive empirical studies on multiple benchmark datasets demonstrate that the proposed RadioLLM achieves superior performance over current baselines.

🔥 NEWS

  • [2025-02-01] 📝 The preprint of the RadioLLM paper is available on arXiv. Check the paper page for more details.
  • [2025-05-13] 📝 The revised version of the RadioLLM paper is now available on arXiv. See the paper page for details.
  • [2025-05-20] 📝 Part of the implementation code for RadioLLM is now publicly available.
  • [2025-06-14] 📝 This release includes supplemental module uploads that were accidentally excluded from prior distributions. More related code: Signal All You Need.

📅 TODO

  • [-] Collect the codes of RadioLLM's classification network and other comparison models.

💻 Requirements

The code is implemented in Python 3.9. We recommend using the provided Dockerfile to set up the environment, as all dependencies are already specified in it. You can build and run the Docker image with:

docker build -t radiollm:latest .
docker run --rm -it radiollm:latest

Alternatively, you can manually create a conda environment and install dependencies as previously described. You can install the required packages by running the following command:

conda create --name radiollm python=3.9
conda activate radiollm
pip install git+https://github.com/huggingface/transformers
pip install git+https://github.com/huggingface/peft

📖 Citation

Please cite the following paper if you use this study in your research:

@article{chen2025radiollm,
  title={RadioLLM: Introducing Large Language Model into Cognitive Radio via Hybrid Prompt and Token Reprogrammings},
  author={Chen, Shuai and Zu, Yong and Feng, Zhixi and Yang, Shuyuan and Li, Mengchang and Ma, Yue and Liu, Jun and Pan, Qiukai and Zhang, Xinlei and Sun, Changjun},
  journal={arXiv preprint arXiv:2501.17888},
  year={2025}
}

Related Skills

View on GitHub
GitHub Stars51
CategoryDevelopment
Updated8d ago
Forks0

Languages

Python

Security Score

80/100

Audited on Mar 18, 2026

No findings