Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • Perspective
  • Published:

Large language models in 6G from standard to on-device networks

Abstract

Sixth-generation radiocommunications (6G) systems are adopting language models for intent-driven control, context-aware adaptation to environmental and network dynamics, and end-to-end orchestration of communication. Conventional artificial intelligence (AI) typically lacks capabilities in task generalization, communication and reasoning; however, the rapid development of efficient large language models (LLMs) can automate mobile and network operations and inform the design of 6G. In this Perspective, we discuss the use of LLMs in 6G networks. We show how cloud LLMs can improve the self-organization, efficiency and local deployment of 6G networks. Next, we describe key techniques for implementing LLMs on devices for 6G. Finally, we propose LLMs for in multi-agent scenarios, emphasizing the importance of telecom-specific adaptations and security. LLMs have the potential to improve network design, operation and service delivery by enabling telecom stakeholders to convert intents into trustworthy, privacy-aware decisions that deliver more reliable services for users.

Key points

  • Large language models (LLMs) are gigantic neural networks which are trained on a huge corpus of unlabelled data to learn a universal representation of data within a given modality or across different modalities.

  • Model compression is essential for efficiently deploying multimodal LLMs at the network edge or in multi-agent sixth-generation radiocommunications (6G) systems, where agents must make high-level, interdependent decisions for tasks such as radio resource control and intent-driven operations.

  • The formation of networks of LLMs will be natural with 6G and will allow very advanced tasks to be realized through collaborative schemes, but will also introduce competitive scenarios that require further research efforts to be understood.

  • Integrating large artificial intelligence (AI) models such as LLMs into 6G is crucial to enable applications such as wireless sensing, intelligent transport systems and intent-driven networks.

This is a preview of subscription content, access via your institution

Access options

Buy this article

USD 39.95

Prices may be subject to local taxes which are calculated during checkout

Fig. 1: Timeline of language models.
Fig. 2: The full life cycle of training, aligning and deploying LLMs in wireless networks.
Fig. 3: A comprehensive framework for deploying LLMs at the edge.
Fig. 4: Multi-agent communication.
Fig. 5: Potential applications of LLMs in 6G communications.

Similar content being viewed by others

References

  1. Ouyang, L. et al. Training language models to follow instructions with human feedback. Adv. Neural Inf. Process. Syst. 35, 27730–27744 (2022).

    Google Scholar 

  2. Wu, S. et al. BloombergGPT: a large language model for finance. Preprint at https://doi.org/10.48550/arXiv.2303.17564 (2023).

  3. Singhal, K. et al. Towards expert-level medical question answering with large language models. Nat. Med. 31, 943–950 (2023).

    Article  Google Scholar 

  4. Colombo, P. et al. SaulLM-7B: a pioneering large language model for law. Preprint at https://doi.org/10.48550/arXiv.2403.03883 (2024).

  5. Azerbayev, Z. et al. Llemma: an open language model for mathematics. In The 12th International Conference on Learning Representations (IEEE, 2024).

  6. Bengio, Y., Ducharme, R., Vincent, P. & Janvin, C. A neural probabilistic language model. J. Mach. Learn. Res. 3, 1137–1155 (2003).

    Google Scholar 

  7. Mikolov, T., Chen, K., Corrado, G. S. & Dean, J. Efficient estimation of word representations in vector space. In The 1st International Conference on Learning Representations (IEEE, 2013).

  8. Pennington, J., Socher, R. & Manning, C. D. GloVe: global vectors for word representation. In Proc. 2014 Conference on Empirical Methods in Natural Language Processing 1532–1543 (ACL, 2014).

  9. Le, Q. V. & Mikolov, T. Distributed representations of sentences and documents. In The 31st International Conference Machine Learning 32, 1188–1196 (PMLR, 2014).

  10. Sutskever, I., Vinyals, O. & Le, Q. V. Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 27, 3104–3112 (2014).

    Google Scholar 

  11. Vaswani, A. et al. Attention is all you need. Adv. Neural Inf. Process. Syst. 30, 5998–6008 (2017).

    Google Scholar 

  12. Brown, T. et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 33, 1877–1901 (2020).

    Google Scholar 

  13. Devlin, J., Chang, M.-W., Lee, K. & Toutanova, K. BERT: pre-training of deep bidirectional transformers for language understanding. In Proc. 2019 Conference of the North American Chapter of the Association for Computational Linguistics 4171–4186 (ACL, 2019).

  14. Anil, R. et al. Gemini: a family of highly capable multimodal models. Preprint at https://doi.org/10.48550/arXiv.2312.11805 (2023).

  15. Bai, J. et al. Qwen technical report. Preprint at https://doi.org/10.48550/arXiv.2309.16609 (2023).

  16. Touvron, H. et al. LLaMA 2: open foundation and fine-tuned chat models. Preprint at https://doi.org/10.48550/arXiv.2307.09288 (2023).

  17. Aaron, G. et al. The Llama 3 herd of models. Preprint at https://doi.org/10.48550/arXiv.2407.21783 (2024).

  18. Jiang, A. Q. et al. Mixtral of experts. Preprint at https://doi.org/10.48550/arXiv.2401.04088 (2024).

  19. Liu, A. et al. DeepSeek-V3 technical report. Preprint at https://doi.org/10.48550/arXiv.2412.19437 (2024).

  20. Kaplan, J. et al. Scaling laws for neural language models. Preprint at https://doi.org/10.48550/arXiv.2001.08361 (2020).

  21. Hoffmann, J. et al. Training compute-optimal large language models. Adv. Neural Inf. Process. Syst. 35, 30016–30030 (2022).

    Google Scholar 

  22. Almazrouei, E. et al. The Falcon series of open language models. Preprint at https://doi.org/10.48550/arXiv.2311.16867 (2023).

  23. OpenAI. OpenAI o1 system card. Preprint at https://doi.org/10.48550/arXiv.2412.16720 (2024).

  24. Guo, D. et al. DeepSeek-R1: incentivizing reasoning capability in LLMs via reinforcement learning. Nature 645, 633–638 (2025).

    Article  Google Scholar 

  25. Wei, J. et al. Chain-of-thought prompting elicits reasoning in large language models. Adv. Neural Inf. Process. Syst. 35, 24824–24837 (2022).

    Google Scholar 

  26. Yao, S. et al. Tree of thoughts: deliberate problem solving with large language models. Adv. Neural Inf. Process. Syst. 36, 11809–11822 (2023).

    Google Scholar 

  27. Besta, M. et al. Graph of thoughts: solving elaborate problems with large language models. In The 38th AAAI Conference on Artificial Intelligence 38, 17682–17690 (AAAI, 2023).

  28. Zou, H. et al. GenAINet: enabling wireless collective intelligence via knowledge transfer and reasoning. IEEE Access 13, 77764–77777 (2024).

    Article  Google Scholar 

  29. Shao, J. et al. WirelessLLM: empowering large language models towards wireless intelligence. J. Commun. Inf. Netw. 9, 99–112 (2024).

    Article  Google Scholar 

  30. Arslan, M., Ghanem, H., Munawar, S. & Cruz, C. A survey on RAG with LLMs. Procedia Comput. Sci. 246, 3781–3790 (2024).

    Article  Google Scholar 

  31. Bornea, A.-L. et al. Telco-RAG: navigating the challenges of retrieval-augmented language models for telecommunications. In IEEE Global Communications Conference 2359–2364 (IEEE, 2024).

  32. Fu, Z. et al. On the effectiveness of parameter-efficient fine-tuning. In The 38th AAAI Conference on Artificial Intelligence 37, 12799–12807 (AAAI, 2023).

  33. Hu, E. J. et al. LoRA: low-rank adaptation of large language models. Int. Conf. Learn. Representations 1, 3 (2022).

    Google Scholar 

  34. Wang, L. et al. Parameter-efficient fine-tuning in large language models: a survey of methodologies. Artif. Intell. Rev. 58, 227 (2025).

    Article  Google Scholar 

  35. Li, X. L. & Liang, P. Prefix-tuning: optimizing continuous prompts for generation. In The 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing 1, 4582–4597 (ACL, 2021).

  36. Zou, H. et al. TelecomGPT: a framework to build telecom-specific large language models. IEEE Trans. Mach. Learn. Commun. Netw. 3, 948–975 (2025).

    Article  Google Scholar 

  37. Maatouk, A., Ampudia, K. C., Ying, R. & Tassiulas, L. Tele-LLMs: a series of specialized large language models for telecommunications. Preprint at https://doi.org/10.48550/arXiv.2409.05314 (2024).

  38. Guo, P. et al Selective aggregation for low-rank adaptation in federated learning. In The 13th International Conference on Learning Representations (ICLR, 2025).

  39. Zhang, J. et al. Towards building the federated GPT: federated instruction tuning. In IEEE International Conference on Acoustics, Speech and Signal Processing 6915–6919 (IEEE, 2024).

  40. Wang, Y. et al. Aligning large language models with human: a survey. Preprint at https://doi.org/10.48550/arXiv.2307.12966 (2023).

  41. Song, S. et al. How to bridge the gap between modalities: survey on multimodal large language model. IEEE Trans. Knowl. Data Eng. 37, 5311–5329 (2025).

    Article  Google Scholar 

  42. Radford, A. et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning 139, 8748–8763 (PMLR, 2021).

  43. Hsu, W. N. et al. HuBert: self-supervised speech representation learning by masked prediction of hidden units. IEEE/ACM Trans. Audio, Speech, Lang. Process. 29, 3451–3460 (2021).

    Article  Google Scholar 

  44. Wu, Y. et al. Large-scale contrastive language-audio pretraining with feature fusion and keyword-to-caption augmentation. In IEEE Internationol Conference on Acoustics, Speech and Signal Processing 1–5 (IEEE, 2023).

  45. Xu, S. et al. Large multi-modal models (LMMs) as universal foundation models for AI-native wireless systems. IEEE Netw. 38, 10–20 (2024).

    Article  Google Scholar 

  46. Wang, L. et al. Generative AI for RF sensing in IoT systems. IEEE Internet Things Mag. 8, 112–120 (2025).

    Article  Google Scholar 

  47. Tao T. et al. 6G hyper reliable and low-latency communication—requirement analysis and proof of concept. In IEEE 98th Vehicular Technology Conference 1–5 (IEEE, 2023).

  48. Guo, Y., Cheng, Y., Zhao, Y., Wang, J. & Xu, H. A survey on model compression and acceleration for pretrained language models. Proc. AAAI Conf. Artificial Intelligence 37, 10566–10575 (2023).

    Article  Google Scholar 

  49. Huang, W. et al. BiLLM: pushing the limit of post-training quantization for LLMs. In International Conference on Machine Learning 235, 20023–20042 (PMLR, 2024).

  50. Dettmers, T., Pagnoni, A., Holtzman, A. & Zettlemoyer, L. QLoRA: efficient finetuning of quantized LLMs. Adv. Neural Inf. Process. Syst. 36, 10088–10115 (2024).

    Google Scholar 

  51. Liu, Z. et al. LLM-QAT: data-free quantization aware training for large language models. In Findings of the Association for Computational Linguistics 467–484 (ACL,2024).

  52. Kim, J. et al. Memory-efficient fine-tuning of compressed large language models via sub-4-bit integer quantization. Adv. Neural Inf. Process. Syst. 36, 36187–36207 (2023).

    Google Scholar 

  53. Wang, H. et al. Bitnet: 1-bit pre-training for large language models. J. Mach. Learn. Res. 26, 1–29 (2025).

    Google Scholar 

  54. Frantar, E., Ashkboos, S., Hoefler, T. & Alistarh, D. GPTQ: accurate post-training quantization for generative pre-trained transformers. In The 11th International Conference on Learning Representations (ICLR, 2023).

  55. Yao, Z. et al. ZeroQuant: efficient and affordable post-training quantization for large-scale transformers. Adv. Neural Inf. Process. Syst. 35, 27168–27183 (2022).

    Google Scholar 

  56. Lin, J. et al. AWQ: activation-aware weight quantization for on-device LLM compression and acceleration. Proc. Mach. Learn. Syst. 6, 87–100 (2024).

    Google Scholar 

  57. Hershcovitch, M. et al. ZipNN: lossless compression for AI models. In IEEE 18th International Conference on Cloud Computing 186–198 (IEEE, 2025).

  58. Yu, R. et al. NISP: pruning networks using neuron importance score propagation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition 9194–9203 (IEEE, 2018).

  59. Frantar, E. & Alistarh, D. SparseGPT: massive language models can be accurately pruned in one-shot. In International Conference on Machine Learning 202, 10323–10337 (PMLR, 2023).

  60. Ma, X., Fang, G. & Wang, X. LLM-pruner: on the structural pruning of large language models. Adv. Neural Inf. Process. Syst. 36, 21702–21720 (2023).

    Google Scholar 

  61. Hinton, G., Vinyals, O. & Dean, J. Distilling the knowledge in a neural network. Preprint at https://doi.org/10.48550/arXiv.1503.02531 (2015).

  62. Buciluă, C., Caruana, R. & Niculescu-Mizil, A. Model compression. In 12th ACM International Conference on Knowledge Discovery and Data Mining 535–541 (ACM, 2006).

  63. Sanh, V., Debut, L., Chaumond, J. & Wolf, T. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. Preprint at https://doi.org/10.48550/arXiv.1910.01108 (2019).

  64. Dao, T., Fu, D., Ermon, S., Rudra, A. & Ré, C. FlashAttention: fast and memory-efficient exact attention with IO-awareness. Adv. Neural Inf. Process. Syst. 35, 16344–16359 (2022).

    Google Scholar 

  65. Kwon, W. et al. Efficient memory management for large language model serving with PagedAttention. In 29th Symposium on Operating Systems Principles 611–626 (ACM, 2023).

  66. Wang, L. et al. A survey on large language model based autonomous agents. Front. Comput. Sci. 18, 186345 (2024).

    Article  Google Scholar 

  67. Xi, Z. et al. The rise and potential of large language model based agents: a survey. Sci. China Inf. Sci. 68, 121101 (2025).

    Article  Google Scholar 

  68. Mohammadi, M., Li, Y., Lo, J. & Yip, W. Evaluation and benchmarking of llm agents: a survey. In 31st ACM SIGKDD Conference on Knowledge Discovery Data Mining 2, 6129–6139 (2025).

    Google Scholar 

  69. Shinn, N., Cassano, F., Gopinath, A., Narasimhan, K. & Yao, S. Reflexion: language agents with verbal reinforcement learning. Adv. Neural Inf. Process. Syst. 36, 8634–8652 (2023).

    Google Scholar 

  70. Fırat, M. & Kuleli, S. What if GPT4 became autonomous: the Auto-GPT project and use cases. J. Emerg. Comput. Technol. 3, 1–6 (2024).

    Article  Google Scholar 

  71. Guo, T. et al. Large language model based multi-agents: a survey of progress and challenges. In 33rd International Joint Conference on Artificial Intelligence (IJCAI, 2024).

  72. Shen, Y. et al. HuggingGPT: solving AI tasks with ChatGPT and its friends in Hugging Face. Adv. Neural Inf. Process. Syst. 36, 38154–38180 (2023).

    Google Scholar 

  73. Wang, J. et al. Mobile-Agent: autonomous multi-modal mobile device agent with visual perception. In The 12th International Conference on Learning Representations (ICLR, 2024).

  74. Liu, F. et al. Large language models for computer networking operations and management: a survey on applications, key techniques, and opportunities. Computer Netw. 271, 111614 (2025).

    Article  Google Scholar 

  75. Xu, M. et al. When large language model agents meet 6G networks: perception, grounding, and alignment. IEEE Wireless Commun. 31, 63–71 (2024).

    Article  Google Scholar 

  76. Jiang, F. et al. Large language model enhanced multi-agent systems for 6G communications. IEEE Wireless Commun. 31, 48–55 (2024).

    Article  Google Scholar 

  77. Zhang, Q., Yu, Y., FU, Q. and Ye, D. More agents is all you need. Trans. Machine Learning Res. https://openreview.net/forum?id=bgzUSZ8aeg (2024).

  78. Li, G. et al. CAMEL: communicative agents for “mind” exploration of large language model society. Adv. Neural Inf. Process. Syst. 36, 51991–52008 (2023).

    Google Scholar 

  79. Park, J. S. et al. Generative agents: interactive simulacra of human behavior. In The 36th Annual ACM Symposium on User Interface Software and Technology 1–22 (ACM, 2023).

  80. Hong, S. et al. MetaGPT: meta programming for a multi-agent collaborative framework. In The 12th International Conference on Learning Representations (ICLR, 2024).

  81. Wu, Q. et al. AutoGen: enabling next-gen LLM applications via multi-agent conversation. In First Conference on Language Modeling (COLM, 2024).

  82. Du, Y. et al. Improving factuality and reasoning in language models through multiagent debate. In International Conference on Machine Learning 235, 11733–11763 (PMLR, 2024).

  83. Chan, C.-M. et al. ChatEval: towards better LLM-based evaluators through multi-agent debate. In The 12th International Conference on Learning Representations (ICLR, 2024).

  84. Hua, Y. et al. Assistive large language model agents for socially-aware negotiation dialogues. In Findings of the Association for Computational Linguistics 8047–8074 (ACL, 2024).

  85. Akata, E. et al. Playing repeated games with large language models. Nat. Hum. Behav. 9, 1380–1390 (2025).

    Article  Google Scholar 

  86. Yiming, Z. & Dongning, G. Multi-agent reinforcement learning for multi-cell spectrum and power allocation. IEEE Trans. Commun. 73, 5980–5992 (2025).

    Article  Google Scholar 

  87. Andrea, L. et al. dApps: enabling real-time AI-based open RAN control. Comput. Netw. 269, 111342 (2025).

    Article  Google Scholar 

  88. Zhengquan, Z. et al. 6G wireless networks: vision, requirements, architecture, and key technologies. IEEE Veh. Technol. Mag. 14, 28–41 (2019).

    Article  Google Scholar 

  89. Kaesberg, L. B., Becker, J., Wahle, J. P., Ruas T., & Gipp B., Voting or consensus? Decision-making in multi-agent debate. In Findings of the Association for Computational Linguistics 11640–11671 (ACL, 2025).

  90. Achiam, J. et al. GPT-4 technical report. Preprint at https://doi.org/10.48550/arXiv.2303.08774 (2023).

  91. Hao, S. et al. Reasoning with language model is planning with world model. In Conference on Empirical Methods in Natural Language Processing 8154–8173 (ACL, 2023).

  92. Bariah, L. et al. Large generative AI models for telecom: the next big thing? IEEE Commun. Mag. 62, 84–90 (2024).

    Article  Google Scholar 

  93. Tian, Y. et al. Multimodal transformers for wireless communications: a case study in beam prediction. ITU J. Future Evol. Technol. 4, 461–471 (2023).

    Article  Google Scholar 

  94. Tian, Y., Alhammadi, A., Quran, A. & Ali, A. S. A novel approach to wavenet architecture for RF signal separation with learnable dilation and data augmentation. In IEEE International Confernce on Acoustics, Speech, and Signal Processing Workshops 79–80 (IEEE, 2024).

  95. Wu, K., Tian, Y., Almazrouei, E. & Bader, F. A simple and effective design for intelligent sub-nyquist modulation recognition. In International Conference on Artificial Intelligence and Big Data 772–776 (Springer, 2023).

  96. Le Scao, T. et al. BLOOM: a 176B-parameter open-access multilingual language model. Preprint at https://doi.org/10.48550/arXiv.2211.05100 (2022).

  97. Laurençon, H. et al. The BigScience ROOTS corpus: a 1.6 TB composite multilingual dataset. Adv. Neural Inf. Process. Syst. 35, 31809–31826 (2022).

    Google Scholar 

  98. Anil, R. et al. PaLM 2 technical report. Preprint at https://doi.org/10.48550/arXiv.2305.10403 (2023).

  99. Penedo, G. et al. The RefinedWeb dataset for Falcon LLM: outperforming curated corpora with web data, and web data only. Preprint at https://doi.org/10.48550/arXiv.2306.01116 (2023).

  100. Cai, Z. et al. InternLM2 technical report. Preprint at https://doi.org/10.48550/arXiv.2403.17297 (2024).

  101. Zeng, A. et al. GLM-130B: an open bilingual pre-trained model. Preprint at https://doi.org/10.48550/arXiv.2210.02414 (2022).

  102. Taylor, R. et al. Galactica: a large language model for science. Preprint at https://doi.org/10.48550/arXiv.2211.09085 (2022).

  103. Lozhkov, A. et al. StarCoder 2 and the stack v2: the next generation. Preprint at https://doi.org/10.48550/arXiv.2402.19173 (2024).

  104. Raffel, C. et al. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 140:1–140:67 (2019).

    MathSciNet  Google Scholar 

  105. Zhu, Y. et al. Aligning books and movies: towards story-like visual explanations by watching movies and reading books. In IEEE International Conference on Computer Vision 19–27 (IEEE, 2015).

Download references

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed substantially to discussion of the content. H.Z., Q.Z., S.L., C.Z. and Y.T. researched and wrote the paper. L.B., F.B. and M.D. edited and reviewed the manuscript before submission.

Corresponding author

Correspondence to Samson Lasaulce.

Ethics declarations

Competing interests

The authors declare no competing interests.

Peer review

Peer review information

Nature Reviews Electrical Engineering thanks Markus Dominik Mueck and the other, anonymous, reviewer(s) for their contribution to the peer review of this work.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zou, H., Zhao, Q., Lasaulce, S. et al. Large language models in 6G from standard to on-device networks. Nat Rev Electr Eng 3, 123–134 (2026). https://doi.org/10.1038/s44287-025-00239-6

Download citation

  • Accepted:

  • Published:

  • Version of record:

  • Issue date:

  • DOI: https://doi.org/10.1038/s44287-025-00239-6

Search

Quick links

Nature Briefing AI and Robotics

Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing: AI and Robotics