Ghost in the minecraft: Generally capable agents for open-world environments via large language models with text-based knowledge and memory. arXiv preprint arXiv:2305.17144, 2023 17. Sclar M, Kumar S, West P, Suhr A, Choi Y, Tsvetkov Y. Minding language models’(lack of) theory of mind:...
A Survey of Large Language Modelsarxiv.org/abs/2303.18223 A quick understanding of how LLMs work Typically, large language models (LLMs) refer to Transformer language models that contain hundreds of billions (or more) of parameters. The basic background for LLMs: 1) scaling laws 2) em...
A Survey on Neural Network Language Models. arXiv 2019 paper bib Kun Jing, Jungang Xu A Survey on Recent Approaches for Natural Language Processing in Low-Resource Scenarios. NAACL-HLT 2021 paper bib Michael A. Hedderich, Lukas Lange, Heike Adel, Jannik Strötgen, Dietrich Klakow An Intro...
A project associated with this survey has been created at https://github.com/jingyi0000/VLM_survey.Index Terms: Visual recognition, vision-language model, pre-training, transfer learning, knowledge distillation, image classification, object detection, semantic segmentation, deep neural network, deep ...
A Survey on Neural Network Language Models. arXiv 2019 paper bib Kun Jing, Jungang Xu Evaluation of Text Generation: A Survey. arXiv 2020 paper bib Asli Celikyilmaz, Elizabeth Clark, Jianfeng Gao Neural Text Generation: Past, Present and Beyond. arXiv 2018 paper bib Sidi Lu, Yaoming Zhu...
Further,word2vec[19, 20] was proposed to build a simplified shallow neural network for learning distributed word representations, which were demonstrated to be very effective across a variety of NLP tasks. These studies have initiated the use of language models for representation learning (beyond wo...
a survey of large language model Large language models are becoming increasingly popular in the field of natural language processing (NLP), as they can generate human-like text and improve the performance of various NLP tasks. A large language model is a neural network that can take a large ...
5.1. Pretrained language models The training of most pre-trained language models (PLMs) is based on self-supervised learning. Self-supervised learning mainly uses auxiliary tasks to mine its supervised information from large-scale unsupervised data, and trains the network through this constructed superv...
Xie, Ziang, et al. “Data noising as smoothing in neural network language models.” arXiv preprint arXiv:1703.02573 (2017). Yu, Adams Wei, et al. “Qanet: Combining local convolution with global self-attention for reading comprehension.” arXiv preprint arXiv:1804.09541 (2018). ...
A Survey on Knowledge Graphs: Representation, Acquisition, and Applications 知识图谱方向重要综述,几乎是必看!文章太长了,小破网和小破电脑快要卡爆了... 从三个角度来浏览文章,泛读,精读以及总结。 文章巨长预警!!!⚠️⚠️⚠️ 文章巨长预警!!!⚠️⚠️⚠️ ...