如有转载,请注明出处。欢迎关注微信公众号:低调奋进。打算开始写LLM系列文章,主要从数据、训练框架、对齐等方面进行LLM整理。
Baichuan 2: Open Large-scale Language Models
原始文章链接
https://cdn.baichuan-ai.com/paper/Baichuan2-technical-report.pdf
github
https://github.com/baichuan-inc
hugginggface
https://huggingface.co/baichuan-inc
训练LLM的同行可以精读文章llama、llama2和baichuan2等文章,干货较多。本文不做翻译,主要罗列个人关注的重点。阅读本文的前提是已经对LLM熟悉,最好已经积累一定训练经验。本文干货较多,有的实验可以作为自己试验的指向标。
同时想阅读LLM的综述文章可以读以下文章:
A Survey of Large Language Models
https://arxiv.org/pdf/2303.18223.pdf
Large Language Models
https://arxiv.org/pdf/2307.05782.pdf
A Comprehensive Overview of Large Language Models
https://arxiv.org/pdf/2307.06435.pdf
A Survey on Evaluation of Large Language Models
https://arxiv.org/pdf/2307.03109.pdf
Is Prompt All You Need? No. A Comprehensive and Broader View of Instruction Learning
https://arxiv.org/pdf/2303.10475.pdf