์ „์ฒด ๊ธ€

Welcome! I'm a student studying about deep learning(NLP) ๐Ÿ˜‰ The goal of my study is to develop a competent LLM helping people!
Insight ๐Ÿ˜Ž

LM์˜ context window, ๊ธธ์–ด์•ผ ํ• ๊นŒ? ์งง์•„์•ผ ํ• ๊นŒ? ๐Ÿ“๐Ÿคจ

Newly spotlighted elements of LM โœจ LM์€ ์‹œ์‹œ๊ฐ๊ฐ ๋ณ€ํ™”ํ•ด๊ฐ€๊ณ  ์žˆ๋‹ค. ๋ฉฐ์น  ์ „์— ์ƒˆ๋กญ๊ฒŒ ๋ฐœํ‘œ๋œ ๋ชจ๋ธ์ด ์˜ค๋Š˜์—์„œ๋Š” ๊ทธ ๋ฉด๋ชจ๊ฐ€ ๋‚ฑ๋‚ฑ์ด ํŒŒ์•…๋˜์–ด ๋ถ€์กฑํ•œ ์ ๋“ค์ด๋‚˜ ๋‹จ์ ๋“ค์ด ์ง€์ ๋ฐ›๊ณ  ์žˆ๋Š” ์š”์ฆ˜์ด๋‹ค. ๐Ÿ˜ฅ ๊ทธ๋งŒํผ LM์€ ๊ทธ๊ฒƒ์ด ํŒŒ๋ผ๋ฏธํ„ฐ๋“  ๋ฐ์ดํ„ฐ๋“  ๋‹ค๋ฐฉ๋ฉด์œผ๋กœ ๋น ๋ฅด๊ฒŒ ๋ณ€ํ™”ํ•ด๋‚˜๊ฐ€๊ณ  ์žˆ๋Š”๋ฐ, ์ด๋ฒˆ ํฌ์ŠคํŒ…์—์„œ ๋‹ค๋ค„๋ณด๊ณ ์ž ํ•˜๋Š” ๋‚ด์šฉ์€ ์˜ค๋žœ ์‹œ๊ฐ„ ๋™์•ˆ ๋ณ„๋กœ ๊ฑด๋“œ๋ ค์ง€์ง€ ์•Š๋‹ค๊ฐ€ ์ตœ๊ทผ์— ์—ฌ๋Ÿฌ ์—ฐ๊ตฌ(Chen et al., 2023, Ding et al., 2023, Liu et al., 2023)๋ฅผ ํ†ตํ•ด ๋‹ค์‹œ ๊ฐ๊ด‘๋ฐ›๊ณ  ์žˆ๋Š” ๋‚ด์šฉ์ธ LM์˜ context window์— ๋Œ€ํ•ด์„œ ์–˜๊ธฐํ•ด๋ณด๊ณ ์ž ํ•œ๋‹ค! ๐Ÿ˜Š What is the 'context window'? ๐Ÿค” ์‹œ์ž‘ํ•˜๊ธฐ์— ์•ž์„œ์„œ ์ด๋ฒˆ ํฌ์ŠคํŒ…์—์„œ ์ค‘์š”ํ•˜๊ฒŒ ๋‹ค๋ค„๋ณผ ๋‚ด์šฉ์ธ ..

Insight ๐Ÿ˜Ž

Closed-source๐Ÿ”’? Open-source๐Ÿ”“? ๊ทธ๊ฒŒ ๋ญ”๋ฐ?? ๐Ÿคจ๐Ÿค”

Starting from ChatGPT ๐Ÿค– which is closed-source ์ž‘๋…„ 12์›”, ์ฆ‰ 2022๋…„ 12์›”์— ์ „ ์„ธ๊ณ„์˜ ์‚ฌ๋žŒ๋“ค์—๊ฒŒ ์ ์ž–์ด ์‹ ์„ ํ•œ ์ถฉ๊ฒฉ์„ ์ค€ ์‚ฌ๊ฑด์ด ๋ฐœ์ƒํ•˜์˜€๋‹ค. ๋ฐ”๋กœ ๊ทธ ์œ ๋ช…ํ•œ 'ChatGPT'์˜ ๋ฐœํ‘œ๋‹ค! OpenAI์—์„œ ๋ฐœํ‘œํ•œ ์ด ๊ฑฐ๋Œ€ ์–ธ์–ด ๋ชจ๋ธ(Large Language Model, LLM)์€ ์ง€๊ธˆ๊นŒ์ง€์™€๋Š” ์ฐจ์›์ด ๋‹ค๋ฅธ ์—„์ฒญ๋‚œ ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์ฃผ๋ฉด์„œ ์‚ฌ๋žŒ๋“ค์˜ ์‚ฌํšŒ ๋ฐ ์‚ถ์— ์ „๋ฐ˜์ ์œผ๋กœ ์Šค๋ฉฐ๋“ค์–ด๊ฐ€๊ธฐ ์‹œ์ž‘ํ–ˆ๋‹ค. ํ•˜์ง€๋งŒ, ์ด ์™„๋ฒฝํ•ด ๋ณด์ด๋Š” ChatGPT๋„ ์—ฌ๋Ÿฌ ๊ฐ€์ง€ ๋‹จ์ ์„ ๊ฐ€์ง€๊ณ  ์žˆ๋Š”๋ฐ, ๊ทธ์ค‘์—์„œ ์ด๋ฒˆ ํฌ์ŠคํŒ…์—์„œ ๋‹ค๋ค„๋ณด๊ณ ์ž ํ•˜๋Š” ๋‚ด์šฉ์€ ๋ฐ”๋กœ 'Closed-source' model์ด๋ผ๋Š” ์ ์ด๋‹ค. ๐Ÿšซ closed-source๊ฐ€ ๋ฌด์—‡์ผ๊นŒ? ์ด ์šฉ์–ด๋ฅผ ์ฒ˜์Œ ๋“ฃ๊ฒŒ ๋œ๋‹ค๋ฉด ๋‹ค์†Œ ์ƒ์†Œํ• ํ…๋ฐ, clos..

Insight ๐Ÿ˜Ž

How has scaling law developed in NLP? ๐Ÿค” - NLP์—์„œ scaling law๋Š” ์–ด๋–ป๊ฒŒ ๋ฐœ์ „๋˜์—ˆ์„๊นŒ?

Before Starting.. 2017๋…„ NLP๋ฅผ ํฌํ•จํ•œ ์ง€๊ธˆ๊นŒ์ง€์˜ ๋”ฅ๋Ÿฌ๋‹์˜ ํŒ๋„๋ฅผ ๋’ค์ง‘์–ด์—Ž๋Š” ํ˜์‹ ์ ์ธ ๋ชจ๋ธ์ธ 'Transformer'๊ฐ€ ์ œ์•ˆ๋˜์—ˆ๋‹ค. ์ด๋ฒˆ ํฌ์ŠคํŒ…์—์„œ ๋‹ค๋ค„๋ณผ ๋‚ด์šฉ์€ Transformer์— ๋Œ€ํ•œ ์ž์„ธํ•œ ๋‚ด์šฉ์ด ์•„๋‹ˆ๊ธฐ์— ๋”ฐ๋กœ ๊นŠ์ด ์•Œ์•„๋ณด์ง€๋Š” ์•Š๊ฒ ์ง€๋งŒ, ์ด๋ฒˆ ํฌ์ŠคํŒ…์„ ์ดํ•ดํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” ์ด ๋ชจ๋ธ์˜ ์‚ฌ์ด์ฆˆ์— ๋Œ€ํ•ด์„œ๋Š” ์•Œ์•„๋‘˜ ํ•„์š”๊ฐ€ ์žˆ๋‹ค. Transformer์˜ ์‚ฌ์ด์ฆˆ๋Š” 465M ๊ฐœ์˜ ํŒŒ๋ผ๋ฏธํ„ฐ๋ฅผ ๊ฐ€์ง€๋Š” ๋ชจ๋ธ์ด์—ˆ๋‹ค. ํ•˜์ง€๋งŒ, ๋ถˆ๊ณผ 3๋…„ ๋งŒ์— ์ด ์‚ฌ์ด์ฆˆ๊ฐ€ ์ •๋ง ์ž‘๊ฒŒ ๋Š๊ปด์ง€๊ฒŒ ํ•  ๋งŒํผ ํฐ ์‚ฌ์ด์ฆˆ์˜ ๋ชจ๋ธ์ธ GPT-3(175B)๊ฐ€ ๋‚˜์˜ค๊ฒŒ ๋˜์—ˆ๋‹ค. ๊ทธ๋ฆฌ๊ณ  ํ˜„์žฌ๊นŒ์ง€๋„ ์ด๋ณด๋‹ค ๋” ํฐ ๋ชจ๋ธ๋“ค์€ ๊ณ„์† ๋‚˜์˜ค๊ณ  ์žˆ๋‹ค. LM์˜ ์‚ฌ์ด์ฆˆ๊ฐ€ ์ด๋ ‡๊ฒŒ ์ ์  ์ปค์ง€๊ฒŒ ๋œ ์ด์œ ๋Š” ๋ฌด์—‡์ผ๊นŒ? ๊ทธ ์ด์œ ๋Š” Kaplan et al. 2020..

Paper Reading ๐Ÿ“œ/Natural Language Processing

SelFee: Iterative Self-Revising LLM Empowered by Self-Feedback Generation ๋ฆฌ๋ทฐ

Introduction SelFee SelFee๋Š” KAIST์˜ LK Lab์—์„œ ๋งŒ๋“  ์ƒˆ๋กœ์šด instruction-following LM์œผ๋กœ ์‘๋‹ต์—์„œ self-feedback์„ ์ƒ์„ฑํ•˜๊ณ  ํ”ผ๋“œ๋ฐฑ์— ๊ธฐ๋ฐ˜ํ•ด์„œ self-revise ํ•˜๋Š” ๋ชจ๋ธ์ด๋‹ค. ChatGPT์— ์˜ํ•ด ์ƒ์„ฑ๋œ self-feedback๊ณผ revision data๋ฅผ ํฌํ•จํ•˜๋Š” 178K ๊ฐœ์˜ training instance๋ฅผ ์‚ฌ์šฉํ•ด์„œ LLaMA model(7B & 13B)์„ fine-tune ํ•˜์˜€๋‹ค. SelFee์˜ ์ž‘๋™ ์˜ˆ์‹œ Vicuna Evaluation์—์„œ ๋‘ SelFee(7B & 13B) ๋ชจ๋ธ์€ LLaMA, Alpaca, Vicuna, Guanaco๋ฅผ ๋Šฅ๊ฐ€ํ•˜๊ณ  ChatGPT์™€ ๋น„์Šทํ•œ ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์คฌ๋‹ค. SelFee๋Š” ํŠนํžˆ high-quality te..

Paper Reading ๐Ÿ“œ/Natural Language Processing

Self-Refine: Iterative Refinement with Self-Feedback ๋…ผ๋ฌธ ๋ฆฌ๋ทฐ

The overview of this paper ์ด ๋…ผ๋ฌธ์—์„œ๋Š” Self-Refine์„ ์†Œ๊ฐœํ•˜์˜€๋‹ค. Self-Refine์€ ๋ฐ˜๋ณต์ ์ธ ํ”ผ๋“œ๋ฐฑ๊ณผ ๊ฐœ์„ ์„ ํ†ตํ•ด LLM์˜ ์ดˆ๊ธฐ output์„ ๊ฐœ์„ ์‹œํ‚ค๊ธฐ ์œ„ํ•œ ๋ฐฉ๋ฒ•์ด๋‹ค. Self-Refine์˜ ์ฃผ๋œ ์•„์ด๋””์–ด๋Š” LLM์„ ์‚ฌ์šฉํ•ด ์ดˆ๊ธฐ output์„ ์ƒ์„ฑํ•˜๊ณ , ๊ทธ๋‹ค์Œ์— ๋˜‘๊ฐ™์€ LLM์ด output์— ๋Œ€ํ•ด ํ”ผ๋“œ๋ฐฑ์„ ์ œ๊ณตํ•˜๊ณ  ์ด ํ”ผ๋“œ๋ฐฑ์„ ์‚ฌ์šฉํ•ด ๋ฐ˜๋ณต์ ์œผ๋กœ ์ž๊ธฐ ์ž์‹ ์„ ๊ฐœ์„ ํ•ด ๋‚˜๊ฐ€๋Š” ๊ฒƒ์ด๋‹ค. ํ•œ ๋งˆ๋””๋กœ Self-Refine์€ ํ•˜๋‚˜์˜ LLM์„ generator, refiner, feedback provider๋กœ ์‚ฌ์šฉํ•˜๋Š” ๊ฒƒ์ด๋‹ค. Self-Refine์€ ๋ชจ๋“  ํ‰๊ฐ€๋œ task์—์„œ Self-Refine์œผ๋กœ ์ƒ์„ฑ๋œ output์€ ๊ธฐ์กด์˜ ๋˜‘๊ฐ™์€ LLM์œผ๋กœ ์ƒ์„ฑ๋œ output๋ณด๋‹ค human..

Paper Reading ๐Ÿ“œ/Natural Language Processing

Reflexion: Language Agents with Verbal Reinforcement Learning ๋…ผ๋ฌธ ๋ฆฌ๋ทฐ

The overview of this paper ์ด ๋…ผ๋ฌธ์—์„œ๋Š” ๊ฐ€์ค‘์น˜๋ฅผ ์—…๋ฐ์ดํŠธํ•˜์ง€ ์•Š๊ณ  ๋Œ€์‹ ์— ์–ธ์–ด์  ํ”ผ๋“œ๋ฐฑ์„ ํ†ตํ•ด language agent๋ฅผ ๊ฐ•ํ™”ํ•˜๊ธฐ ์œ„ํ•œ ์ƒˆ๋กœ์šด ํ”„๋ ˆ์ž„์›Œํฌ์ธ Reflexion์„ ์†Œ๊ฐœํ•˜์˜€๋‹ค. ๊ตฌ์ฒด์ ์œผ๋กœ, Reflexion agent๋Š” task ํ”ผ๋“œ๋ฐฑ ์‹ ํ˜ธ์— ๋Œ€ํ•ด ์–ธ์–ด๋กœ ๋‚˜ํƒ€๋‚ด๊ณ , ๊ทธ๋‹ค์Œ์— ์ดํ›„์˜ ์‹œ๋„์— ๋” ๋‚˜์€ ์˜์‚ฌ ๊ฒฐ์ •์„ ์œ ๋ฐœํ•˜๊ธฐ ์œ„ํ•ด ๋ฉ”๋ชจ๋ฆฌ ๋ฒ„ํผ์— ์ด๋“ค๋งŒ์˜ reflective text๋ฅผ ์œ ์ง€ํ•œ๋‹ค. Reflexion์€ ๋‹ค์–‘ํ•œ ํƒ€์ž…๊ณผ ์†Œ์Šค์˜ ํ”ผ๋“œ๋ฐฑ ์‹ ํ˜ธ๋ฅผ ํฌํ•จํ•  ์ˆ˜ ์žˆ์„ ์ •๋„๋กœ ์ถฉ๋ถ„ํžˆ ์œ ์—ฐํ•˜๊ณ , ๋‹ค์–‘ํ•œ task์— ๊ฑธ์ณ์„œ baseline agent์— ๋น„ํ•ด์„œ ์ƒ๋‹นํ•œ ๊ฐœ์„ ์„ ์–ป์—ˆ๋‹ค. Table of Contents 1. Introduction 2. Reflexion: reinforceme..

Paper Reading ๐Ÿ“œ/Natural Language Processing

GPT-4๋„ ์ž˜ ๋ชปํ•œ API ํ˜ธ์ถœ์„ ํ•œ๋‹ค๊ณ ?!? - Gorilla๐Ÿฆ: Large Language Model Connected with Massive APIs ๋…ผ๋ฌธ ๋ฆฌ๋ทฐ

The overview of this paper LLM์€ ์ตœ๊ทผ์— ์—„์ฒญ ๋ฐœ์ „ํ–ˆ์œผ๋‚˜, ์ด๋“ค์˜ API ํ˜ธ์ถœ์„ ํ†ตํ•œ ํšจ๊ณผ์ ์ธ ํˆด ์‚ฌ์šฉ์— ๋Œ€ํ•œ ์ž ์žฌ์„ฑ์€ ๋งŒ์กฑ๋˜์ง€ ์•Š์€ ์ฑ„ ๋‚จ์•„์žˆ๋‹ค. ์ด ๋…ผ๋ฌธ์—์„œ๋Š” API ํ˜ธ์ถœ ์ž‘์„ฑ์—์„œ GPT-4์˜ ์„ฑ๋Šฅ์„ ๋Šฅ๊ฐ€ํ•˜๋Š” fine-tuned LLaMA-based model์ธ Gorilla๐Ÿฆ๋ฅผ ์†Œ๊ฐœํ•˜์˜€๋‹ค. Gorilla๋Š” document retriever์™€ ํ•จ๊ป˜ ์‚ฌ์šฉ๋  ๋•Œ, test-time ๋ฌธ์„œ ๋ณ€ํ™”์— ์ ์‘ํ•˜๊ธฐ ์œ„ํ•œ ๊ฐ•๋ ฅํ•œ ๋Šฅ๋ ฅ์„ ๋ณด์—ฌ์ฃผ๊ณ , ์œ ์—ฐํ•œ ์‚ฌ์šฉ์ž ์—…๋ฐ์ดํŠธ ๋˜๋Š” ๋ฒ„์ „ ๋ณ€ํ™”๋ฅผ ๊ฐ€๋Šฅํ•˜๊ฒŒ ํ•ด ์ฃผ์—ˆ๋‹ค. ์ด๊ฒƒ์€ LLM์„ direct ํ•˜๊ฒŒ prompting ํ•  ๋•Œ ์ผ๋ฐ˜์ ์œผ๋กœ ๋งž๋‹ฅ๋œจ๋ฆฌ๋Š” hallucination์˜ ๋ฌธ์ œ์ ์„ ์ƒ๋‹นํžˆ ์™„ํ™”ํ•˜์˜€๋‹ค. ๋˜ํ•œ ๋…ผ๋ฌธ์—์„œ๋Š” Gorilla์˜ ๋Šฅ๋ ฅ์„ ํ‰๊ฐ€ํ•˜๊ธฐ ์œ„ํ•ด ..

Paper Reading ๐Ÿ“œ/Natural Language Processing

Open-domain instruction์˜ ํšจ๊ณผ ๐Ÿช„ - WizardLM: Empowering Large Language Models to Follow Complex Instructions ๋…ผ๋ฌธ ๋ฆฌ๋ทฐ

The overview of this paper open-domain instruction๊ณผ ํ•จ๊ป˜ LLM์„ ํ•™์Šต์‹œํ‚ค๋Š” ๊ฒƒ์€ ์ƒ๋‹นํ•œ ์„ฑ๊ณต์„ ๊ฐ€์ ธ์™”๋‹ค. ์ด ๋…ผ๋ฌธ์—์„œ๋Š” ์‚ฌ๋žŒ ๋Œ€์‹ ์— LLM์„ ์‚ฌ์šฉํ•ด์„œ ๋‹ค์–‘ํ•œ ๋ ˆ๋ฒจ์˜ ๋ณต์žก๋„๋ฅผ ๊ฐ€์ง€๋Š” ๋งŽ์€ ์–‘์˜ instruction data๋ฅผ ์ƒ์„ฑํ•˜๊ธฐ ์œ„ํ•œ ๋ฐฉ์•ˆ์„ ๋ณด์—ฌ์ค€๋‹ค. ์ดˆ๊ธฐ instruction set์™€ ํ•จ๊ป˜ ์‹œ์ž‘ํ•ด์„œ, ์ด instruction set๋ฅผ Evol-instruct๋ฅผ ์‚ฌ์šฉํ•ด์„œ ๋”์šฑ ๋ณต์žกํ•œ instruction์œผ๋กœ step-by-step ์ž‘์„ฑํ•˜์˜€๋‹ค. ๊ทธ๋‹ค์Œ์—, ๋ชจ๋“  ์ƒ์„ฑ๋œ instruction ๋ฐ์ดํ„ฐ๋ฅผ LLaMA๋ฅผ fine-tune ํ•˜๊ธฐ ์œ„ํ•ด ์„ž์—ˆ๋‹ค. ์ด๋ ‡๊ฒŒ ํ•ด์„œ ๋‚˜์˜จ ๋ชจ๋ธ์ด ๋ฐ”๋กœ WizardLM์ด๋‹ค. Human Evaluation & Vicuna Evaluatio..

Paper Reading ๐Ÿ“œ/Natural Language Processing

ํ•„์š”ํ•œ ๊ฑด ์˜ค์ง ๊ต๊ณผ์„œ ์ˆ˜์ค€์˜ ๋ฐ์ดํ„ฐ๋ฟ!! ๐Ÿ“– - phi-1: Textbooks Are All You Need ๋…ผ๋ฌธ ๋ฆฌ๋ทฐ

The overview of this paper ๋…ผ๋ฌธ์—์„œ๋Š” ๋‹ค๋ฅธ ๋ชจ๋ธ๋ณด๋‹ค ํ›จ์”ฌ ์ž‘๊ณ  code๋ฅผ ์œ„ํ•œ LLM์ธ phi-1์„ ์†Œ๊ฐœํ•˜์˜€๋‹ค. phi-1์€ 1.3B Transformer model์ด๊ณ , ์›น์œผ๋กœ๋ถ€ํ„ฐ textbook ํ€„๋ฆฌํ‹ฐ ๋ฐ์ดํ„ฐ์˜ ์„ ํƒ์  ๋ชจ์Œ๊ณผ ์ข…ํ•ฉ์ ์œผ๋กœ ์ƒ์„ฑ๋œ textbook์„ ์‚ฌ์šฉํ•˜๊ณ , GPT-3.5๋กœ ํ›ˆ๋ จ๋˜์—ˆ๋‹ค. phi-1์€ ์ž‘์€ ๊ทœ๋ชจ์—๋„ ๋ถˆ๊ตฌํ•˜๊ณ  ๋†’์€ pass@1 accuracy๋ฅผ ๋‹ฌ์„ฑํ•˜์˜€๋‹ค. Table of Contents 1. Introduction 2. Training details and the importance of high-quality data 3. Spikes of model capability after finetuning on CodeExercises 4. Evaluati..

Cartinoe
Cartinoe's paper review