LLM
ยท NLP
Text2All
ยท All2All
Multi-modal
ยท Multi-task
Large language models (LLMs) have achieved remarkable progress in various natural language processing tasks with emergent abilities. However, they face inherent limitations, such as an inability to access up-to-date information, utilize external tools, or perform precise mathematical reasoning. In this paper, we introduce Chameleon, a plug-and-play compositional reasoning framework that augments LLMs to help address these challenges. Chameleon synthesizes programs to compose various tools, including LLM models, off-the-shelf vision models, web search engines, Python functions, and rule-based modules tailored to user interests. Built on top of an LLM as a natural language planner, Chameleon infers the appropriate sequence of tools to compose and execute in order to generate a final response. We showcase the adaptability and effectiveness of Chameleon on two tasks: ScienceQA and TabMWP. Notably, Chameleon with GPT-4 achieves an 86.54% accuracy on ScienceQA, significantly improving upon the best published few-shot model by 11.37%; using GPT-4 as the underlying LLM, Chameleon achieves a 17.8% increase over the state-of-the-art model, leading to a 98.78% overall accuracy on TabMWP. Further studies suggest that using GPT-4 as a planner exhibits more consistent and rational tool selection and is able to infer potential constraints given the instructions, compared to other LLMs like ChatGPT.
Believable proxies of human behavior can empower interactive applications ranging from immersive environments to rehearsal spaces for interpersonal communication to prototyping tools. In this paper, we introduce generative agents--computational software agents that simulate believable human behavior. Generative agents wake up, cook breakfast, and head to work; artists paint, while authors write; they form opinions, notice each other, and initiate conversations; they remember and reflect on days past as they plan the next day. To enable generative agents, we describe an architecture that extends a large language model to store a complete record of the agent's experiences using natural language, synthesize those memories over time into higher-level reflections, and retrieve them dynamically to plan behavior. We instantiate generative agents to populate an interactive sandbox environment inspired by The Sims, where end users can interact with a small town of twenty five agents using natural language. In an evaluation, these generative agents produce believable individual and emergent social behaviors: for example, starting with only a single user-specified notion that one agent wants to throw a Valentine's Day party, the agents autonomously spread invitations to the party over the next two days, make new acquaintances, ask each other out on dates to the party, and coordinate to show up for the party together at the right time. We demonstrate through ablation that the components of our agent architecture--observation, planning, and reflection--each contribute critically to the believability of agent behavior. By fusing large language models with computational, interactive agents, this work introduces architectural and interaction patterns for enabling believable simulations of human behavior.
Recent advancements in decision-making large language model (LLM) agents have demonstrated impressive performance across various benchmarks. However, these state-of-the-art approaches typically necessitate internal model fine-tuning, external model fine-tuning, or policy optimization over a defined state space. Implementing these methods can prove challenging due to the scarcity of high-quality training data or the lack of well-defined state space. Moreover, these agents do not possess certain qualities inherent to human decision-making processes, specifically the ability to learn from mistakes. Self-reflection allows humans to efficiently solve novel problems through a process of trial and error. Building on recent research, we propose Reflexion, an approach that endows an agent with dynamic memory and self-reflection capabilities to enhance its existing reasoning trace and task-specific action choice abilities. To achieve full automation, we introduce a straightforward yet effective heuristic that enables the agent to pinpoint hallucination instances, avoid repetition in action sequences, and, in some environments, construct an internal memory map of the given environment. To assess our approach, we evaluate the agent's ability to complete decision-making tasks in AlfWorld environments and knowledge-intensive, search-based question-and-answer tasks in HotPotQA environments. We observe success rates of 97% and 51%, respectively, and provide a discussion on the emergent property of self-reflection.
Like people, LLMs do not always generate the best text for a given generation problem on their first try (e.g., summaries, answers, explanations). Just as people then refine their text, we introduce SELF-REFINE, a framework for similarly improving initial outputs from LLMs through iterative feedback and refinement. The main idea is to generate an output using an LLM, then allow the same model to provide multi-aspect feedback for its own output; finally, the same model refines its previously generated output given its own feedback. Unlike earlier work, our iterative refinement framework does not require supervised training data or reinforcement learning, and works with a single LLM. We experiment with 7 diverse tasks, ranging from review rewriting to math reasoning, demonstrating that our approach outperforms direct generation. In all tasks, outputs generated with SELF-REFINE are preferred by humans and by automated metrics over those generated directly with GPT-3.5 and GPT-4, improving on average by absolute 20% across tasks.
Solving complicated AI tasks with different domains and modalities is a key step toward advanced artificial intelligence. While there are abundant AI models available for different domains and modalities, they cannot handle complicated AI tasks. Considering large language models (LLMs) have exhibited exceptional ability in language understanding, generation, interaction, and reasoning, we advocate that LLMs could act as a controller to manage existing AI models to solve complicated AI tasks and language could be a generic interface to empower this. Based on this philosophy, we present HuggingGPT, a framework that leverages LLMs (e.g., ChatGPT) to connect various AI models in machine learning communities (e.g., Hugging Face) to solve AI tasks. Specifically, we use ChatGPT to conduct task planning when receiving a user request, select models according to their function descriptions available in Hugging Face, execute each subtask with the selected AI model, and summarize the response according to the execution results. By leveraging the strong language capability of ChatGPT and abundant AI models in Hugging Face, HuggingGPT is able to cover numerous sophisticated AI tasks in different modalities and domains and achieve impressive results in language, vision, speech, and other challenging tasks, which paves a new way towards advanced artificial intelligence.
Auto-GPT is an experimental open-source application showcasing the capabilities of the GPT-4 language model. This program, driven by GPT-4, chains together LLM "thoughts", to autonomously achieve whatever goal you set. As one of the first examples of GPT-4 running fully autonomously, Auto-GPT pushes the boundaries of what is possible with AI.
์ฌ๋ด์ ์ ํ ์ ๋ณด, ๋ฌผ๋ฅ ์ ๋ณด, ์ธ์ฌ๊ท์ , ํ๊ณ๊ธฐ์ค๊ณผ ๊ฐ์ ์ ๋ณด๋ ์ฌ๋ด์ ์ ์ง๋์ด์ผ ํ๋ฉฐ, ํด๋น ์ฌํญ์ ๋ํ ์ง์์ ๋ต๋ณ์ ๋ํด์๋ ๋น๋ฐ์ด ์ ์ง๋์ด์ผ ํฉ๋๋ค. ๊ธฐ์กด ์ธ๋ถ ํด๋ผ์ฐ๋์์ ์ ๊ณต๋๋ ์ธ์ด๋ชจ๋ธ์ ๊ฒฝ์ฐ ์ฌ๋ด ์ ๋ณด๊ฐ ์ ์ถ๋ ๊ฐ๋ฅ์ฑ์ ํต์ ํ ์ ์๋ ๊ธฐ์ ์ ์ธ ๋ฐฉ๋ฒ์ด ์์ผ๋ฏ๋ก, ์ธ์ด๋ชจ๋ธ์ ์ฌ๋ด์ ์ค์นํ์ฌ ์ฌ์ฉํ๋ ๋ฐฉ๋ฒ์ด ์ ์ผํฉ๋๋ค. LiOn์ ์ฌ๋ด์ ์ค์นํ์ฌ ์ฌ์ฉํ ์ ์๋ ๊ฒฝ๋ํ๋ ์ด๊ฑฐ๋ ์ธ์ด๋ชจ๋ธ๋ก์ ์ฌ๋ด์ ์ ๋ณด๋ฅผ ์์ ํ๊ฒ ์ ์งํ๋ฉด์ ๊ตฌ์ฑ์๋ค์ด ์์ ํ๊ฒ ์ฌ์ฉํ ์ ์๋ ๋์์ ์ ๊ณตํ ์ ์์ต๋๋ค. ์๋๋ ๊ทธ์ค ํ๋์ ์์์ด๋ฉฐ์ฌ๋ด์์์ ์ง์๋ค๊ณผ์ ๋ถํ์ ๋ํ ์๋ด์ ์์ด LiOn์ด ์๋ดํ๋ ์ฌ๋ก๋ฅผ ๋ณด์ค ์ ์์ต๋๋ค. ์ด ์ธ์๋ LiOn์ ์ฌ๋ด์์ ์ผ์ด๋ ์ ์๋ ์๋ง์ ์ํฉ์์ ๋ค์ํ ํด๊ฒฐ๋ฐฉ๋ฒ์ ์ ๊ณตํจ์ผ๋ก์ 24/7 ๊ตฌ์ฑ์๋ค์ ์ ๋ฌด๋ฅผ ๋๋ ๊ฒ์ด ๊ฐ๋ฅํฉ๋๋ค.
-
Free Dolly: Introducing the World's First Truly Open Instruction-Tuned LLM
-
Language Is Not All You Need: Aligning Perception with Language Models
-
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
-
Do Prompt-Based Models Really Understand the Meaning of their Prompts?
-
Improving language models by retrieving from trillions of tokens
-
Structure and Content-Guided Video Synthesis with Diffusion Models
-
InstructGPT : Training language models to follow instructions with human feedback
-
BioGPT: Generative Pre-trained Transformer for Biomedical Text Generation and Mining
-
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
-
What learning algorithm is in-context learning? Investigations with linear models
-
Toolformer: Language Models Can Teach Themselves to Use Tools
-
Improving alignment of dialogue agents via targeted human judgements
-
RLHF: Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
-
BaGuaLu: Targeting Brain Scale Pretrained Models with over 37 Million Cores
-
Flamingo: a Visual Language Model for Few-Shot Learning, Blog
-
Scaling Language Models: Methods, Analysis & Insights from Training Gopher
-
GPU and learning method required for KoChatLlaMA fine-tuning
-
GPT-4 is coming next week โ and it will be multimodal, says Microsoft Germany
-
Google USM: Scaling Automatic Speech Recognition Beyond 100 Languages
-
Tightly-Integrated Generative Encoder-Decoder Representation
-
Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models
-
SpikeGPT: Generative Pre-trained Language Model with Spiking Neural Networks
-
Low-rank Adaptation for Fast Text-to-Image Diffusion Fine-tuning
-
Improving language models by retrieving from trillions of tokens
-
T0: Multitask Prompted Training Enables Zero-Shot Task Generalization
-
The Flan Collection: Designing Data and Methods for Effective Instruction Tuning
-
The Wisdom of Hindsight Makes Language Models Better Instruction Followers
-
Exploring the Benefits of Training Expert Language Models over Instruction Tuning
-
Unsupervised Imputation of Non-ignorably Missing Data Using Importance-Weighted Autoencoders
-
Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
-
Do Prompt-Based Models Really Understand the Meaning of their Prompts?
-
Muse: Text-To-Image Generation via Masked Generative Transformers
-
Structure and Content-Guided Video Synthesis with Diffusion Models
-
Accurate global machine learning force fields for molecules with hundreds of atoms
-
Algorithms with More Granular Differential Privacy Guarantees
-
Anomaly Clustering: Grouping Images into Coherent Clusters of Anomaly Types
-
Are we cobblers without shoes? Making Computer Science data FAIR
-
Creating, Calibrating, and Validating Large-Scale Microscopic Traffic Simulation
-
Increasing Impact of Mobile Health Programs: SAHELI for Maternal and Child Care
-
Designing Responsible AI: Adaptations of UX Practice to Meet Responsible AI Challenges
-
Developer Productivity for Humans: A Human-Centered Approach to Developer Productivity
-
Development of a Machine Learning Model for Sonographic Assessment of Gestational Age
-
Estimates of broadband upwelling irradiance from GOES-16 ABI
-
Flexible Budgets in Restless Bandits: A Primal-Dual Algorithm for Efficient Budget Allocation
-
Helpful Neighbors: Leveraging Neighbors in Geographic Feature Pronunciation
-
High-Performance GPU-to-CPU Transpilation and Optimization via High-Level Parallel Constructs
-
Helpful Neighbors: Leveraging Neighbors in Geographic Feature Pronunciation
-
KwikBucks: Correlation Clustering with Cheap-Weak and Expensive-Strong Signals
-
Machine Learning for Healthcare: A Bibliometric Study of Contributions from Africa
-
Propeller: A Profile Guided, Relinking Optimizer for Warehouse-Scale Applications
-
Deepmind: Improving language models by retrieving from trillions of tokens
-
Deepmind: Mastering Stratego, the classic game of imperfect information
-
Deepmind: AlphaFold reveals the structure of the protein universe
-
Deepmind: Exploring the beauty of pure mathematics in novel ways
-
Deepmind: Putting the power of AlphaFold into the worldโs hands
-
Google Research: Deciphering clinical abbreviations with privacy protecting ML
-
Google Research: Google Research, 2022 & beyond: Language, vision and generative models
-
Google Research: Google Research, 2022 & beyond: Responsible AI
-
Google Research: Google Research, 2022 & beyond: ML & computer systems
-
Google Research: Real-time tracking of wildfire boundaries using satellite imagery
-
Google Research: DiffQG: Generating Questions on Paired Sentences
-
Google Research: Assessment of Security Defense of Native Programs Against Software Faults
-
Google Research: Adaptive mixing of auxiliary losses in supervised learning
-
[2013/01] Efficient Estimation of Word Representations in Vector Space
-
[2014/12] Dependency-Based Word Embeddings
-
[2015/07] Neural Machine Translation of Rare Words with Subword Units
-
[2014/07] GloVe: Global Vectors for Word Representation : GloVe
-
[2016/06] Siamese CBOW: Optimizing Word Embeddings for Sentence Representations : Siamese CBOW
-
[2016/07] Enriching Word Vectors with Subword Information : fastText
-
[2014/09] Sequence to Sequence Learningwith Neural Networks : seq2seq
-
[2017/07] Attention Is All You Need : Transformer
-
[2017/08] Learned in Translation: Contextualized Word Vectors : CoVe
-
[2018/01] Universal Language Model Fine-tuning for Text Classification : ULMFIT
-
[2018/02] Deep contextualized word representations : ELMo
-
[2018/06] Improving Language Understanding by Generative Pre-Training : GPT-1
-
[2018/10] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding : BERT
-
[2019/02] Language Models are Unsupervised Multitask Learners : GPT-2
-
[2019/04] Language Models with Transformers
-
[2019/01] Cross-lingual Language Model Pretraining XLM
-
[2019/01] Multi-Task Deep Neural Networks for Natural Language Understanding : MT-DNN
-
[2019/01] Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context : Transformer-XL
-
[2019/06] XLNet: Generalized Autoregressive Pretraining for Language Understanding : XLNet
-
[2019/09] Fine-Tuning Language Models from Human Preferences
-
[2019/01] BioBERT: a pre-trained biomedical language representation model for biomedical text mining : BioBERT
-
[2019/03] SciBERT: A Pretrained Language Model for Scientific Text : SciBERT
-
[2019/04] ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission : ClinicalBERT
-
[2019/06] HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization : HIBERT
-
[2019/07] SpanBERT: Improving Pre-training by Representing and Predicting Spans : SpanBERT
-
[2019/08] Pre-Training with Whole Word Masking for Chinese BERT
-
[2019/07] R-Transformer: Recurrent Neural Network Enhanced Transformer : R-Transformer
-
[2019/09] FREELB: ENHANCED ADVERSARIAL TRAINING FOR LANGUAGE UNDERSTANDING : FREELB
-
[2019/09] Mixup Inference: Better Exploiting Mixup to Defend Adversarial Attacks
-
[2019/10] Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer : T5
-
[2018/07] Subword-level Word Vector Representations for Korean
-
[2019/08] Zero-shot Word Sense Disambiguation using Sense Definition Embeddings
-
[2019/06] Bridging the Gap between Training and Inference for Neural Machine Translation
-
[2019/06] Emotion-Cause Pair Extraction: A New Task to Emotion Analysis in Texts
-
[2019/07] A Simple Theoretical Model of Importance for Summarization
-
[2019/05] Transferable Multi-Domain State Generator for Task-Oriented Dialogue Systems
-
[2019/07] We need to talk about standard splits
-
[2019/07] ERNIE 2.0: A Continual Pre-training Framework for Language Understanding : ERNIE 2.0
-
[2019/05] SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems : SuperGLUE
-
[2020/01] Towards a Human-like Open-Domain Chatbot + Google AI Blog
-
[2020/03] ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators : ELECTRA
-
[2019/04] Mask-Predict: Parallel Decoding of Conditional Masked Language Models : Mask-Predict
-
[2020/01] Reformer: The Efficient Transformer : Reformer
-
[2020/04] Longformer: The Long-Document Transformer : Longformer
-
[2019/11] DialoGPT: Large-Scale Generative Pre-training for Conversational Response Generation : DialoGPT
-
[2020/01] Towards a Human-like Open-Domain Chatbot
-
[2020/04] You Impress Me: Dialogue Generation via Mutual Persona Perception
-
[2020/04] ToD-BERT: Pre-trained Natural Language Understanding for Task-Oriented Dialogues : ToD-BERT
-
[2020/04] SOLOIST: Few-shot Task-Oriented Dialog with A Single Pre-trained Auto-regressive Model : SOLOIST
-
[2020/05] A Simple Language Model for Task-Oriented Dialogue
-
[2019/07] ReCoSa: Detecting the Relevant Contexts with Self-Attention for Multi-turn Dialogue Generation : ReCoSa
-
[2020/04] FastBERT: a Self-distilling BERT with Adaptive Inference Time : FastBERT
-
[2020/01] PoWER-BERT: Accelerating BERT inference for Classification Tasks : PoWER-BERT
-
[2019/10] DistillBERT, a distilled version of BERT: smaller, faster, cheaper and lighter : DistillBERT
-
[2019/10] TinyBERT: Distilling BERT for Natural Language Understanding : TinyBERT
-
[2018/12] Conditional BERT Contextual Augmentation
-
[2020/03] Data Augmentation using Pre-trained Transformer Models
-
[2020/04] FLAT: Chinese NER Using Flat-Lattice Transformer : FLAT
-
[2019/12] Big Transfer (BiT): General Visual Representation Learning : BiT
-
[2019/04] ERNIE: Enhanced Representation through Knowledge Integration : ERNIE
-
[2019/07] ERNIE 2.0: A Continual Pre-training Framework for Language Understanding : ERNIE 2.0
-
[2020/06] ERNIE-ViL: Knowledge Enhanced Vision-Language Representations Through Scene Graph : ERNIE-ViL
-
[2020/12] ERNIE-Doc: A Retrospective Long-Document Modeling Transformer : ERNIE-Doc
-
[2021/07] ERNIE 3.0: Large-scale Knowledge Enhanced Pre-training for Language Understanding and Generation : ERNIE 3.0
-
[2022/10] Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
-
[2017/03] Distilling Task-Specific Knowledge from BERT into Simple Neural Networks
-
[2020/10] DiPair: Fast and Accurate Distillation for Trillion-Scale Text Matching and Pair Modeling : DiPair
-
[2021/08] Distilling Transformers for Neural Cross-Domain Search
-
[2020/06] DeBERTa: Decoding-enhanced BERT with Disentangled Attention : DeBERTa
-
[2020/11] VEGA: Towards an End-to-End Configurable AutoML Pipeline : VEGA
-
[2020/12] FILTER: An Enhanced Fusion Method for Cross-lingual Language Understanding : FILTER
-
[2019/12] StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding : StructBERT
-
[2019/04] Improving Multi-Task Deep Neural Networks via Knowledge Distillation for Natural Language Understanding : MT-DNN
-
[2021/05] Comparing Kullback-Leibler Divergence and Mean Squared Error Loss in Knowledge Distillation
์ต์ MLLM ๊ด๋ จ ์คํฐ๋. ๊ธฐ๋ณธ ์คํ์ ์งํ. ๋ ผ๋ฌธ, ๊ฐ์, ์ฝ๋, ๋ด์ค, ๋ธ๋ก๊ทธ ๋ฑ ๋ค์ํ ์๋ฃ๋ก ํ์ต.
MLLM, LLM, NLG, Dialogue, Reinforcement learning, Distillation, Efficient, Sentence similarity, multiple tasks, multimodal, Stable diffusion, TTS, Text-To-Video, All-To-All, ์ฐ์ฃผ, ์๋ช , ์ง๋ฅ, ์ค๋ฆฌ, ๊ท์ , ๋ฒ, ๋ ธํ, ์ํ, ํฌ์, ๊ฐ๋ฐ, ์ธํ๋ผ, ๋์์ธ, ๊ฒฝ์, ETC...
์ ๋ง ์คํํธ์ C๋ ๋ฒจ, ๊ตญ๋ด์ธ ํํฐ์ด ์ฐ๊ตฌ์, ๊ตญ๋ด์ธ ํํฐ์ด ๋ํ, ๋ํ์ ์ฌํ์๊ณผ ์กธ์ ์, ์ํ, ๊ต์ ๋ฑ A๊ธ ์ธ์ฌ๋ค์ด ์ต์ ๋ ผ๋ฌธ, ๊ฐ์ ๋ฑ ์คํฐ๋ ๋ฐ ํ๋ก์ ํธ ์งํ.
๊ธฐ๋ณธ ๋งค์ฃผ ์์์ผ ์คํ 7์๋ฐ. ์ฌ์ ํ์ต ์์ด ๋ ผ๋ฌธ ์ฝ๊ธฐ ์ต๋ 20๋ถ, ํ ๋ก ์ต๋ 40๋ถ. ํ ๋ฒ์ 1 ~ 10๊ฐ ๋ ผ๋ฌธ, ๊ฐ์ ๋ฑ ์งํ. ์ง๊ธ๊น์ง๋ ํญ์ 3๊ฐ. ์ฃผ์ ๋ ผ๋ฌธ ์ ์ ์ ์์ . ํํฐ์ด ํํ ๋ ผ๋ฌธ ๋ฐ ํ๋ก์ ํธ ์ ์ ์์ .
์ฃผ๋ง์ ํฌํจํ์ฌ, ๊ฑฐ์ ๋งค์ผ ์ถ๊ฐ ์คํฐ๋ ์กด์ฌ. ํฅ๋ฏธ๋ก์ด ์ฃผ์ ๊ฑฐ๋ ์ฐธ์ฌ ๋๋ ๋ ๋ง ์ค๊ฐ์ ๋ค์ด์์ ์ค๊ฐ์ ๋๊ฐ๋ ๋ฌด๊ด. ๋ชจ๋ ๊ท์น์ ํ์ ๊ฐ๋ฅ. ์คํ๋ผ์ธ ๋ชจ์๋ ์์ . ์์จ ์ฐธ์ฌ.
- ์์ด๋ง ์ฌ์ฉ์ ๊ธ์ง. ํ๊ตญ์ด ์ค์ฌ ์ฌ์ฉ. ํน์ ์ฉ์ด๋ ์์ด ์ฌ์ฉ.
- 1์ฃผ์ผ์ ๋ ผ๋ฌธ 2๊ฐ ์ด์ ์คํฐ๋. ๋๋ ์ฌ๋์ 10๊ฐ ์ด์.
- 3๋ถ์์ 20๋ถ ํ์ฅ์์ ๋ ผ๋ฌธ ์ฝ๊ธฐ. 5๋ถ์์ 30๋ถ ํ ๋ก .
- 1์๊ฐ ์คํฐ๋ ์, ๋ฐ๋ก ๋๊ฐ๋ ๋จ. ์ํ ๋ 10๋ถ ์ดํ ์ฐธ์ฌ๋ ๋ฌด๊ด. ์์ ๋กญ๊ฒ ์งํ. 2์๊ฐ ๋งค์ผ๋ ๊ฐ๋ฅ.
- ๊ฐ์ ๋ ๋ฐ์ด๋ ๊ฒ ์๋ค๋ ๊ฒ์ ์ธ์ง. ๋ค๋ค ๋๋จํ ๋ถ๋ค์ด๋ ์ง๋ฌธ ๋ง์ด ํ๊ณ , ์ ๋ณด ๊ณต์ ์์ฃผ.
- ๋ณธ์ธ์ด ํ๊ธฐ๋ก ํ ์ผ๋ง์ ์ํ. ํ๋ค๊ณ ๋งํ๊ณ , ์ ํ๋ ๊ฒ์ ๋ฏผํ๋ค.
- ๊ธฐ๋ณธ์ ์ผ๋ก ๋ นํ ํ ๋ด๋ถ ๊ณต์ .
- ์ ๋ณด๋ฅผ ํผ์ ์๊ฒ ์ฐ์ง ๋ง๊ณ , ๋ค ๊ฐ์ด ์๊ฒ ๋งํ๊ธฐ.
- ๊ฐ์ธ ์ฌ์ ์ผ๋ก ์คํฐ๋ ํํด ์, ์๊ธฐ์๊ฐ์ ์ธ์ฌ ์์ฑ.
- ์ฌ๋ฌ ๊ธฐ๊ด ์ข์ ๊ท์น ๋ถ์ฌ๋ฃ๊ธฐ.
- ํ์ ๋์์ด ๋๋ค๊ณ ํ๋จํ๋ฉด, ์ ๊ท์น์ ๋ชจ๋ ๋ฌด์ํ๊ณ ํ๋.
- ์ถ๊ฐ.
mathematics | machine learning | Transformer | Hugging Face |
---|---|---|---|
mathematics for machine learning | Pattern Recognition and Machine Learning | Getting Started with Google BERT | Natural Language Processing with Transformers |