1. MinGPT 项目简介

  • MinGPT 是 GPT 模型的一个流行的开源 PyTorch 复现项目,包括训练和推理。该项目由特斯拉人工智能研究负责人、前 OpenAI 研究科学家 Andrej Karpathy 开发并于 2020 年在 github 发布。当时大多数可用的GPT模型实现都有点杂乱,而 minGPT 简洁干净可解释,因而颇具教育意义。MinGPT 的开源地址为:karpathy/minGPT
  • GPT 不是一个复杂的模型,它做的只是将一系列 token 输入到 Transformer 中,然后得到序列中下一个 token 的概率分布,大多数复杂性只是为了提高效率而巧妙地使用批处理 (both across examples and over sequence length)。minGPT 麻雀虽小,五脏俱全。如果说 GPT 模型是所向披靡的战舰,那么 minGPT 大概算是个个头虽小但仍能乘风破浪的游艇了吧。
    在这里插入图片描述

2. 相关论文

  • 这个库的 Readme 中简介了 GPT-1, GPT-2, GPT-3 的模型结构,有一定指导意义,这里贴过来
    在这里插入图片描述

2.1 GPT-1

  • 论文地址:Improving Language Understanding by Generative Pre-Training
  • GPT-1 模型大体遵循了原始 transformer,训练了 12 层的 Decoder-only Transformer、具备 masked-self attention head(768 维状态和 12 个注意力头),具体实现细节为
    • 我们的模型在很大程度上遵循了原始的Transformer工作
    • 我们训练了一个12层的 Decoder-only Transformer,使用了带有 masked-attention head 的结构(768维状态和12个注意力头)。对于 position-wise FFD 网络,我们使用了3072维的内部状态
    • Adam最大学习率为 2.5e-4。(之后GPT-3针对这个模型大小使用了6e-4)
    • 学习率衰减:在前2000次更新中线性增加到最大值,然后使用余弦计划将其退化为0
    • 我们使用 64 个随机抽样的连续序列,每个序列包含 512 个 torch,进行100个 epoch 的训练
    • 由于 Layernorm 在整个模型中被广泛使用,简单的 N ( 0 , 0.02 ) N(0, 0.02) N(0,0.02) 权重初始化就足够了
    • 使用了 40,000 merges 的 Byte Pair Encoding(BPE)词汇表。
    • 为了正则化,residual, embedding, 和 attention 都设置了 0.1 的 dropout
    • 采用了修改过的 L2 正则化方法,在所有非偏置或增益权重上设置 w = 0.01 w = 0.01 w=0.01
    • 激活函数采用了高斯误差线性单元(GELU)
    • 我们使用了学习得到的 position embedding,而不是原始工作中提出的正弦版本
    • 对于微调(finetuning):我们在分类器上添加了 0.1 的 dropout layer。学习率为 6.25e-5,批大小为 32,进行 3 个epoch的训练。我们使用线性学习率衰减计划,在训练的前 0.2% 部分进行 warmup,λ 设置为 0.5
    • GPT-1 模型有 12 层,嵌入维度 d_model=768,大约有 117M 个参数

2.2 GPT-2

  • 论文地址:Language Models are Unsupervised Multitask Learners
  • GPT-2 将 LayerNorm 移动到每个子模块的输入位置,类似于预激活残差网络,并在最后的自注意力模块后添加了一个额外的层归一化。此外,该模型还更改了模型初始化(包括残差层初始化权重等)、扩展了词汇量、将 context 规模从 512 个 token 增加到 1024、使用更大的 batch size 等。具体实现细节为:
    • LayerNorm 被移动到每个子块(attention layer、ffd layer…)的输入处,类似于预激活残差网络
    • 在最后的自注意力块之后添加了额外的 LayerNorm 操作
    • 考虑到模型深度上残差路径上的累积效应,将残差层的权重在初始化时缩放一个因子 1 / N 1/\sqrt{N} 1/N ,其中 N N N 是残差层的数量。(weird because in their released code i can only find a simple use of the old 0.02… in their release of image-gpt I found it used for c_proj, and even then only for attn, not for mlp. huh)
    • 词汇表扩展到 50,257
    • 将上下文大小从 512 增加到 1024 个 token
    • 使用更大的 batch size=512
    • GPT-2 模型有 48 层,嵌入维度 d_model=1600,大约有 1.542B 个参数

2.3 GPT-3

  • 论文地址:Language Models are Few-Shot Learners
  • GPT-3 使用了和 GPT-2 相同的模型和架构,区别在于 GPT-3 在 transformer 的各层上都使用了交替密集和局部带状稀疏的注意力模式,类似于 Sparse Transformer。具体实现细节为:
    • 我们使用与GPT-2相同的模型和架构,包括其中描述的修改初始化、预归一化和可逆标记化
    • 在 Transformer 的各个层中,我们使用交替的 dense 和 locally banded sparse 注意力模式,类似于Sparse Transformer
    • 我们的 FFD layer 尺寸始终是瓶颈层大小的四倍,dff = 4 * d_model
    • 将上下文长度增加到 2048 个 token
    • Adam优化器的参数设置为 β1 = 0.9,β2 = 0.95,eps = 10^-8。
    • 所有模型使用 0.1 的权重 dropout 进行轻微正则化(注意:我相信GPT-1使用的是0.01,请参考上文)
    • 将梯度的全局范数剪切到 1.0
    • 在前 375 亿个 token 进行线性学习率 warmup,然后在 2600亿个 token 的时间内余弦衰减到其值的 10%
    • 在训练的前 40-120 亿个 token 之间,根据模型尺寸,从小值(32k token)逐渐线性增加 batch size
    • 始终使用完整的 2048 大小的上下文窗口,并添加特殊的 “END OF DOCUMENT” 标记分隔符。

3. 代码详解

3.1 项目结构

  • 看一下项目的目录结构
    在这里插入图片描述

  • minGPT 库由三个文件组成

    1. minGPT/model.py 包含实际的 Transformer 模型定义
    2. minGPT/bpe.py 包含一个稍微重构的 Byte Pair 编码器,它用于将文本转换为整数序列,为 tokenize 做准备
    3. minGPT/trainer.py 是训练模型的 (与GPT无关的) PyTorch样板代码
  • 在 projects 文件夹中给出了一些使用 minGPT 库的 demos 和 projects,包括

    1. projects/adder 从零训练 GPT 做加法(灵感来自GPT-3论文中的添加部分)
    2. projects/chargpt 用一些输入文本将 GPT 训练成一个 character-level language model
    3. demo.ipynb 在一个简单的排序示例中以笔记本格式展示了 GPTTrainer 的 minimal usage
    4. generate.ipynb 展示了如何加载预训练的 GPT2 并在给定 prompt 的情况下生成文本

3.2 GPT 模型代码详解

  • 本段详细分析 minGPT/model.py 中的 GPT 模型实现

3.2.1 Transformer block

  • 先看下 Transformer block 的定义,它是 GPT 的核心组件

    class Block(nn.Module):
        """ an unassuming Transformer block """
    
        def __init__(self, config):
            super().__init__()
            self.ln_1 = nn.LayerNorm(config.n_embd)
            self.attn = CausalSelfAttention(config)
            self.ln_2 = nn.LayerNorm(config.n_embd)
            self.mlp = nn.ModuleDict(dict(
                c_fc    = nn.Linear(config.n_embd, 4 * config.n_embd),
                c_proj  = nn.Linear(4 * config.n_embd, config.n_embd),
                act     = NewGELU(),
                dropout = nn.Dropout(config.resid_pdrop),
            ))
            m = self.mlp
            self.mlpf = lambda x: m.dropout(m.c_proj(m.act(m.c_fc(x)))) # MLP forward
    
        def forward(self, x):
            x = x + self.attn(self.ln_1(x))
            x = x + self.mlpf(self.ln_2(x))
            return x
    

    结构很清晰,如下图所示
    在这里插入图片描述
    可见使用了 GPT2/3 的 LayerNorm 前置结构,其中两个核心模块是 masked self-attention 层 self.attn 和 FFD 层 self.mlp

  • FFD 层:很简单就是一个带 dropout 正则的两层 MLP,特殊之处在于使用了 gelu 激活函数,如下所示

    class NewGELU(nn.Module):
        """
        Implementation of the GELU activation function currently in Google BERT repo (identical to OpenAI GPT).
        Reference: Gaussian Error Linear Units (GELU) paper: https://arxiv.org/abs/1606.08415
        """
        def forward(self, x):
            return 0.5 * x * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (x + 0.044715 * torch.pow(x, 3.0))))
    

    这个激活函数是对 relu 的一个优化,增加了更多非线性成分,如下所示
    在这里插入图片描述

  • masked self attention 层:这是 GPT 模型的核心,需要计算因果注意力

    class CausalSelfAttention(nn.Module):
        """
        A vanilla multi-head masked self-attention layer with a projection at the end.
        It is possible to use torch.nn.MultiheadAttention here but I am including an
        explicit implementation here to show that there is nothing too scary here.
        """
    
        def __init__(self, config):
            super().__init__()
            assert config.n_embd % config.n_head == 0
            # key, query, value projections for all heads, but in a batch
            self.c_attn = nn.Linear(config.n_embd, 3 * config.n_embd)
            # output projection
            self.c_proj = nn.Linear(config.n_embd, config.n_embd)
            # regularization
            self.attn_dropout = nn.Dropout(config.attn_pdrop)
            self.resid_dropout = nn.Dropout(config.resid_pdrop)
            # causal mask to ensure that attention is only applied to the left in the input sequence
            self.register_buffer("bias", torch.tril(torch.ones(config.block_size, config.block_size))
                                         .view(1, 1, config.block_size, config.block_size))
            self.n_head = config.n_head
            self.n_embd = config.n_embd
    
        def forward(self, x):
            B, T, C = x.size() # batch size, sequence length, embedding dimensionality (n_embd)
    
            # calculate query, key, values for all heads in batch and move head forward to be the batch dim
            q, k ,v  = self.c_attn(x).split(self.n_embd, dim=2)
            k = k.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs)
            q = q.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs)
            v = v.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs)
    
            # causal self-attention; Self-attend: (B, nh, T, hs) x (B, nh, hs, T) -> (B, nh, T, T)
            att = (q @ k.transpose(-2, -1)) * (1.0 / math.sqrt(k.size(-1)))
            att = att.masked_fill(self.bias[:,:,:T,:T] == 0, float('-inf'))
            att = F.softmax(att, dim=-1)
            att = self.attn_dropout(att)
            y = att @ v # (B, nh, T, T) x (B, nh, T, hs) -> (B, nh, T, hs)
            y = y.transpose(1, 2).contiguous().view(B, T, C) # re-assemble all head outputs side by side
    
            # output projection
            y = self.resid_dropout(self.c_proj(y))
            return y
    
    1. 每个注意力头提取的 value,query 和 key 的维度 hs 为嵌入维度 n_embd 的 1/n_head。虽然编程时这里仅需保证 query 和 key 的空间维度一致即可,但是考虑到嵌入维度和 vocab table 大小的关系,每个注意力头处理的嵌入维度不宜过大,这样和 n_embd 绑定是比较方便的做法。请参考:最小熵原理(六):词向量的维度应该怎么选择?
    2. masked self attention 通过一个下三角矩阵 self.bias 来实现,序列长度(config.block_size)为 5 时该张量形如
      tensor([[[[1., 0., 0., 0., 0.],
                [1., 1., 0., 0., 0.],
                [1., 1., 1., 0., 0.],
                [1., 1., 1., 1., 0.],
                [1., 1., 1., 1., 1.]]]])
      
      后面 att.masked_fill(self.bias[:,:,:T,:T] == 0, float('-inf')) 这句就会把这些为 0 的位置替换为 -inf,从而使这些位置在经过 softmax 后权重趋近 0,实现对未来序列的遮盖。另需注意 bias 张量是通过 self.register_buffer 方法登记的,这样登记过的张量可以求梯度也可以随模型在 CPU/GPU 之间移动,但是不进行参数优化

3.2.2 GPT

  • 先看初始化部分
    class GPT(nn.Module):
        """ GPT Language Model """
    
        @staticmethod
        def get_default_config():
            C = CfgNode()
            # either model_type or (n_layer, n_head, n_embd) must be given in the config
            C.model_type = 'gpt'
            C.n_layer = None
            C.n_head = None
            C.n_embd =  None
            # these options must be filled in externally
            C.vocab_size = None
            C.block_size = None
            # dropout hyperparameters
            C.embd_pdrop = 0.1
            C.resid_pdrop = 0.1
            C.attn_pdrop = 0.1
            return C
    
        def __init__(self, config):
            super().__init__()
            assert config.vocab_size is not None
            assert config.block_size is not None
            self.block_size = config.block_size
    
            type_given = config.model_type is not None
            params_given = all([config.n_layer is not None, config.n_head is not None, config.n_embd is not None])
            assert type_given ^ params_given # exactly one of these (XOR)
            if type_given:
                # translate from model_type to detailed configuration
                config.merge_from_dict({
                    # names follow the huggingface naming conventions
                    # GPT-1
                    'openai-gpt':   dict(n_layer=12, n_head=12, n_embd=768),  # 117M params
                    # GPT-2 configs
                    'gpt2':         dict(n_layer=12, n_head=12, n_embd=768),  # 124M params
                    'gpt2-medium':  dict(n_layer=24, n_head=16, n_embd=1024), # 350M params
                    'gpt2-large':   dict(n_layer=36, n_head=20, n_embd=1280), # 774M params
                    'gpt2-xl':      dict(n_layer=48, n_head=25, n_embd=1600), # 1558M params
                    # Gophers
                    'gopher-44m':   dict(n_layer=8, n_head=16, n_embd=512),
                    # (there are a number more...)
                    # I made these tiny models up
                    'gpt-mini':     dict(n_layer=6, n_head=6, n_embd=192),
                    'gpt-micro':    dict(n_layer=4, n_head=4, n_embd=128),
                    'gpt-nano':     dict(n_layer=3, n_head=3, n_embd=48),
                }[config.model_type])
    
            self.transformer = nn.ModuleDict(dict(
                wte = nn.Embedding(config.vocab_size, config.n_embd),
                wpe = nn.Embedding(config.block_size, config.n_embd),
                drop = nn.Dropout(config.embd_pdrop),
                h = nn.ModuleList([Block(config) for _ in range(config.n_layer)]),
                ln_f = nn.LayerNorm(config.n_embd),
            ))
            self.lm_head = nn.Linear(config.n_embd, config.vocab_size, bias=False)
    
            # init all weights, and apply a special scaled init to the residual projections, per GPT-2 paper
            self.apply(self._init_weights)
            for pn, p in self.named_parameters():
                if pn.endswith('c_proj.weight'):
                    torch.nn.init.normal_(p, mean=0.0, std=0.02/math.sqrt(2 * config.n_layer))
    
            # report number of parameters (note we don't count the decoder parameters in lm_head)
            n_params = sum(p.numel() for p in self.transformer.parameters())
            print("number of parameters: %.2fM" % (n_params/1e6,))
    
        def _init_weights(self, module):
            if isinstance(module, nn.Linear):
                torch.nn.init.normal_(module.weight, mean=0.0, std=0.02)
                if module.bias is not None:
                    torch.nn.init.zeros_(module.bias)
            elif isinstance(module, nn.Embedding):
                torch.nn.init.normal_(module.weight, mean=0.0, std=0.02)
            elif isinstance(module, nn.LayerNorm):
                torch.nn.init.zeros_(module.bias)
                torch.nn.init.ones_(module.weight)
    
    1. 作者使用了一个 CfgNode 对象来管理参数,这个是一个类似 yacs 的轻量级参数管理类对象,具体实现在 minGPT/utils.py 文件中,请自行查阅。另外注意到作者设置了一系列预制参数,可以轻松定义不同规模的 GPT 模型
    2. GPT 模型的所有组件包含在 self.transformer 成员中,wtewpe 分别是 token embedding 和 position embedding 模块,注意到它们都是 nn.Embedding 层;’h 中包含了所有堆叠的 transformer block 层;另外 self.lm_head 是 GPT 模型最后的分类头,GPT 模型宏观上就是一个输出空间为 vocab_size 的分类器
    3. 接下来作者通过 self.apply(self._init_weights) 对模型参数进行了初始化,这一通用性是比较强的,可以记录下复用到自己的项目中
  • 下面作者提供了一个方法用于加载预训练的 gpt2 check point,基本都是数据形式变换,这里不多讲了
    @classmethod
    def from_pretrained(cls, model_type):
        """
        Initialize a pretrained GPT model by copying over the weights
        from a huggingface/transformers checkpoint.
        """
        assert model_type in {'gpt2', 'gpt2-medium', 'gpt2-large', 'gpt2-xl'}
        from transformers import GPT2LMHeadModel
    
        # create a from-scratch initialized minGPT model
        config = cls.get_default_config()
        config.model_type = model_type
        config.vocab_size = 50257 # openai's model vocabulary
        config.block_size = 1024  # openai's model block_size
        model = GPT(config)
        sd = model.state_dict()
    
        # init a huggingface/transformers model
        model_hf = GPT2LMHeadModel.from_pretrained(model_type)
        sd_hf = model_hf.state_dict()
    
        # copy while ensuring all of the parameters are aligned and match in names and shapes
        keys = [k for k in sd_hf if not k.endswith('attn.masked_bias')] # ignore these
        transposed = ['attn.c_attn.weight', 'attn.c_proj.weight', 'mlp.c_fc.weight', 'mlp.c_proj.weight']
        # basically the openai checkpoints use a "Conv1D" module, but we only want to use a vanilla nn.Linear.
        # this means that we have to transpose these weights when we import them
        assert len(keys) == len(sd)
        for k in keys:
            if any(k.endswith(w) for w in transposed):
                # special treatment for the Conv1D weights we need to transpose
                assert sd_hf[k].shape[::-1] == sd[k].shape
                with torch.no_grad():
                    sd[k].copy_(sd_hf[k].t())
            else:
                # vanilla copy over the other parameters
                assert sd_hf[k].shape == sd[k].shape
                with torch.no_grad():
                    sd[k].copy_(sd_hf[k])
    
        return model
    
  • 接下来作者实现了一个优化器构造方法
    def configure_optimizers(self, train_config):
        """
        This long function is unfortunately doing something very simple and is being very defensive:
        We are separating out all parameters of the model into two buckets: those that will experience
        weight decay for regularization and those that won't (biases, and layernorm/embedding weights).
        We are then returning the PyTorch optimizer object.
        """
    
        # separate out all parameters to those that will and won't experience regularizing weight decay
        decay = set()
        no_decay = set()
        whitelist_weight_modules = (torch.nn.Linear, )
        blacklist_weight_modules = (torch.nn.LayerNorm, torch.nn.Embedding)
        for mn, m in self.named_modules():
            for pn, p in m.named_parameters():
                fpn = '%s.%s' % (mn, pn) if mn else pn # full param name
                # random note: because named_modules and named_parameters are recursive
                # we will see the same tensors p many many times. but doing it this way
                # allows us to know which parent module any tensor p belongs to...
                if pn.endswith('bias'):
                    # all biases will not be decayed
                    no_decay.add(fpn)
                elif pn.endswith('weight') and isinstance(m, whitelist_weight_modules):
                    # weights of whitelist modules will be weight decayed
                    decay.add(fpn)
                elif pn.endswith('weight') and isinstance(m, blacklist_weight_modules):
                    # weights of blacklist modules will NOT be weight decayed
                    no_decay.add(fpn)
    
        # validate that we considered every parameter
        param_dict = {pn: p for pn, p in self.named_parameters()}
        inter_params = decay & no_decay
        union_params = decay | no_decay
        assert len(inter_params) == 0, "parameters %s made it into both decay/no_decay sets!" % (str(inter_params), )
        assert len(param_dict.keys() - union_params) == 0, "parameters %s were not separated into either decay/no_decay set!" \
                                                    % (str(param_dict.keys() - union_params), )
    
        # create the pytorch optimizer object
        optim_groups = [
            {"params": [param_dict[pn] for pn in sorted(list(decay))], "weight_decay": train_config.weight_decay},
            {"params": [param_dict[pn] for pn in sorted(list(no_decay))], "weight_decay": 0.0},
        ]
        optimizer = torch.optim.AdamW(optim_groups, lr=train_config.learning_rate, betas=train_config.betas)
        return optimizer
    
    这里主要是通过权重衰减方法来进行正则化,避免过拟合。注意到作者通过一个二重遍历考察 GPT 模型所有 sub module 的所有 parameters,仅对所有 torch.nn.Linear 层的 weight 参数进行衰减,bias 参数及所有 torch.nn.LayerNormtorch.nn.Embedding 模块的参数都不做处理。由于模块是递归组织的,这个二重变量会重复访问很多参数,所以通过 set 自动去重,最后根据处理结果定义 torch.optim.AdamW 优化器返回

    关于权重衰减的理论说明,参考:机器学习基础(6)—— 使用权重衰减和丢弃法缓解过拟合问题

  • 下面就是 GPT 模型的前向方法
     def forward(self, idx, targets=None):
         device = idx.device
         b, t = idx.size()
         assert t <= self.block_size, f"Cannot forward sequence of length {t}, block size is only {self.block_size}"
         pos = torch.arange(0, t, dtype=torch.long, device=device).unsqueeze(0) # shape (1, t)
    
        # forward the GPT model itself
        tok_emb = self.transformer.wte(idx)             # (b, t, n_embd)
        pos_emb = self.transformer.wpe(pos)             # (1, t, n_embd)
        x = self.transformer.drop(tok_emb + pos_emb)    # (b, t, n_embd)
        for block in self.transformer.h:
            x = block(x)
        x = self.transformer.ln_f(x)                    # (b, t, n_embd)
        logits = self.lm_head(x)                        # (b, t, vocab_size)
    
        # if we are given some desired targets also calculate the loss
        loss = None
        if targets is not None:
            loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1), ignore_index=-1)
    
        return logits, loss
    
    1. 流程很清晰:输入一个 batch 长度为 t 的 vocab index 序列 idx,用 wte 进行 embedding 并加上 wpe 位置编码后得到尺寸 (batch_size, t, n_embd) 的 token embedding 序列 x,经过若干 transformer block 和后置 layer norm 层 ln_f,最后经过分类头 self.lm_head 得到尺寸 (batch_size, t, n_embd) 的预测序列 logits
    2. 注意 GPT 模型会对每个输入 token 产生对应位置的输出,最后 logits 序列长度为 t
    3. 使用 teacher-forcing 方式进行训练,所以需要输入尺寸 (batch_size, t) 的 targets序列。将 xtargets 都拉平来计算分类的交叉熵损失。注意交叉熵函数中设置了 ignore_index=-1,这样我们可以对 target 序列设置一些 -1 来 mask 掉无需计算 loss 的 prompt 序列
  • 最后来看 GPT 模型的预测方法
      @torch.no_grad()
      def generate(self, idx, max_new_tokens, temperature=1.0, do_sample=False, top_k=None):
          """
          Take a conditioning sequence of indices idx (LongTensor of shape (b,t)) and complete
          the sequence max_new_tokens times, feeding the predictions back into the model each time.
          Most likely you'll want to make sure to be in model.eval() mode of operation for this.
          """
          for _ in range(max_new_tokens):
              # if the sequence context is growing too long we must crop it at block_size
              idx_cond = idx if idx.size(1) <= self.block_size else idx[:, -self.block_size:] # (batch_size, seq_len)
              # forward the model to get the logits for the index in the sequence
              logits, _ = self(idx_cond)                                                      # (batch_size, seq_len, vocab_size)
              # pluck the logits at the final step and scale by desired temperature
              logits = logits[:, -1, :] / temperature                                         # (batch_size, vocab_size)
              # optionally crop the logits to only the top k options
              if top_k is not None:
                  v, _ = torch.topk(logits, top_k)
                  logits[logits < v[:, [-1]]] = -float('Inf')
              # apply softmax to convert logits to (normalized) probabilities
              probs = F.softmax(logits, dim=-1)                                               # (batch_size, vocab_size)
              # either sample from the distribution or take the most likely element
              if do_sample:
                  idx_next = torch.multinomial(probs, num_samples=1)                          # (batch_size, 1)
              else:
                  _, idx_next = torch.topk(probs, k=1, dim=-1)
              # append sampled index to the running sequence and continue
              idx = torch.cat((idx, idx_next), dim=1)                                         # (batch_size, seq_len+1)
    
          return idx
    
    注意这里是 AutoRegress 框架:
    1. 输入尺寸为 (batch_size, seq_len) 的 vocab index 序列 idx,这时序列长度 seq_len 完全是 prompt 长度
    2. 将其经过 GPT 前向过程得到尺寸为 (batch_size, seq_len, vocab_size) 的预测 logits
    3. 通过 logits = logits[:, -1, :] / temperature 拿出最后一个位置的预测
    4. 根据生成方法设置(是否仅保留topk候选、是否采样生成…)得到预测的 token 结果 idx_next,将其拼接回原序列
    5. 回到第一步,自回归地生成下一个 token

3.3 两位数加法实验

  • 本节介绍 projects/adder.py 中的两位数加法实验

3.3.1 数据集构造

  • 本实验仅考虑两个不超过两位的整数间的加法,这样计算的结果可能是两位数或者三位数。下面给出两个加法算式以及其对应的序列样本
    85 + 50 = 135 ⟶ 8550531 6 + 39 = 45 ⟶ 0639054 85 + 50 = 135 \longrightarrow 8550531 \\ 6 + 39 = 45 \longrightarrow 0639054 85+50=13585505316+39=450639054

    1. 丢弃了不必要的 + + + = = =,仅对数字进行编码
    2. 结果被反向编码,这样加法更容易学习(回想一下竖式加法计算的过程)
    3. 进行 zero-padding 以确保总是生成长度为 2+2+(2+1)=7 的序列
    4. 在训练和评估过程中,我们总是输入前 2+2 个 token,并希望 GPT 模型预测出接下来的 (2+1) 位完成序列
  • 下面给出数据集的构造代码

    class AdditionDataset(Dataset):
        """
        Creates n-digit addition problems. For example, if n=2, then an example
        addition problem would be to add 85 + 50 = 135. This problem would be
        represented as the following string for the GPT:
    
        "8550531"
    
        This is because:
        - we are discarding the + and =, which are not necessary. We just encode the digits
          of the input numbers concatenated together.
        - the result 135 is encoded backwards to make the addition easier to learn for the
          GPT model, because of how the addition algorithm works.
    
        As one more example, the problem 6 + 39 = 45 would be encoded as:
    
        "0639054"
    
        where you will notice that we are padding with zeros to make sure that we always
        produce strings of the exact same size: n + n + (n + 1). When n=2, this is 7.
        At test time, we will feed in an addition problem by giving the first 2n digits,
        and hoping that the GPT model completes the sequence with the next (n+1) digits
        correctly.
        """
    
        @staticmethod
        def get_default_config():
            C = CN()
            C.ndigit = 2
            return C
    
        def __init__(self, config, split):
            self.config = config
            self.split = split # train/test
    
            # split up all addition problems into either training data or test data
            ndigit = self.config.ndigit
            assert ndigit <= 3, "the lines below would be very memory inefficient, in future maybe refactor to support"
            num = (10**ndigit)**2 # total number of possible addition problems with ndigit numbers
            rng = torch.Generator()
            rng.manual_seed(1337)
            perm = torch.randperm(num, generator=rng)
            num_test = min(int(num*0.2), 500) # 20% of the whole dataset, or only up to 500
            self.ixes = perm[:num_test] if split == 'test' else perm[num_test:]
    
        def get_vocab_size(self):
            return 10 # digits 0..9
    
        def get_block_size(self):
            # a,b,a+b, and +1 due to potential carry overflow,
            # but then also -1 because very last digit doesn't ever plug back
            # as there is no explicit <EOS> token to predict, it is implied
            return 3*self.config.ndigit + 1 - 1
    
        def __len__(self):
            return self.ixes.nelement()
    
        def __getitem__(self, idx):
            ndigit = self.config.ndigit
            # given a problem index idx, first recover the associated a + b
            idx = self.ixes[idx].item()
            nd = 10**ndigit
            a = idx // nd
            b = idx %  nd
            # calculate the "label" of the addition problem a + b
            c = a + b
            # encode the digits of a, b, c into strings
            astr = f'%0{ndigit}d' % a
            bstr = f'%0{ndigit}d' % b
            cstr = (f'%0{ndigit+1}d' % c)[::-1] # reverse c to make addition easier
            render = astr + bstr + cstr
            dix = [int(s) for s in render] # convert each character to its token index
            # x will be input to GPT and y will be the associated expected outputs
            x = torch.tensor(dix[:-1], dtype=torch.long)
            y = torch.tensor(dix[1:], dtype=torch.long) # predict the next token in the sequence
            y[:ndigit*2-1] = -1 # we will only train in the output locations. -1 will mask loss to zero
            return x, y
    
  • 打印一些生成的样本和标签看一看

    (tensor([9, 9, 3, 2, 1, 3]), tensor([-1, -1, -1,  1,  3,  1]))
    (tensor([5, 6, 2, 1, 7, 7]), tensor([-1, -1, -1,  7,  7,  0]))
    (tensor([5, 5, 0, 3, 8, 5]), tensor([-1, -1, -1,  8,  5,  0]))
    (tensor([3, 8, 6, 2, 0, 0]), tensor([-1, -1, -1,  0,  0,  1]))
    (tensor([9, 9, 0, 5, 4, 0]), tensor([-1, -1, -1,  4,  0,  1]))
    (tensor([8, 9, 9, 6, 5, 8]), tensor([-1, -1, -1,  5,  8,  1]))
    (tensor([4, 0, 8, 9, 9, 2]), tensor([-1, -1, -1,  9,  2,  1]))
    ...
    

    这一下子可能看不清,我把第一条和第三条的 AutoRegress 过程示意图放在下面

    99 + 32 = 131 
    ------------------------
    -1  -1  -1  1   3   1
    9   9   3   2   1   3   1
    
    
    55 + 3 = 58
    ------------------------
    -1  -1  -1  8   5   0
    5   5   0   3   8   5   0
    

    注意设为 -1 的部分被 mask 掉不计算 loss(联系 3.2.2 节的 GPT forward 函数中损失计算部分)

3.3.2 训练器

  • 在 trainer 类中实现训练循环
    class Trainer:
    
        @staticmethod
        def get_default_config():
            C = CN()
            # device to train on
            C.device = 'auto'
            # dataloder parameters
            C.num_workers = 4
            # optimizer parameters
            C.max_iters = None
            C.batch_size = 64
            C.learning_rate = 3e-4
            C.betas = (0.9, 0.95)
            C.weight_decay = 0.1 # only applied on matmul weights
            C.grad_norm_clip = 1.0
            return C
    
        def __init__(self, config, model, train_dataset):
            self.config = config
            self.model = model
            self.optimizer = None
            self.train_dataset = train_dataset
            self.callbacks = defaultdict(list)
    
            # determine the device we'll train on
            if config.device == 'auto':
                self.device = 'cuda' if torch.cuda.is_available() else 'cpu'
            else:
                self.device = config.device
            self.model = self.model.to(self.device)
            print("running on device", self.device)
    
            # variables that will be assigned to trainer class later for logging and etc
            self.iter_num = 0
            self.iter_time = 0.0
            self.iter_dt = 0.0
    
        def add_callback(self, onevent: str, callback):
            self.callbacks[onevent].append(callback)
    
        def set_callback(self, onevent: str, callback):
            self.callbacks[onevent] = [callback]
    
        def trigger_callbacks(self, onevent: str):
            for callback in self.callbacks.get(onevent, []):
                callback(self)
    
        def run(self):
            model, config = self.model, self.config
    
            # setup the optimizer
            self.optimizer = model.configure_optimizers(config)
    
            # setup the dataloader
            train_loader = DataLoader(
                self.train_dataset,
                sampler=torch.utils.data.RandomSampler(self.train_dataset, replacement=True, num_samples=int(1e10)),
                shuffle=False,
                pin_memory=True,
                batch_size=config.batch_size,
                num_workers=config.num_workers,
            )
    
            model.train()
            self.iter_num = 0
            self.iter_time = time.time()
            data_iter = iter(train_loader)
            while True:
    
                # fetch the next batch (x, y) and re-init iterator if needed
                try:
                    batch = next(data_iter)
                except StopIteration:
                    data_iter = iter(train_loader)
                    batch = next(data_iter)
                batch = [t.to(self.device) for t in batch]
                x, y = batch
    
                # forward the model
                logits, self.loss = model(x, y)
    
                # backprop and update the parameters
                model.zero_grad(set_to_none=True)
                self.loss.backward()
                torch.nn.utils.clip_grad_norm_(model.parameters(), config.grad_norm_clip)
                self.optimizer.step()
    
                self.trigger_callbacks('on_batch_end')
                self.iter_num += 1
                tnow = time.time()
                self.iter_dt = tnow - self.iter_time
                self.iter_time = tnow
    
                # termination conditions
                if config.max_iters is not None and self.iter_num >= config.max_iters:
                    break
    
    这个训练实现得相当直观,只需注意
    1. 这里不是按把训练数据集轮一遍算做一个 epoch,训练若干 epoch 这样,而是设定一个目标的迭代周期 config.max_iters,不断轮询训练集直到训练 batch 数量达标为止
    2. 作者注册了一个回调函数 self.trigger_callbacks('on_batch_end'),每个 batch 训练后调用它,这个函数的函数体将在训练主函数中实现

3.3.3 模型参数设置

  • 作者使用 CfgNode 对象来管理参数,这是个类似 yacs 的轻量级参数管理类对象,具体实现在 minGPT/utils.py 文件中,请自行查阅
    def get_config():
    
        C = CN()
    
        # system
        C.system = CN()
        C.system.seed = 3407
        C.system.work_dir = './out/adder'
    
        # data
        C.data = AdditionDataset.get_default_config()
    
        # model
        C.model = GPT.get_default_config()
        C.model.model_type = 'gpt-nano'
    
        # trainer
        C.trainer = Trainer.get_default_config()
        C.trainer.learning_rate = 5e-4 # the model we're using is so small that we can go a bit faster
    
        return C
    

3.3.4 训练过程

  • 没有特别值得说的,基本就是定义数据集、定义 trainer、定义模型,开始训练后按一定周期建议模型性能。直接贴代码吧
    if __name__ == '__main__':
    
        # get default config and overrides from the command line, if any
        config = get_config()
        config.merge_from_args(sys.argv[1:])
        print(config)
        setup_logging(config)
        set_seed(config.system.seed)
    
        # construct train and test datasets
        train_dataset = AdditionDataset(config.data, split='train')
        test_dataset  = AdditionDataset(config.data, split='test')
    
        # construct the model
        config.model.vocab_size = train_dataset.get_vocab_size()
        config.model.block_size = train_dataset.get_block_size()
        model = GPT(config.model)
    
        # construct the trainer object
        config.trainer.max_iters = 5001
        trainer = Trainer(config.trainer, model, train_dataset)
    
        train_losses = []
        accuracies = []
    
        # helper function for the evaluation of a model
        def eval_split(trainer, split, max_batches=None):
            dataset = {'train':train_dataset, 'test':test_dataset}[split]
            ndigit = config.data.ndigit
            results = []
            mistakes_printed_already = 0
            factors = torch.tensor([[10**i for i in range(ndigit+1)][::-1]]).to(trainer.device)
            loader = DataLoader(dataset, batch_size=100, num_workers=0, drop_last=False)
            for b, (x, y) in enumerate(loader):
                x = x.to(trainer.device)
                # isolate the first two digits of the input sequence alone
                d1d2 = x[:, :ndigit*2]
                # let the model sample the rest of the sequence
                d1d2d3 = model.generate(d1d2, ndigit+1, do_sample=False) # using greedy argmax, not sampling
                # isolate the last digit of the sampled sequence
                d3 = d1d2d3[:, -(ndigit+1):]
                d3 = d3.flip(1) # reverse the digits to their "normal" order
                # decode the integers from individual digits
                d1i = (d1d2[:,:ndigit] * factors[:,1:]).sum(1)
                d2i = (d1d2[:,ndigit:ndigit*2] * factors[:,1:]).sum(1)
                d3i_pred = (d3 * factors).sum(1)
                d3i_gt = d1i + d2i # manually calculate the ground truth
                # evaluate the correctness of the results in this batch
                correct = (d3i_pred == d3i_gt).cpu() # Software 1.0 vs. Software 2.0 fight RIGHT on this line haha
                for i in range(x.size(0)):
                    results.append(int(correct[i]))
                    if not correct[i] and mistakes_printed_already < 5: # only print up to 5 mistakes to get a sense
                        mistakes_printed_already += 1
                        print("GPT claims that %d + %d = %d but gt is %d" % (d1i[i], d2i[i], d3i_pred[i], d3i_gt[i]))
                if max_batches is not None and b+1 >= max_batches:
                    break
            rt = torch.tensor(results, dtype=torch.float)
            print("%s final score: %d/%d = %.2f%% correct" % (split, rt.sum(), len(results), 100*rt.mean()))
            if split == 'test':
                accuracies.append(100*rt.mean())
            return rt.sum()
    
        # iteration callback
        top_score = 0
        def batch_end_callback(trainer):
            global top_score
            train_losses.append(trainer.loss.item())
            
            if trainer.iter_num % 10 == 0:
                print(f"iter_dt {trainer.iter_dt * 1000:.2f}ms; iter {trainer.iter_num}: train loss {trainer.loss.item():.5f}")
    
            if trainer.iter_num % 500 == 0:
                # evaluate both the train and test score
                train_max_batches = {1: None, 2: None, 3: 5}[config.data.ndigit] # if ndigit=2 we can afford the whole train set, ow no
                model.eval()
                with torch.no_grad():
                    train_score = eval_split(trainer, 'train', max_batches=train_max_batches)
                    test_score  = eval_split(trainer, 'test',  max_batches=None)
                score = train_score + test_score
                # save the model if this is the best score we've seen so far
                if score > top_score:
                    top_score = score
                    print(f"saving model with new top score of {score}")
                    ckpt_path = os.path.join(config.system.work_dir, "model.pt")
                    torch.save(model.state_dict(), ckpt_path)
                # revert model to training mode
                model.train()
    
        trainer.set_callback('on_batch_end', batch_end_callback)
    
        # run the optimization
        trainer.run()
    
        # show performance
        fig, (ax1, ax2) = plt.subplots(2, 1, figsize=(8, 6))
        x = np.arange(len(train_losses))
        ax1.plot(x, train_losses, label='Train Loss')
        ax2.set_xlabel('iter')
        ax1.set_ylabel('Loss')
        ax1.legend()
        x = np.arange(len(accuracies)) * 500
        ax2.plot(x, accuracies, label='Accuracy')
        ax2.set_xlabel('iter')
        ax2.set_ylabel('Accuracy')
        ax2.legend()
        plt.subplots_adjust(hspace=0.4)
        plt.show()
    
  • 代码中绘图部分是我添加的,train loss 和 accuracy 的变化曲线如下
    在这里插入图片描述
Logo

开放原子开发者工作坊旨在鼓励更多人参与开源活动,与志同道合的开发者们相互交流开发经验、分享开发心得、获取前沿技术趋势。工作坊有多种形式的开发者活动,如meetup、训练营等,主打技术交流,干货满满,真诚地邀请各位开发者共同参与!

更多推荐