AI大模型微调技术预研:基于Transformer架构的个性化模型训练与部署最佳实践

 
更多

AI大模型微调技术预研:基于Transformer架构的个性化模型训练与部署最佳实践

引言

随着人工智能技术的快速发展,大规模预训练模型(Large Language Models, LLMs)已成为自然语言处理、计算机视觉等领域的核心技术。这些模型通过在海量数据上进行预训练,获得了强大的泛化能力和语言理解能力。然而,在实际应用中,如何将这些通用模型适配到特定任务或领域,成为了业界关注的焦点。

微调技术(Fine-tuning)作为连接通用预训练模型与特定应用场景的桥梁,其重要性不言而喻。本文将深入探讨基于Transformer架构的AI大模型微调技术,分析其核心原理、前沿方法和最佳实践,为企业的AI应用落地提供技术预研参考。

Transformer架构核心原理

自注意力机制

Transformer架构的核心在于自注意力机制(Self-Attention),它允许模型在处理序列数据时关注序列中的不同位置。自注意力的计算公式如下:

Attention(Q, K, V) = softmax(QK^T/√d_k)V

其中,Q(Query)、K(Key)、V(Value)分别代表查询、键和值矩阵,d_k是键向量的维度。

多头注意力

为了增强模型的表达能力,Transformer采用了多头注意力机制:

class MultiHeadAttention(nn.Module):
    def __init__(self, d_model, num_heads):
        super(MultiHeadAttention, self).__init__()
        self.d_model = d_model
        self.num_heads = num_heads
        self.d_k = d_model // num_heads
        
        self.W_q = nn.Linear(d_model, d_model)
        self.W_k = nn.Linear(d_model, d_model)
        self.W_v = nn.Linear(d_model, d_model)
        self.W_o = nn.Linear(d_model, d_model)
        
    def forward(self, query, key, value, mask=None):
        batch_size = query.size(0)
        
        # 线性变换
        Q = self.W_q(query).view(batch_size, -1, self.num_heads, self.d_k).transpose(1, 2)
        K = self.W_k(key).view(batch_size, -1, self.num_heads, self.d_k).transpose(1, 2)
        V = self.W_v(value).view(batch_size, -1, self.num_heads, self.d_k).transpose(1, 2)
        
        # 计算注意力
        scores = torch.matmul(Q, K.transpose(-2, -1)) / math.sqrt(self.d_k)
        if mask is not None:
            scores = scores.masked_fill(mask == 0, -1e9)
        
        attention = F.softmax(scores, dim=-1)
        context = torch.matmul(attention, V)
        
        # 合并多头
        context = context.transpose(1, 2).contiguous().view(batch_size, -1, self.d_model)
        output = self.W_o(context)
        
        return output

位置编码

由于Transformer不包含循环或卷积结构,需要通过位置编码来引入序列信息:

class PositionalEncoding(nn.Module):
    def __init__(self, d_model, max_len=5000):
        super(PositionalEncoding, self).__init__()
        
        pe = torch.zeros(max_len, d_model)
        position = torch.arange(0, max_len, dtype=torch.float).unsqueeze(1)
        div_term = torch.exp(torch.arange(0, d_model, 2).float() * 
                            (-math.log(10000.0) / d_model))
        pe[:, 0::2] = torch.sin(position * div_term)
        pe[:, 1::2] = torch.cos(position * div_term)
        pe = pe.unsqueeze(0).transpose(0, 1)
        
        self.register_buffer('pe', pe)
        
    def forward(self, x):
        return x + self.pe[:x.size(0), :]

参数高效微调方法

适配器微调(Adapter Fine-tuning)

适配器微调通过在预训练模型的每一层中插入小型神经网络模块来实现参数高效的微调:

class AdapterLayer(nn.Module):
    def __init__(self, d_model, bottleneck_size=64):
        super(AdapterLayer, self).__init__()
        self.down_project = nn.Linear(d_model, bottleneck_size)
        self.activation = nn.ReLU()
        self.up_project = nn.Linear(bottleneck_size, d_model)
        self.dropout = nn.Dropout(0.1)
        
    def forward(self, x):
        residual = x
        x = self.down_project(x)
        x = self.activation(x)
        x = self.up_project(x)
        x = self.dropout(x)
        return x + residual

class TransformerWithAdapter(nn.Module):
    def __init__(self, vocab_size, d_model, nhead, num_layers, bottleneck_size=64):
        super(TransformerWithAdapter, self).__init__()
        self.embedding = nn.Embedding(vocab_size, d_model)
        self.pos_encoding = PositionalEncoding(d_model)
        
        self.transformer_layers = nn.ModuleList([
            nn.TransformerEncoderLayer(d_model, nhead) 
            for _ in range(num_layers)
        ])
        
        self.adapters = nn.ModuleList([
            AdapterLayer(d_model, bottleneck_size) 
            for _ in range(num_layers)
        ])
        
        self.output_layer = nn.Linear(d_model, vocab_size)
        
    def forward(self, src, src_mask=None):
        x = self.embedding(src) * math.sqrt(self.embedding.embedding_dim)
        x = self.pos_encoding(x)
        
        for layer, adapter in zip(self.transformer_layers, self.adapters):
            x = layer(x, src_mask)
            x = adapter(x)
            
        return self.output_layer(x)

LoRA(Low-Rank Adaptation)

LoRA通过在权重矩阵中添加低秩分解的更新矩阵来实现参数高效的微调:

class LoRALayer(nn.Module):
    def __init__(self, in_features, out_features, rank=8, alpha=16):
        super(LoRALayer, self).__init__()
        self.in_features = in_features
        self.out_features = out_features
        self.rank = rank
        self.alpha = alpha
        
        # 冻结原始权重
        self.weight = nn.Parameter(torch.Tensor(out_features, in_features))
        self.weight.requires_grad = False
        
        # LoRA参数
        self.lora_A = nn.Parameter(torch.zeros(in_features, rank))
        self.lora_B = nn.Parameter(torch.zeros(rank, out_features))
        self.scaling = alpha / rank
        
        self.reset_parameters()
        
    def reset_parameters(self):
        nn.init.kaiming_uniform_(self.weight, a=math.sqrt(5))
        nn.init.normal_(self.lora_A, std=0.02)
        nn.init.zeros_(self.lora_B)
        
    def forward(self, x):
        return F.linear(x, self.weight + self.lora_B @ self.lora_A * self.scaling)

# 使用示例
class LoraTransformer(nn.Module):
    def __init__(self, vocab_size, d_model, nhead, num_layers, lora_rank=8):
        super(LoraTransformer, self).__init__()
        self.embedding = nn.Embedding(vocab_size, d_model)
        self.pos_encoding = PositionalEncoding(d_model)
        
        encoder_layer = nn.TransformerEncoderLayer(d_model, nhead)
        self.transformer = nn.TransformerEncoder(encoder_layer, num_layers)
        
        # 替换注意力层为LoRA层
        self.replace_attention_with_lora(lora_rank)
        
        self.output_layer = nn.Linear(d_model, vocab_size)
        
    def replace_attention_with_lora(self, rank):
        for layer in self.transformer.layers:
            # 替换QKV投影矩阵
            layer.self_attn.in_proj_weight = nn.Parameter(
                torch.cat([
                    LoRALayer(layer.self_attn.embed_dim, layer.self_attn.embed_dim, rank).weight,
                    LoRALayer(layer.self_attn.embed_dim, layer.self_attn.embed_dim, rank).weight,
                    LoRALayer(layer.self_attn.embed_dim, layer.self_attn.embed_dim, rank).weight
                ], dim=0)
            )
            
    def forward(self, src, src_mask=None):
        x = self.embedding(src) * math.sqrt(self.embedding.embedding_dim)
        x = self.pos_encoding(x)
        x = self.transformer(x, src_mask)
        return self.output_layer(x)

前缀微调(Prefix Tuning)

前缀微调通过学习任务特定的前缀向量来引导模型生成:

class PrefixTuning(nn.Module):
    def __init__(self, d_model, nhead, num_layers, prefix_length=10, dropout=0.1):
        super(PrefixTuning, self).__init__()
        self.prefix_length = prefix_length
        self.d_model = d_model
        self.nhead = nhead
        
        # 前缀参数
        self.prefix_tokens = nn.Parameter(torch.randn(prefix_length, d_model))
        self.prefix_dropout = nn.Dropout(dropout)
        
        # 前缀映射网络
        self.prefix_mlp = nn.Sequential(
            nn.Linear(d_model, d_model * 2),
            nn.Tanh(),
            nn.Linear(d_model * 2, d_model * 2)
        )
        
    def get_prefix(self):
        prefix = self.prefix_tokens.unsqueeze(0)  # (1, prefix_length, d_model)
        prefix = self.prefix_mlp(prefix)
        prefix = self.prefix_dropout(prefix)
        
        # 分离为key和value
        batch_size = prefix.size(0)
        prefix = prefix.view(batch_size, self.prefix_length, 2, self.nhead, self.d_model // self.nhead)
        prefix = prefix.permute(2, 0, 3, 1, 4)  # (2, batch_size, nhead, prefix_length, d_model//nhead)
        
        return prefix[0], prefix[1]  # key, value

class PrefixTransformer(nn.Module):
    def __init__(self, vocab_size, d_model, nhead, num_layers, prefix_length=10):
        super(PrefixTransformer, self).__init__()
        self.embedding = nn.Embedding(vocab_size, d_model)
        self.pos_encoding = PositionalEncoding(d_model)
        self.prefix_tuning = PrefixTuning(d_model, nhead, num_layers, prefix_length)
        
        encoder_layer = nn.TransformerEncoderLayer(d_model, nhead)
        self.transformer = nn.TransformerEncoder(encoder_layer, num_layers)
        self.output_layer = nn.Linear(d_model, vocab_size)
        
    def forward(self, src, src_mask=None):
        batch_size = src.size(0)
        
        # 获取前缀
        prefix_k, prefix_v = self.prefix_tuning.get_prefix()
        prefix_k = prefix_k.expand(-1, batch_size, -1, -1, -1)
        prefix_v = prefix_v.expand(-1, batch_size, -1, -1, -1)
        
        # 嵌入和位置编码
        x = self.embedding(src) * math.sqrt(self.embedding.embedding_dim)
        x = self.pos_encoding(x)
        
        # 修改transformer的前向传播以包含前缀
        for layer in self.transformer.layers:
            x = layer(x, src_mask)
            
        return self.output_layer(x)

模型压缩技术

知识蒸馏

知识蒸馏通过将大型教师模型的知识转移到小型学生模型来实现模型压缩:

class KnowledgeDistillation:
    def __init__(self, teacher_model, student_model, temperature=3.0, alpha=0.7):
        self.teacher_model = teacher_model
        self.student_model = student_model
        self.temperature = temperature
        self.alpha = alpha
        self.criterion = nn.KLDivLoss(reduction='batchmean')
        self.ce_loss = nn.CrossEntropyLoss()
        
    def distillation_loss(self, student_outputs, teacher_outputs, labels):
        # 软标签损失
        soft_loss = self.criterion(
            F.log_softmax(student_outputs / self.temperature, dim=1),
            F.softmax(teacher_outputs / self.temperature, dim=1)
        ) * (self.temperature ** 2)
        
        # 硬标签损失
        hard_loss = self.ce_loss(student_outputs, labels)
        
        return self.alpha * soft_loss + (1 - self.alpha) * hard_loss
    
    def train_step(self, inputs, labels):
        # 教师模型推理(不计算梯度)
        with torch.no_grad():
            teacher_outputs = self.teacher_model(inputs)
        
        # 学生模型训练
        student_outputs = self.student_model(inputs)
        loss = self.distillation_loss(student_outputs, teacher_outputs, labels)
        
        return loss

# 使用示例
def train_with_distillation(teacher_model, student_model, train_loader, optimizer, device):
    distiller = KnowledgeDistillation(teacher_model, student_model)
    
    student_model.train()
    teacher_model.eval()
    
    for batch_idx, (data, target) in enumerate(train_loader):
        data, target = data.to(device), target.to(device)
        
        optimizer.zero_grad()
        loss = distiller.train_step(data, target)
        loss.backward()
        optimizer.step()

剪枝技术

结构化剪枝通过移除不重要的神经元或连接来压缩模型:

class ModelPruner:
    def __init__(self, model, pruning_ratio=0.2):
        self.model = model
        self.pruning_ratio = pruning_ratio
        
    def compute_importance(self, model, dataloader, device):
        """计算参数重要性"""
        importance_scores = {}
        
        # 使用梯度范数作为重要性指标
        for name, param in model.named_parameters():
            if 'weight' in name:
                importance_scores[name] = torch.zeros_like(param)
                
        model.train()
        for data, target in dataloader:
            data, target = data.to(device), target.to(device)
            output = model(data)
            loss = F.cross_entropy(output, target)
            loss.backward()
            
            for name, param in model.named_parameters():
                if 'weight' in name and param.grad is not None:
                    importance_scores[name] += torch.abs(param.grad)
                    
        return importance_scores
    
    def prune_model(self, importance_scores):
        """执行剪枝"""
        for name, module in self.model.named_modules():
            if isinstance(module, nn.Linear):
                weight_name = f"{name}.weight"
                if weight_name in importance_scores:
                    # 获取重要性分数
                    scores = importance_scores[weight_name]
                    
                    # 计算阈值
                    threshold = torch.quantile(scores.flatten(), self.pruning_ratio)
                    
                    # 应用剪枝
                    mask = (scores >= threshold).float()
                    module.weight.data *= mask
                    
                    # 剪枝偏置
                    if module.bias is not None:
                        module.bias.data *= (mask.sum(dim=1) > 0).float()

量化技术

量化通过降低参数精度来减少模型大小和计算开销:

class ModelQuantizer:
    def __init__(self, model):
        self.model = model
        
    def quantize_weights(self, bits=8):
        """权重量化"""
        for name, module in self.model.named_modules():
            if isinstance(module, (nn.Linear, nn.Conv2d)):
                # 计算量化范围
                weight = module.weight.data
                min_val = weight.min()
                max_val = weight.max()
                
                # 量化到指定位数
                qmin = -(2 ** (bits - 1))
                qmax = 2 ** (bits - 1) - 1
                
                scale = (max_val - min_val) / (qmax - qmin)
                zero_point = qmin - min_val / scale
                
                # 执行量化
                quantized_weight = torch.round(weight / scale + zero_point).clamp(qmin, qmax)
                dequantized_weight = (quantized_weight - zero_point) * scale
                
                module.weight.data = dequantized_weight
                
    def quantize_activations(self):
        """激活量化"""
        # 添加量化钩子
        def quantize_hook(module, input, output):
            # 量化输出激活
            min_val = output.min()
            max_val = output.max()
            
            qmin = -128
            qmax = 127
            
            scale = (max_val - min_val) / (qmax - qmin)
            zero_point = qmin - min_val / scale
            
            quantized_output = torch.round(output / scale + zero_point).clamp(qmin, qmax)
            dequantized_output = (quantized_output - zero_point) * scale
            
            return dequantized_output
            
        # 为所有层注册钩子
        hooks = []
        for module in self.model.modules():
            if isinstance(module, (nn.ReLU, nn.Sigmoid, nn.Tanh)):
                hook = module.register_forward_hook(quantize_hook)
                hooks.append(hook)
                
        return hooks

推理优化策略

动态批处理

动态批处理通过合并多个推理请求来提高GPU利用率:

import asyncio
import queue
import threading
from typing import List, Any

class DynamicBatcher:
    def __init__(self, model, max_batch_size=32, max_wait_time=0.01):
        self.model = model
        self.max_batch_size = max_batch_size
        self.max_wait_time = max_wait_time
        self.request_queue = queue.Queue()
        self.result_futures = {}
        
        # 启动批处理线程
        self.batch_thread = threading.Thread(target=self._batch_processing_loop)
        self.batch_thread.daemon = True
        self.batch_thread.start()
        
    def _batch_processing_loop(self):
        while True:
            batch_requests = []
            futures = []
            
            # 收集请求
            start_time = time.time()
            while len(batch_requests) < self.max_batch_size:
                try:
                    request, future = self.request_queue.get(timeout=self.max_wait_time)
                    batch_requests.append(request)
                    futures.append(future)
                except queue.Empty:
                    break
                    
            if batch_requests:
                # 批处理推理
                try:
                    batch_inputs = torch.stack(batch_requests)
                    with torch.no_grad():
                        batch_outputs = self.model(batch_inputs)
                    
                    # 分发结果
                    for i, future in enumerate(futures):
                        future.set_result(batch_outputs[i])
                        
                except Exception as e:
                    for future in futures:
                        future.set_exception(e)
    
    async def inference(self, input_tensor):
        """异步推理接口"""
        loop = asyncio.get_event_loop()
        future = loop.create_future()
        
        self.request_queue.put((input_tensor, future))
        
        result = await future
        return result

# 使用示例
async def main():
    # 初始化模型和批处理器
    model = YourModel()
    batcher = DynamicBatcher(model, max_batch_size=16)
    
    # 并发推理
    tasks = []
    for i in range(100):
        input_tensor = torch.randn(1, 768)  # 假设输入维度
        task = asyncio.create_task(batcher.inference(input_tensor))
        tasks.append(task)
    
    results = await asyncio.gather(*tasks)
    print(f"处理了 {len(results)} 个请求")

模型并行化

对于超大模型,可以采用模型并行化策略:

import torch.distributed as dist
from torch.nn.parallel import DistributedDataParallel as DDP

class ModelParallelTransformer(nn.Module):
    def __init__(self, vocab_size, d_model, nhead, num_layers, num_gpus):
        super(ModelParallelTransformer, self).__init__()
        self.num_gpus = num_gpus
        self.d_model = d_model
        
        # 将层分配到不同GPU
        layers_per_gpu = num_layers // num_gpus
        self.layer_groups = nn.ModuleList()
        
        for gpu_id in range(num_gpus):
            start_layer = gpu_id * layers_per_gpu
            end_layer = start_layer + layers_per_gpu
            
            layers = nn.ModuleList([
                nn.TransformerEncoderLayer(d_model, nhead)
                for _ in range(start_layer, end_layer)
            ])
            
            # 将层移动到对应GPU
            layers = layers.to(f'cuda:{gpu_id}')
            self.layer_groups.append(layers)
            
        self.embedding = nn.Embedding(vocab_size, d_model).to('cuda:0')
        self.pos_encoding = PositionalEncoding(d_model).to('cuda:0')
        self.output_layer = nn.Linear(d_model, vocab_size).to(f'cuda:{num_gpus-1}')
        
    def forward(self, src, src_mask=None):
        # 在第一个GPU上进行嵌入
        x = self.embedding(src) * math.sqrt(self.embedding.embedding_dim)
        x = self.pos_encoding(x)
        x = x.to(f'cuda:0')
        
        # 在不同GPU间传递数据
        for gpu_id, layers in enumerate(self.layer_groups):
            x = x.to(f'cuda:{gpu_id}')
            for layer in layers:
                x = layer(x, src_mask)
                
        # 在最后一个GPU上进行输出
        x = x.to(f'cuda:{self.num_gpus-1}')
        return self.output_layer(x)

# 分布式训练设置
def setup_distributed():
    dist.init_process_group(backend='nccl')
    torch.cuda.set_device(dist.get_rank())
    
def train_model_parallel():
    # 初始化分布式环境
    setup_distributed()
    
    # 创建模型
    model = ModelParallelTransformer(
        vocab_size=30000,
        d_model=1024,
        nhead=16,
        num_layers=24,
        num_gpus=dist.get_world_size()
    )
    
    # 使用DDP包装模型
    model = DDP(model, device_ids=[dist.get_rank()])
    
    # 训练循环
    optimizer = torch.optim.Adam(model.parameters(), lr=1e-4)
    
    for epoch in range(10):
        for batch in dataloader:
            optimizer.zero_grad()
            output = model(batch)
            loss = compute_loss(output, batch.target)
            loss.backward()
            optimizer.step()

缓存优化

利用KV缓存来加速自回归生成:

class KVCache:
    def __init__(self, max_length, num_heads, head_dim, batch_size=1):
        self.max_length = max_length
        self.num_heads = num_heads
        self.head_dim = head_dim
        self.batch_size = batch_size
        self.current_length = 0
        
        # 预分配缓存空间
        self.key_cache = torch.zeros(
            batch_size, num_heads, max_length, head_dim
        )
        self.value_cache = torch.zeros(
            batch_size, num_heads, max_length, head_dim
        )
        
    def update(self, new_keys, new_values):
        """更新缓存"""
        batch_size, num_heads, seq_len, head_dim = new_keys.shape
        
        # 将新的KV存储到缓存中
        self.key_cache[:, :, self.current_length:self.current_length+seq_len] = new_keys
        self.value_cache[:, :, self.current_length:self.current_length+seq_len] = new_values
        
        self.current_length += seq_len
        
    def get_cache(self):
        """获取当前缓存"""
        return (
            self.key_cache[:, :, :self.current_length],
            self.value_cache[:, :, :self.current_length]
        )

class CachedTransformer(nn.Module):
    def __init__(self, vocab_size, d_model, nhead, num_layers):
        super(CachedTransformer, self).__init__()
        self.embedding = nn.Embedding(vocab_size, d_model)
        self.pos_encoding = PositionalEncoding(d_model)
        self.transformer = nn.Transformer(d_model, nhead, num_layers)
        self.output_layer = nn.Linear(d_model, vocab_size)
        
    def forward_with_cache(self, src, cache=None):
        """带缓存的前向传播"""
        x = self.embedding(src)
        x = self.pos_encoding(x)
        
        # 使用缓存进行注意力计算
        if cache is not None:
            key_cache, value_cache = cache.get_cache()
            # 在注意力计算中使用缓存的KV
            # 这里需要修改注意力实现以支持缓存
            
        return self.transformer(x)
    
    def generate(self, input_ids, max_length=100):
        """使用缓存进行文本生成"""
        cache = KVCache(max_length, self.transformer.nhead, self.transformer.d_model // self.transformer.nhead)
        
        # 编码输入
        encoded = self.forward_with_cache(input_ids, cache)
        
        # 逐个生成token
        for _ in range(max_length - input_ids.size(1)):
            # 使用缓存进行解码
            next_token_logits = self.output_layer(encoded[:, -1:])
            next_token = torch.argmax(next_token_logits, dim=-1)
            
            # 更新输入和缓存
            input_ids = torch.cat([input_ids, next_token], dim=1)
            encoded = self.forward_with_cache(next_token, cache)
            
        return input_ids

部署最佳实践

容器化部署

使用Docker进行模型服务化部署:

# Dockerfile
FROM nvidia/cuda:11.8-devel-ubuntu20.04

# 安装基础依赖
RUN apt-get update && apt-get install -y \
    python3.8 \
    python3.8-pip \
    && rm -rf /var/lib/apt/lists/*

# 设置工作目录
WORKDIR /app

# 复制依赖文件
COPY requirements.txt .
RUN pip3 install -r requirements.txt

# 复制模型文件和代码
COPY models/ ./models/
COPY src/ ./src/

# 暴露端口
EXPOSE 8000

# 启动服务
CMD ["python3", "src/server.py"]
# server.py
from flask import Flask, request, jsonify
import torch
from transformers import AutoTokenizer, AutoModel

app = Flask(__name__)

# 加载模型
model_path = "models/your_model"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModel.from_pretrained(model_path)
model.eval()

@app.route('/predict', methods=['POST'])
def predict():
    try:
        data = request.json
        text = data['text']
        
        # 预处理
        inputs = tokenizer(text, return_tensors='pt', padding=True, truncation=True)
        
        # 推理
        with torch.no_grad():
            outputs = model(**inputs)
            
        # 后处理
        result = {
            'embeddings': outputs.last_hidden_state.tolist(),
            'pooled_output': outputs.pooler_output.tolist()
        }
        
        return jsonify(result)
        
    except Exception as e:
        return jsonify({'error': str(e)}), 500

if __name__ == '__main__':
    app.run(host='0.0.0.0', port=8000)

性能监控

实现模型服务的性能监控:

import time
import logging
from functools import wraps
from prometheus_client import Counter, Histogram, start_http_server

# 定义监控指标
REQUEST_COUNT = Counter('model_requests_total', 'Total number of requests')
REQUEST_DURATION = Histogram('model_request_duration_seconds', 'Request duration in seconds')
ERROR_COUNT = Counter('model_errors_total', 'Total number of errors')

def monitor_performance(func

打赏

本文固定链接: https://www.cxy163.net/archives/7575 | 绝缘体

该日志由 绝缘体.. 于 2021年05月16日 发表在 未分类 分类下, 你可以发表评论,并在保留原文地址及作者的情况下引用到你的网站或博客。
原创文章转载请注明: AI大模型微调技术预研:基于Transformer架构的个性化模型训练与部署最佳实践 | 绝缘体
关键字: , , , ,

AI大模型微调技术预研:基于Transformer架构的个性化模型训练与部署最佳实践:等您坐沙发呢!

发表评论


快捷键:Ctrl+Enter