当前位置: 首页 > news >正文

企业网站最下面的那栏叫啥俄文网站推广

企业网站最下面的那栏叫啥,俄文网站推广,wordpress主题 产品展示,佛山快速建站哪家服务专业目录 环境配置与脚本编写 前向传播过程 网络结构 环境配置与脚本编写 按照官网执行并没有顺利完成#xff0c;将yaml文件中的 pip 项 手动安装的 conda create -n artrack python3.9 # 启动该环境#xff0c;并跳转到项目主目录路径下 astor0.8.1 configparser5.2.0 data…目录 环境配置与脚本编写 前向传播过程 网络结构 环境配置与脚本编写 按照官网执行并没有顺利完成将yaml文件中的 pip 项 手动安装的 conda create -n artrack python3.9 # 启动该环境并跳转到项目主目录路径下 astor0.8.1 configparser5.2.0 data0.4 docker-pycreds0.4.0 easydict1.9 einops0.4.1 formulaic0.5.2 funcsigs1.0.2 future0.18.2 gitdb4.0.9 gitpython3.1.27 interface-meta1.3.0 iopath0.1.9 jpeg4py0.1.4 jsonpatch1.32 jsonpointer2.3 latex0.7.0 libarchive-c2.9 linearmodels4.29 lmdb1.3.0 loguru0.6.0 mat730.59 memory-profiler0.60.0 msgpack1.0.2 ninja1.11.1 opencv-python4.5.5.64 pathtools0.1.2 promise2.3 property-cached1.6.4 protobuf3.20.0 pycocotools2.0.4 pyhdfe0.1.2 ruamel-yaml-conda0.15.100 sentry-sdk1.5.8 setproctitle1.2.2 setuptools-scm7.1.0 shapely1.8.1.post1 shortuuid1.0.8 shutilwhich1.1.0 smmap5.0.0 tables3.6.1 tempdir0.7.1 tensorboardx2.5.1 thop0.1.0.post2207010342 tikzplotlib0.10.1 timm0.5.4 tomli2.0.1 torch1.11.0 torchfile0.1.0 visdom0.1.8.9 wandb0.12.11 webcolors1.12 yaspin2.1.0 里面的默认路径需要改写 python tracking/create_default_local_file.py --workspace_dir . --data_dir ./data --save_dir ./outpu 官网下载训练好的模型创建路径将模型放在该路径下 ARTrack-main/output/checkpoints/train/artrack_seq/artrack_seq_256_full/ARTrackSeq_ep0060.pth.tar 创建encoder的预训练模型路径并把预训练模型放入这里在yaml文件中进行更改并且源脚本文件 artrack_seq.py中也需要更改 mkdir pretrained_model # mae_pretrain_vit_base.pth 文件名# artrack_seq_256_full.yaml 中用绝对路径改写 PRETRAIN_PTH: /root/data/zjx/Code-subject/ARTrack/ARTrack-main/pretrained_models# 同时将artrack_seq.py --100 中的 load_from cfg.MODEL.PRETRAIN_PTH # 改为 load_from cfg.MODEL.PRETRAIN_PTH / cfg.MODEL.PRETRAIN_FILE #同时将 artrack_seq.py -- 103 中的 missing_keys, unexpected_keys model.load_state_dict(checkpoint[net], strictFalse) # 改为 missing_keys, unexpected_keys model.load_state_dict(checkpoint[model], strictFalse) 代码中没有实现 run video 的脚本这里需要自定义一个脚本实现 from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literalsimport os import random import argparse import multiprocessingimport cv2 import torch import torch.nn as nn import numpy as np from glob import globfrom lib.test.evaluation.tracker import Trackerimport sysprj_path os.path.join(os.path.dirname(__file__), ..) if prj_path not in sys.path:sys.path.append(prj_path)torch.set_num_threads(1)parser argparse.ArgumentParser(descriptionRun tracker on sequence or dataset.) parser.add_argument(tracker_name, typestr, helpName of tracking method.) parser.add_argument(tracker_param, typestr, helpName of config file.) parser.add_argument(--runid, typeint, defaultNone, helpThe run id.) parser.add_argument(--video_path, typestr, defaultNone, helpName of dataset (otb, nfs, uav, tpl, vot, tn, gott, gotv, lasot).) parser.add_argument(--sequence, typestr, defaultNone, helpSequence number or name.) parser.add_argument(--debug, typeint, default0, helpDebug level.) parser.add_argument(--threads, typeint, default0, helpNumber of threads.) parser.add_argument(--num_gpus, typeint, default8)args parser.parse_args()def main(): # 这里已经是图片了colors [random.randint(0, 255) for _ in range(3)]print([INFO] Loading the model)# load configtrackers Tracker(args.tracker_name, args.tracker_param, None, args.runid)try:worker_name multiprocessing.current_process().nameworker_id int(worker_name[worker_name.find(-) 1:]) - 1gpu_id worker_id % args.num_gpustorch.cuda.set_device(gpu_id)except:passtrackers.run_video(args.video_path, None, None, None, False)if __name____main__:main()执行 python tracking/run_video.py artrack_seq artrack_seq_256_full --video_path /root/data/zjx/Code-subject/OSTrack-main/experiments/video/soccer1.avi 前向传播过程 裁剪模板区域和OSTrack代码一样初始化的时候为需要保留的N帧的bbox的坐标信息创建了一个buffer--self.store_result初始化时全为 init bboxN的值此时设置为7 for i in range(self.save_all - 1):self.store_result.append(info[init_bbox].copy()) 搜索区域的裁剪和OSTrack的一样。将之前帧的坐标进行变换  以前一帧预测的坐标为参考点计算相对坐标因为当前帧的裁剪的搜索区域的就是以上一帧预测的bbox为中心进行裁剪的所以搜索区域的中心实则是前一帧预测的bbox的中心。只不过前一帧预测的bbox为原img的尺度而搜索区域为crop size上的尺度因此只需要将计算原img尺度上的也就是之前帧的预测的坐标与前一帧预测的坐标的相对坐标再乘以resize factor就可以将相对坐标转换到crop size 的尺度下。并且前一帧的预测的bbox转换实则移到了搜索区域的中心点也就是 crop_size/2, crop_ size/2)。  转换后除以 crop size 进行了归一化不过这里有可能会 小于0 或者 大于 1因为坐标变换可能会超出边界。接下来将xywh转换成 xyxy 形式并筛选只保留-0.51.5区间的。然后对坐标进行量化。加上0.5 为了防止 出现负数最终将bbox量化到 2*bins-1之间。最终包含时空上下文信息的坐标输入为 seqs_out seqs_out.unsqueeze(0) # 128 将 模板 和 搜索区域送入 ViT backbone中进行特征提取这个过程中一共 16倍 下采样。然后将 提取的 sequence patch、以及位置编码、外加之前转换后的之前帧的bbox的信息 送入 接下来的Transformer中。 首先进入一个 encoder在FeatureFusionEncoder类中进行一些预处理主要的基本模块是 FeatureFusion 模块。这个encoder的主要过程如下所示最终返回 z 和 x 一样shape的特征 patch。 接下来 将 之前帧的 bbox 坐标序列以及开始标志拼接在一起作为decoder的输入 sequence。因为只需要预测bbox的坐标所以不需要额外的结束标志输出的序列长度直接为4即可。 1、 将输入的sequence 进行词汇嵌入词向量的长度是crop img 下采样得到的特征patch的分辨率 2、 将初始输入tgt、模板特征、搜索特征、patch z的位置编码、 x patch的位置编码、identity高斯截断分布、高斯截断分布、查询嵌入、输入序列的掩码 送入decoder decoder主要由TargetQueryDecoderLayer层组成。该模块的前向过程如下所示一共有6层 最终输出和 tgt shape一样的token sequence。得到的输出的shape为1length768这个length为tgt的长度随sequence的预测而逐渐增加。接下来 1、 拿出得到的 query的 最后一个单词嵌入并与词向量的权重矩阵进行矩阵乘法得到与每个位置量化后的相关联的预测值。 2、 取softmax得到关于量化后的坐标的概率分布。 3、 采用argmax sampleing也就是看最大概率的位置。 4、 将当前预测的量化后的坐标加入到 tgt当中执行循环。 5、 最终得到预测的bbox的量化坐标。得到网络的输出预测后 1、 bbox坐标反量化 2、 xyxy 转为 xywh 中心点加长宽 3、 尺度返回到原img 转成 xywh 左顶点加长宽 4、 平滑处理去掉bbox超出图片的部分 5、 对于之前保存的坐标信息将最靠前的弹出去在最靠后的也就是前一帧的坐标加入当前预测的。好比出栈入栈操作。 网络结构 ARTrackSeq((backbone): VisionTransformer((patch_embed): PatchEmbed((proj): Conv2d(3, 768, kernel_size(16, 16), stride(16, 16))(norm): Identity())(pos_drop): Dropout(p0.0, inplaceFalse)(blocks): Sequential((0): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): Identity()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(1): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(2): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(3): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(4): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(5): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(6): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(7): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(8): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(9): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(10): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(11): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse))))(norm): LayerNorm((768,), eps1e-06, elementwise_affineTrue))(pix_head): Pix2Track((word_embeddings): Embedding(802, 768, padding_idx800, max_norm1)(position_embeddings): Embedding(5, 768)(prev_position_embeddings): Embedding(28, 768)(encoder): FeatureFusionEncoder((layers): ModuleList((0): FeatureFusion((z_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_x_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_z_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(x_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(1): FeatureFusion((z_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_x_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_z_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(x_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(2): FeatureFusion((z_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_x_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_z_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(x_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity()))(z_pos_enc): Untied2DPositionalEncoder((pos): Learned2DPositionalEncoder()(norm): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(pos_q_linear): Linear(in_features768, out_features768, biasTrue)(pos_k_linear): Linear(in_features768, out_features768, biasTrue))(x_pos_enc): Untied2DPositionalEncoder((pos): Learned2DPositionalEncoder()(norm): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(pos_q_linear): Linear(in_features768, out_features768, biasTrue)(pos_k_linear): Linear(in_features768, out_features768, biasTrue))(z_rel_pos_bias_table): RelativePosition2DEncoder()(x_rel_pos_bias_table): RelativePosition2DEncoder()(z_x_rel_pos_bias_table): RelativePosition2DEncoder()(x_z_rel_pos_bias_table): RelativePosition2DEncoder())(decoder): TargetQueryDecoderBlock((layers): ModuleList((0): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(1): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(2): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(3): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(4): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(5): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity()))(norm): LayerNorm((768,), eps1e-05, elementwise_affineTrue))) )
http://www.pierceye.com/news/620150/

相关文章:

  • 做视频网站视频空间设计是什么
  • 怎么把网站推广wordpress 百万数据
  • 乐陵市人力资源中心网站网站的内容和功能
  • wordpress网站搬家图片路径做网站的算什么行业
  • 个人网站logo重庆网络优化平台
  • 河南 网站建设静态网站是什么意思
  • 上海正规做网站公司电话基于flash网站设计
  • 每个城市建设规划在哪个网站wordpress 无法安装主题
  • 自建网站做外贸谷歌推广网站内部资源推广方法
  • 网站数据库5g一个人看的免费视频高清直播
  • 怎么做网站注册推广泰州网站建设费用
  • 找南阳建立网站的公司网址大全最新版的
  • 网站建设与维护就业前景小程序开发外包注意事项
  • 胶州网站建设哪里有天润网站建设
  • 网站备案 怎么加搜索引擎网站建设
  • 做外贸自己开公司网站网站建设三个友好
  • 深圳高端品牌网站设计网站建设实训报告收获
  • 万虹点读机如何做系统下载网站网站seo的重要性
  • 同一家公司可以做几个网站吗铁岭网站开发公司
  • 网站推广费用大概需要多少钱个人二级网站怎么做
  • 查询企业的网站有哪些山东平台网站建设找哪家
  • 如何推广外贸型网站wordpress本地环境迁移
  • 网站建设国内外现状网站建设公司 网络服务
  • 百度网站首页福田时代汽车官方网站
  • 网站建设智推网深圳进出口贸易有限公司
  • 网站开发语言pwordpress v4.9.5
  • 东莞建站模板源码东莞注塑切水口东莞网站建设
  • 做文案策划需要看什么网站服装网站开发目的
  • 湖北定制型网站建设微信公众平台网页版
  • 需要做网站的公司有哪些免费网页模板之家