当前位置: 首页 > news >正文

wordpress+支付查看山东网站建设优化技术

wordpress+支付查看,山东网站建设优化技术,wordpress 安装 插件,揭阳智能模板建站目录 环境配置与脚本编写 前向传播过程 网络结构 环境配置与脚本编写 按照官网执行并没有顺利完成#xff0c;将yaml文件中的 pip 项 手动安装的 conda create -n artrack python3.9 # 启动该环境#xff0c;并跳转到项目主目录路径下 astor0.8.1 configparser5.2.0 data…目录 环境配置与脚本编写 前向传播过程 网络结构 环境配置与脚本编写 按照官网执行并没有顺利完成将yaml文件中的 pip 项 手动安装的 conda create -n artrack python3.9 # 启动该环境并跳转到项目主目录路径下 astor0.8.1 configparser5.2.0 data0.4 docker-pycreds0.4.0 easydict1.9 einops0.4.1 formulaic0.5.2 funcsigs1.0.2 future0.18.2 gitdb4.0.9 gitpython3.1.27 interface-meta1.3.0 iopath0.1.9 jpeg4py0.1.4 jsonpatch1.32 jsonpointer2.3 latex0.7.0 libarchive-c2.9 linearmodels4.29 lmdb1.3.0 loguru0.6.0 mat730.59 memory-profiler0.60.0 msgpack1.0.2 ninja1.11.1 opencv-python4.5.5.64 pathtools0.1.2 promise2.3 property-cached1.6.4 protobuf3.20.0 pycocotools2.0.4 pyhdfe0.1.2 ruamel-yaml-conda0.15.100 sentry-sdk1.5.8 setproctitle1.2.2 setuptools-scm7.1.0 shapely1.8.1.post1 shortuuid1.0.8 shutilwhich1.1.0 smmap5.0.0 tables3.6.1 tempdir0.7.1 tensorboardx2.5.1 thop0.1.0.post2207010342 tikzplotlib0.10.1 timm0.5.4 tomli2.0.1 torch1.11.0 torchfile0.1.0 visdom0.1.8.9 wandb0.12.11 webcolors1.12 yaspin2.1.0 里面的默认路径需要改写 python tracking/create_default_local_file.py --workspace_dir . --data_dir ./data --save_dir ./outpu 官网下载训练好的模型创建路径将模型放在该路径下 ARTrack-main/output/checkpoints/train/artrack_seq/artrack_seq_256_full/ARTrackSeq_ep0060.pth.tar 创建encoder的预训练模型路径并把预训练模型放入这里在yaml文件中进行更改并且源脚本文件 artrack_seq.py中也需要更改 mkdir pretrained_model # mae_pretrain_vit_base.pth 文件名# artrack_seq_256_full.yaml 中用绝对路径改写 PRETRAIN_PTH: /root/data/zjx/Code-subject/ARTrack/ARTrack-main/pretrained_models# 同时将artrack_seq.py --100 中的 load_from cfg.MODEL.PRETRAIN_PTH # 改为 load_from cfg.MODEL.PRETRAIN_PTH / cfg.MODEL.PRETRAIN_FILE #同时将 artrack_seq.py -- 103 中的 missing_keys, unexpected_keys model.load_state_dict(checkpoint[net], strictFalse) # 改为 missing_keys, unexpected_keys model.load_state_dict(checkpoint[model], strictFalse) 代码中没有实现 run video 的脚本这里需要自定义一个脚本实现 from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literalsimport os import random import argparse import multiprocessingimport cv2 import torch import torch.nn as nn import numpy as np from glob import globfrom lib.test.evaluation.tracker import Trackerimport sysprj_path os.path.join(os.path.dirname(__file__), ..) if prj_path not in sys.path:sys.path.append(prj_path)torch.set_num_threads(1)parser argparse.ArgumentParser(descriptionRun tracker on sequence or dataset.) parser.add_argument(tracker_name, typestr, helpName of tracking method.) parser.add_argument(tracker_param, typestr, helpName of config file.) parser.add_argument(--runid, typeint, defaultNone, helpThe run id.) parser.add_argument(--video_path, typestr, defaultNone, helpName of dataset (otb, nfs, uav, tpl, vot, tn, gott, gotv, lasot).) parser.add_argument(--sequence, typestr, defaultNone, helpSequence number or name.) parser.add_argument(--debug, typeint, default0, helpDebug level.) parser.add_argument(--threads, typeint, default0, helpNumber of threads.) parser.add_argument(--num_gpus, typeint, default8)args parser.parse_args()def main(): # 这里已经是图片了colors [random.randint(0, 255) for _ in range(3)]print([INFO] Loading the model)# load configtrackers Tracker(args.tracker_name, args.tracker_param, None, args.runid)try:worker_name multiprocessing.current_process().nameworker_id int(worker_name[worker_name.find(-) 1:]) - 1gpu_id worker_id % args.num_gpustorch.cuda.set_device(gpu_id)except:passtrackers.run_video(args.video_path, None, None, None, False)if __name____main__:main()执行 python tracking/run_video.py artrack_seq artrack_seq_256_full --video_path /root/data/zjx/Code-subject/OSTrack-main/experiments/video/soccer1.avi 前向传播过程 裁剪模板区域和OSTrack代码一样初始化的时候为需要保留的N帧的bbox的坐标信息创建了一个buffer--self.store_result初始化时全为 init bboxN的值此时设置为7 for i in range(self.save_all - 1):self.store_result.append(info[init_bbox].copy()) 搜索区域的裁剪和OSTrack的一样。将之前帧的坐标进行变换  以前一帧预测的坐标为参考点计算相对坐标因为当前帧的裁剪的搜索区域的就是以上一帧预测的bbox为中心进行裁剪的所以搜索区域的中心实则是前一帧预测的bbox的中心。只不过前一帧预测的bbox为原img的尺度而搜索区域为crop size上的尺度因此只需要将计算原img尺度上的也就是之前帧的预测的坐标与前一帧预测的坐标的相对坐标再乘以resize factor就可以将相对坐标转换到crop size 的尺度下。并且前一帧的预测的bbox转换实则移到了搜索区域的中心点也就是 crop_size/2, crop_ size/2)。  转换后除以 crop size 进行了归一化不过这里有可能会 小于0 或者 大于 1因为坐标变换可能会超出边界。接下来将xywh转换成 xyxy 形式并筛选只保留-0.51.5区间的。然后对坐标进行量化。加上0.5 为了防止 出现负数最终将bbox量化到 2*bins-1之间。最终包含时空上下文信息的坐标输入为 seqs_out seqs_out.unsqueeze(0) # 128 将 模板 和 搜索区域送入 ViT backbone中进行特征提取这个过程中一共 16倍 下采样。然后将 提取的 sequence patch、以及位置编码、外加之前转换后的之前帧的bbox的信息 送入 接下来的Transformer中。 首先进入一个 encoder在FeatureFusionEncoder类中进行一些预处理主要的基本模块是 FeatureFusion 模块。这个encoder的主要过程如下所示最终返回 z 和 x 一样shape的特征 patch。 接下来 将 之前帧的 bbox 坐标序列以及开始标志拼接在一起作为decoder的输入 sequence。因为只需要预测bbox的坐标所以不需要额外的结束标志输出的序列长度直接为4即可。 1、 将输入的sequence 进行词汇嵌入词向量的长度是crop img 下采样得到的特征patch的分辨率 2、 将初始输入tgt、模板特征、搜索特征、patch z的位置编码、 x patch的位置编码、identity高斯截断分布、高斯截断分布、查询嵌入、输入序列的掩码 送入decoder decoder主要由TargetQueryDecoderLayer层组成。该模块的前向过程如下所示一共有6层 最终输出和 tgt shape一样的token sequence。得到的输出的shape为1length768这个length为tgt的长度随sequence的预测而逐渐增加。接下来 1、 拿出得到的 query的 最后一个单词嵌入并与词向量的权重矩阵进行矩阵乘法得到与每个位置量化后的相关联的预测值。 2、 取softmax得到关于量化后的坐标的概率分布。 3、 采用argmax sampleing也就是看最大概率的位置。 4、 将当前预测的量化后的坐标加入到 tgt当中执行循环。 5、 最终得到预测的bbox的量化坐标。得到网络的输出预测后 1、 bbox坐标反量化 2、 xyxy 转为 xywh 中心点加长宽 3、 尺度返回到原img 转成 xywh 左顶点加长宽 4、 平滑处理去掉bbox超出图片的部分 5、 对于之前保存的坐标信息将最靠前的弹出去在最靠后的也就是前一帧的坐标加入当前预测的。好比出栈入栈操作。 网络结构 ARTrackSeq((backbone): VisionTransformer((patch_embed): PatchEmbed((proj): Conv2d(3, 768, kernel_size(16, 16), stride(16, 16))(norm): Identity())(pos_drop): Dropout(p0.0, inplaceFalse)(blocks): Sequential((0): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): Identity()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(1): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(2): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(3): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(4): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(5): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(6): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(7): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(8): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(9): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(10): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse)))(11): Block((norm1): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(attn): Attention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.0, inplaceFalse))(drop_path): DropPath()(norm2): LayerNorm((768,), eps1e-06, elementwise_affineTrue)(mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(drop1): Dropout(p0.0, inplaceFalse)(fc2): Linear(in_features3072, out_features768, biasTrue)(drop2): Dropout(p0.0, inplaceFalse))))(norm): LayerNorm((768,), eps1e-06, elementwise_affineTrue))(pix_head): Pix2Track((word_embeddings): Embedding(802, 768, padding_idx800, max_norm1)(position_embeddings): Embedding(5, 768)(prev_position_embeddings): Embedding(28, 768)(encoder): FeatureFusionEncoder((layers): ModuleList((0): FeatureFusion((z_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_x_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_z_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(x_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(1): FeatureFusion((z_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_x_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_z_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(x_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(2): FeatureFusion((z_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_self_attn): SelfAttention((qkv): Linear(in_features768, out_features2304, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm2_2): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_x_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(x_z_cross_attention): CrossAttention((q): Linear(in_features768, out_features768, biasTrue)(kv): Linear(in_features768, out_features1536, biasTrue)(attn_drop): Dropout(p0.0, inplaceFalse)(proj): Linear(in_features768, out_features768, biasTrue)(proj_drop): Dropout(p0.1, inplaceFalse))(z_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(x_norm3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(z_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(x_mlp): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity()))(z_pos_enc): Untied2DPositionalEncoder((pos): Learned2DPositionalEncoder()(norm): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(pos_q_linear): Linear(in_features768, out_features768, biasTrue)(pos_k_linear): Linear(in_features768, out_features768, biasTrue))(x_pos_enc): Untied2DPositionalEncoder((pos): Learned2DPositionalEncoder()(norm): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(pos_q_linear): Linear(in_features768, out_features768, biasTrue)(pos_k_linear): Linear(in_features768, out_features768, biasTrue))(z_rel_pos_bias_table): RelativePosition2DEncoder()(x_rel_pos_bias_table): RelativePosition2DEncoder()(z_x_rel_pos_bias_table): RelativePosition2DEncoder()(x_z_rel_pos_bias_table): RelativePosition2DEncoder())(decoder): TargetQueryDecoderBlock((layers): ModuleList((0): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(1): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(2): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(3): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(4): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity())(5): TargetQueryDecoderLayer((norm_1): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(self_attn1): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_2_query): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(norm_2_memory): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(multihead_attn): MultiheadAttention((out_proj): NonDynamicallyQuantizableLinear(in_features768, out_features768, biasTrue))(norm_3): LayerNorm((768,), eps1e-05, elementwise_affineTrue)(mlpz): Mlp((fc1): Linear(in_features768, out_features3072, biasTrue)(act): GELU()(fc2): Linear(in_features3072, out_features768, biasTrue)(drop): Dropout(p0.1, inplaceFalse))(drop_path): Identity()))(norm): LayerNorm((768,), eps1e-05, elementwise_affineTrue))) )
http://www.pierceye.com/news/534616/

相关文章:

  • 网站浮动窗口如何做自己怎么做淘宝客网站
  • 石材外贸在哪个网站做网页版 微信
  • 网站开发属于程序员吗sem 优化软件
  • 公司做网站是管理费用小程序官方文档
  • 公司网站推广技巧响水网站设计
  • 徐州本地网站wap页面是什么
  • 网站开发应用价值做套网站多少钱
  • asp.net网站模板免费下载怎么才能访问自己做的网站
  • 长沙企业网站制作宝安公司网站建设
  • 做网站需要拉多大的宽带dw做的网站怎么做后台
  • 公司网站建设设计公司哪家好wordpress自动封ip
  • 郫县网站制作wordpress搜索打钩
  • 哪些网站可以做招商广告语wordpress发文章的id怎么不连续
  • 家私网站栏目和功能需求策划网页样式库
  • 什么是网站网页主页企业电子邮箱格式
  • 金属建材企业网站建设方案用pycharm做网站
  • 重庆网站空间黄骅港一期码头潮汐表
  • 推广网站如何做做酒店网站所用到的算法
  • 最好的网站建设组织wordpress 删除google
  • 生物科技 网站模板下载在线室内设计
  • 网站兼容性问题线上设计师接单
  • 外包网站平台可以做电算化的网站
  • 教育网站设计案例学校网站设计
  • 网站建设入门教程pdf网络推广和seo
  • 闲鱼钓鱼网站怎么做百度网页版主页
  • 一次备案多个网站alexa排名查询
  • 郑州做招商的网站网站建设的流程推广方案
  • wordpress手机网站插件海口seo关键词优化
  • wordpress随机文章佛山网站优化美姿姿seo
  • 做酒类网站中铁三局最新消息