积分充值
 首页
前端开发
AngularDartElectronFlutterHTML/CSSJavaScriptReactSvelteTypeScriptVue.js构建工具
后端开发
.NetC#C++C语言DenoffmpegGoIdrisJavaJuliaKotlinLeanMakefilenimNode.jsPascalPHPPythonRISC-VRubyRustSwiftUML其它语言区块链开发测试微服务敏捷开发架构设计汇编语言
数据库
Apache DorisApache HBaseCassandraClickHouseFirebirdGreenplumMongoDBMySQLPieCloudDBPostgreSQLRedisSQLSQLiteTiDBVitess数据库中间件数据库工具数据库设计
系统运维
AndroidDevOpshttpdJenkinsLinuxPrometheusTraefikZabbix存储网络与安全
云计算&大数据
Apache APISIXApache FlinkApache KarafApache KyuubiApache OzonedaprDockerHadoopHarborIstioKubernetesOpenShiftPandasrancherRocketMQServerlessService MeshVirtualBoxVMWare云原生CNCF机器学习边缘计算
综合其他
BlenderGIMPKiCadKritaWeblate产品与服务人工智能亿图数据可视化版本控制笔试面试
文库资料
前端
AngularAnt DesignBabelBootstrapChart.jsCSS3EchartsElectronHighchartsHTML/CSSHTML5JavaScriptJerryScriptJestReactSassTypeScriptVue前端工具小程序
后端
.NETApacheC/C++C#CMakeCrystalDartDenoDjangoDubboErlangFastifyFlaskGinGoGoFrameGuzzleIrisJavaJuliaLispLLVMLuaMatplotlibMicronautnimNode.jsPerlPHPPythonQtRPCRubyRustR语言ScalaShellVlangwasmYewZephirZig算法
移动端
AndroidAPP工具FlutterFramework7HarmonyHippyIoniciOSkotlinNativeObject-CPWAReactSwiftuni-appWeex
数据库
ApacheArangoDBCassandraClickHouseCouchDBCrateDBDB2DocumentDBDorisDragonflyDBEdgeDBetcdFirebirdGaussDBGraphGreenPlumHStreamDBHugeGraphimmudbIndexedDBInfluxDBIoTDBKey-ValueKitDBLevelDBM3DBMatrixOneMilvusMongoDBMySQLNavicatNebulaNewSQLNoSQLOceanBaseOpenTSDBOracleOrientDBPostgreSQLPrestoDBQuestDBRedisRocksDBSequoiaDBServerSkytableSQLSQLiteTiDBTiKVTimescaleDBYugabyteDB关系型数据库数据库数据库ORM数据库中间件数据库工具时序数据库
云计算&大数据
ActiveMQAerakiAgentAlluxioAntreaApacheApache APISIXAPISIXBFEBitBookKeeperChaosChoerodonCiliumCloudStackConsulDaprDataEaseDC/OSDockerDrillDruidElasticJobElasticSearchEnvoyErdaFlinkFluentGrafanaHadoopHarborHelmHudiInLongKafkaKnativeKongKubeCubeKubeEdgeKubeflowKubeOperatorKubernetesKubeSphereKubeVelaKumaKylinLibcloudLinkerdLonghornMeiliSearchMeshNacosNATSOKDOpenOpenEBSOpenKruiseOpenPitrixOpenSearchOpenStackOpenTracingOzonePaddlePaddlePolicyPulsarPyTorchRainbondRancherRediSearchScikit-learnServerlessShardingSphereShenYuSparkStormSupersetXuperChainZadig云原生CNCF人工智能区块链数据挖掘机器学习深度学习算法工程边缘计算
UI&美工&设计
BlenderKritaSketchUI设计
网络&系统&运维
AnsibleApacheAWKCeleryCephCI/CDCurveDevOpsGoCDHAProxyIstioJenkinsJumpServerLinuxMacNginxOpenRestyPrometheusServertraefikTrafficUnixWindowsZabbixZipkin安全防护系统内核网络运维监控
综合其它
文章资讯
 上传文档  发布文章  登录账户
IT文库
  • 综合
  • 文档
  • 文章

无数据

分类

全部云计算&大数据(21)机器学习(21)

语言

全部英语(13)中文(简体)(8)

格式

全部PDF文档 PDF(21)
 
本次搜索耗时 0.077 秒,为您找到相关结果约 21 个.
  • 全部
  • 云计算&大数据
  • 机器学习
  • 全部
  • 英语
  • 中文(简体)
  • 全部
  • PDF文档 PDF
  • 默认排序
  • 最新排序
  • 页数排序
  • 大小排序
  • 全部时间
  • 最近一天
  • 最近一周
  • 最近一个月
  • 最近三个月
  • 最近半年
  • 最近一年
  • pdf文档 PyTorch Release Notes

    mixed-precision arithmetic and Tensor Cores on V100 GPUs for faster training times while maintaining target accuracy. This model script is available on GitHub and NGC. ‣ Mask R-CNN model: Mask R-CNN is a arithmetic by using Tensor Cores on NVIDIA V100 GPUs for 1.3x faster training time while maintaining target accuracy. This model script is available on GitHub and NGC. ‣ Tacotron 2 and WaveGlow v1.1 model: mixed-precision arithmetic and Tensor Cores on V100 GPUs for faster training times while maintaining target accuracy. This model script is available on GitHub and NGC. ‣ Mask R-CNN model: Mask R-CNN is a
    0 码力 | 365 页 | 2.94 MB | 1 年前
    3
  • pdf文档 AI大模型千问 qwen 中文文档

    Thread generation_kwargs = dict(model_inputs, streamer=streamer, max_new_tokens=512) thread = Thread(target=model.generate, kwargs=generation_kwargs) thread.start() generated_text = "" for new_text in streamer: @dataclass class LoraArguments: lora_r: int = 64 lora_alpha: int = 16 lora_dropout: float = 0.05 lora_target_modules: List[str] = field( default_factory=lambda: [ "q_proj", "k_proj", "v_proj", "o_proj", "up_proj" lora_alpha: the alpha value for LoRA; • lora_dropout: the dropout rate for LoRA; • lora_target_modules: the target modules for LoRA. By default we tune all linear layers; • lora_weight_path: the path to
    0 码力 | 56 页 | 835.78 KB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 7 - Automation

    and creating a random classification dataset with 20 samples, each one assigned to one of the five target classes. import random import tensorflow as tf import numpy as np from tensorflow.keras import layers read_config=tfds.ReadConfig(try_autocache=False) ) Let's resize the dataset splits to the same size. The target size is identical to the project in chapter 3. # Dataset image size IMG_SIZE = 264 def resize_image(image model and the child networks are the players whose rewards are determined by their performance on the target dataset. The controller model learns to generate better architectures as the search game progresses
    0 码力 | 33 页 | 2.48 MB | 1 年前
    3
  • pdf文档 全连接神经网络实战. pytorch 版

    datasets 是 torchvision 的对象,它返回的数据就是 pytorch 的 Dataset 类型的。 参数 transf orm 表示导出的数据应该怎么转换,我们还可以使用参数 target_transf orm 表 示导出的数据标签应该怎么转换。 注意显示时我们调用了 squeeze() 函数,这是因为原来的数据维度是 (1,28,28) 的三维数据, 使用.squeeze() ” , train=True , #用 来 训 练 的 数 据 download=True , #如 果 根 目 录 没 有 就 下 载 transform=ToTensor () , target_transform=Lambda( lambda y : torch . zeros (10 , dtype=torch . f l o a t ) . scatter_ (0 , torch 是对数据的转换,ToTensor() 函数将 PIL 图像或者 NumPy 的 ndarray 转换为 FloatTensor 类型的,并且把图像的每个像素值压缩到 [0.0,1.0] 之间。 target_transf orm 10 1.2. 导入样本数据 是标签的转换,分类中我们需要将标签表示为向量的形式,例如一共有三类,则表示为: [1 0 0] (1.2.1) [0 1 0] (1
    0 码力 | 29 页 | 1.40 MB | 1 年前
    3
  • pdf文档 pytorch 入门笔记-03- 神经网络

    (output, target) 作为输入,计算一个值来估计网络的输出和目标值相差多少。 译者注:output 为网络的输出,target 为实际值 nn 包中有很多不同的损失函数。 nn.MSELoss是一个比较简单的损失函数,它计算输出和目标间的均方误差, 例如: output = net(input) target = torch.rand(10) target = target.view(1 view(1, -1) criterion = nn.MSELoss() loss = criterion(output, target) print(loss) tensor(0.4526, grad_fn=) 现在,如果在反向过程中跟随 loss , 使用它的 .grad_fn 属性,将看到如下所示的计算图。 input -> conv2d -> relu -> parameters(), lr=0.01) # 迭代训练 optimizer.zero_grad() # 梯度清零 output = net(input) loss = criterion(output, target) # 计算损失 loss.backward() # 反向传播 optimizer.step() # 更新参数 注意 观察如何使用 optimizer.zero_grad() 手动将梯度缓冲区设置为零。
    0 码力 | 7 页 | 370.53 KB | 1 年前
    3
  • pdf文档 Lecture 1: Overview

    Choose exactly what is to be learned, i.e. the target function Choose how to represent the target function Choose a learning algorithm to infer the target function from the ex- perience. Feng Li (SDU) useful target function. Checker boards labeled with the correct move, e.g. extracted from record of expert play Indirect experience: Given feedback which is not direct I/O pairs for a useful target function have some training cases for which its value is known The thing we want to predict is called the target or the response variable Usually, we need training data Feng Li (SDU) Overview September 6, 2023
    0 码力 | 57 页 | 2.41 MB | 1 年前
    3
  • pdf文档 Keras: 基于 Python 的深度学习库

    compile compile(self, optimizer, loss, metrics=None, sample_weight_mode=None, weighted_metrics=None, target_tensors=None) 用于配置训练模型。 参数 • optimizer: 字符串(优化器名)或者优化器对象。详见 optimizers。 • loss: 字符串(目标函数名)或目标函数。详见 sample_weight 或 class_weight 评估和加权的度 量标准列表。 • target_tensors: 默认情况下,Keras 将为模型的目标创建一个占位符,在训练过程中将使用 目标数据。相反,如果你想使用自己的目标张量(反过来说,Keras 在训练期间不会载入 这些目标张量的外部 Numpy 数据),您可以通过 target_tensors 参数指定它们。它应该 是单个张量(对于单输出 Sequential optimizer, loss, metrics=None, loss_weights=None, sample_weight_mode=None, weighted_metrics=None, target_tensors=None) 用于配置训练模型。 参数 • optimizer: 字符串(优化器名)或者优化器对象。详见 optimizers。 • loss: 字符串(目标函数名)或目标函数。详见
    0 码力 | 257 页 | 1.19 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 3 - Learning Techniques

    of handwritten digit that can potentially confuse the human labelers to choose a 1 or a 7 as the target label. Obtaining labels in many cases requires significant human involvement, and for that reason In our example, only the 300 KB vanilla model is acceptable for deployment (it meets 80% accuracy target). Whereas, among the models with the learning techniques, four models with the smallest being the a cat! The label mixing transformations generate samples based on differently labeled inputs. The target label is a composite of the inputs that were combined. A combination of a dog with a hamster image
    0 码力 | 56 页 | 18.93 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architectures

    demonstrates the Skipgram task. Figure 4-5: This figure depicts the sliding window of size 5, the hidden target word, model inputs, and the label for a given sample text in the CBOW task. 7 GloVe - https://nlp arXiv:1301.3781 (2013). Figure 4-6: This figure depicts the sliding window of size 5, the hidden target word, model inputs, and the label for a given sample text in the Skipgram task. Let’s get to solving self-attention. Encoder-decoder attention computes attention between the encoder output sequence and the target sequence. Self-attention is a special type of attention which operates over a single sequence to
    0 码力 | 53 页 | 3.92 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 6 - Advanced Learning Techniques - Technical Review

    domain that transfer well across specific tasks in that domain. They can be adapted to solve the target task by: 1. Adding a new prediction head to the pre-trained model which can translate the general labeled examples otherwise). If we add a middle-step of pre-training using unlabeled data from the same target dataset, the authors report needing fewer labeled examples. Refer to figure 6-6 for a comparison data, ULMFit semi-supervised: pre-training with WikiText-103 as well as unlabeled data from the target dataset and fine-tuning with labeled data). Source: Howard et al. The pre-trained model can then
    0 码力 | 31 页 | 4.03 MB | 1 年前
    3
共 21 条
  • 1
  • 2
  • 3
前往
页
相关搜索词
PyTorchReleaseNotesAI模型千问qwen中文文档EfficientDeepLearningBookEDLChapterAutomation连接神经网络神经网神经网络实战pytorch入门笔记03LectureOverviewKeras基于Python深度学习TechniquesArchitecturesAdvancedTechnicalReview
IT文库
关于我们 文库协议 联系我们 意见反馈 免责声明
本站文档数据由用户上传或本站整理自互联网,不以营利为目的,供所有人免费下载和学习使用。如侵犯您的权益,请联系我们进行删除。
IT文库 ©1024 - 2025 | 站点地图
Powered By MOREDOC AI v3.3.0-beta.70
  • 关注我们的公众号【刻舟求荐】,给您不一样的精彩
    关注我们的公众号【刻舟求荐】,给您不一样的精彩