积分充值
 首页
前端开发
AngularDartElectronFlutterHTML/CSSJavaScriptReactSvelteTypeScriptVue.js构建工具
后端开发
.NetC#C++C语言DenoffmpegGoIdrisJavaJuliaKotlinLeanMakefilenimNode.jsPascalPHPPythonRISC-VRubyRustSwiftUML其它语言区块链开发测试微服务敏捷开发架构设计汇编语言
数据库
Apache DorisApache HBaseCassandraClickHouseFirebirdGreenplumMongoDBMySQLPieCloudDBPostgreSQLRedisSQLSQLiteTiDBVitess数据库中间件数据库工具数据库设计
系统运维
AndroidDevOpshttpdJenkinsLinuxPrometheusTraefikZabbix存储网络与安全
云计算&大数据
Apache APISIXApache FlinkApache KarafApache KyuubiApache OzonedaprDockerHadoopHarborIstioKubernetesOpenShiftPandasrancherRocketMQServerlessService MeshVirtualBoxVMWare云原生CNCF机器学习边缘计算
综合其他
BlenderGIMPKiCadKritaWeblate产品与服务人工智能亿图数据可视化版本控制笔试面试
文库资料
前端
AngularAnt DesignBabelBootstrapChart.jsCSS3EchartsElectronHighchartsHTML/CSSHTML5JavaScriptJerryScriptJestReactSassTypeScriptVue前端工具小程序
后端
.NETApacheC/C++C#CMakeCrystalDartDenoDjangoDubboErlangFastifyFlaskGinGoGoFrameGuzzleIrisJavaJuliaLispLLVMLuaMatplotlibMicronautnimNode.jsPerlPHPPythonQtRPCRubyRustR语言ScalaShellVlangwasmYewZephirZig算法
移动端
AndroidAPP工具FlutterFramework7HarmonyHippyIoniciOSkotlinNativeObject-CPWAReactSwiftuni-appWeex
数据库
ApacheArangoDBCassandraClickHouseCouchDBCrateDBDB2DocumentDBDorisDragonflyDBEdgeDBetcdFirebirdGaussDBGraphGreenPlumHStreamDBHugeGraphimmudbIndexedDBInfluxDBIoTDBKey-ValueKitDBLevelDBM3DBMatrixOneMilvusMongoDBMySQLNavicatNebulaNewSQLNoSQLOceanBaseOpenTSDBOracleOrientDBPostgreSQLPrestoDBQuestDBRedisRocksDBSequoiaDBServerSkytableSQLSQLiteTiDBTiKVTimescaleDBYugabyteDB关系型数据库数据库数据库ORM数据库中间件数据库工具时序数据库
云计算&大数据
ActiveMQAerakiAgentAlluxioAntreaApacheApache APISIXAPISIXBFEBitBookKeeperChaosChoerodonCiliumCloudStackConsulDaprDataEaseDC/OSDockerDrillDruidElasticJobElasticSearchEnvoyErdaFlinkFluentGrafanaHadoopHarborHelmHudiInLongKafkaKnativeKongKubeCubeKubeEdgeKubeflowKubeOperatorKubernetesKubeSphereKubeVelaKumaKylinLibcloudLinkerdLonghornMeiliSearchMeshNacosNATSOKDOpenOpenEBSOpenKruiseOpenPitrixOpenSearchOpenStackOpenTracingOzonePaddlePaddlePolicyPulsarPyTorchRainbondRancherRediSearchScikit-learnServerlessShardingSphereShenYuSparkStormSupersetXuperChainZadig云原生CNCF人工智能区块链数据挖掘机器学习深度学习算法工程边缘计算
UI&美工&设计
BlenderKritaSketchUI设计
网络&系统&运维
AnsibleApacheAWKCeleryCephCI/CDCurveDevOpsGoCDHAProxyIstioJenkinsJumpServerLinuxMacNginxOpenRestyPrometheusServertraefikTrafficUnixWindowsZabbixZipkin安全防护系统内核网络运维监控
综合其它
文章资讯
 上传文档  发布文章  登录账户
IT文库
  • 综合
  • 文档
  • 文章

无数据

分类

全部云计算&大数据(25)机器学习(25)

语言

全部中文(简体)(13)英语(12)

格式

全部PDF文档 PDF(25)
 
本次搜索耗时 0.020 秒,为您找到相关结果约 25 个.
  • 全部
  • 云计算&大数据
  • 机器学习
  • 全部
  • 中文(简体)
  • 英语
  • 全部
  • PDF文档 PDF
  • 默认排序
  • 最新排序
  • 页数排序
  • 大小排序
  • 全部时间
  • 最近一天
  • 最近一周
  • 最近一个月
  • 最近三个月
  • 最近半年
  • 最近一年
  • pdf文档 Lecture 4: Regularization and Bayesian Statistics

    Overfitting Problem 2 Regularized Linear Regression 3 Regularized Logistic Regression 4 MLE and MAP Feng Li (SDU) Regularization and Bayesian Statistics September 20, 2023 2 / 25 Overfitting Problem Bayesian Statistics September 20, 2023 13 / 25 Maximum-a-Posteriori Estimation (MAP) Maximum-a-Posteriori Estimation (MAP): Maximize the posterior prob- ability of θ (i.e., probability in the light of θ)dθ The Bayes Rule lets us update our belief about θ in the light of observed data While doing MAP, we usually maximize the log of the posteriori prob- ability Feng Li (SDU) Regularization and Bayesian
    0 码力 | 25 页 | 185.30 KB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniques

    Keeping all that in mind, it is easy to see that floating-point xmin should map to 0, and xmax should map to 2b-1. How do we map the rest of the floating point values in between xmin and xmax to integer continuous values are also clamped to be in the range [xmin, xmax]. Solution: Note that we have to map all the values from [xmin, xmax] to 2b possible values (let’s call them bins). Figure 2-5 shows a visual is referred to as scale). Hence, [xmin, xmin + s) will map to the bin 0, [xmin + s, xmin + 2s) will map to bin 1, [xmin + 2s, xmin + 3s) will map to bin 2, and so on. Thus, to find which bin the given
    0 码力 | 33 页 | 1.96 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 7 - Automation

    cast(image, tf.uint8) return image, label train_ds = train_ds.map(resize_image) val_ds = val_ds.map(resize_image) test_ds = test_ds.map(resize_image) Note that the create_model() function here has two a Reduction cell. A normal cell's output feature map is identical to the input feature map. In contrast, a reduction cell reduces the output feature map to half. Figure 7-7 shows two child networks that primitive operations. The concatenation operation happens along the filter dimension to keep the feature map intact. Figure 7-9 shows the Normal and Reduction cells predicted by NASNet with the cifar10 dataset
    0 码力 | 33 页 | 2.48 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 3 - Learning Techniques

    cast(image, tf.uint8) return image, label train_ds = train_ds.map(dsitem_to_tuple).map(resize_image).cache() val_ds = val_ds.map(dsitem_to_tuple).map(resize_image).cache() print(train_ds.as_numpy_iterator() layer composite of random flip and rotation. Then, we map each image through this layer to apply the transform. The mapping is done using a handy map() method on the tf.data.Dataset object. augs = tf.keras image = tf.squeeze(image, axis=0) # Squeeze the batch return image, label train_aug_ds = train_ds.map(augfn) That’s it! Simple, isn’t it? Let’s run the training function again with the augmented dataset
    0 码力 | 56 页 | 18.93 MB | 1 年前
    3
  • pdf文档 AI大模型千问 qwen 中文文档

    model = AutoModelForCausalLM.from_pretrained( "Qwen/Qwen1.5-7B-Chat", torch_dtype="auto", device_map="auto" ) tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen1.5-7B-Chat") # Instead of using model model = AutoModelForCausalLM.from_pretrained( "Qwen/Qwen1.5-7B-Chat", torch_dtype="auto", device_map="auto", attn_implementation="flash_attention_2", ) 为了解决下载问题,我们建议您尝试从 ModelScope 进行下载,只需将上述代码的第一行更改为以下内容: model = AutoModelForCausalLM.from_pretrained( "Qwen/Qwen1.5-7B-Chat", torch_dtype="auto", device_map="auto" ) tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen1.5-7B-Chat") # Instead of using model
    0 码力 | 56 页 | 835.78 KB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architectures

    the actual words by their indices in our vocabulary. If a word doesn’t exist in the vocabulary, we map it to the index of the OOV token. Similarly, we replace the label word with the index of that word end_to_end_model = tf.keras.Model(string_input, predictions) We can now classify a new piece of text, and map it to a class name. probabilities = end_to_end_model.predict( [['Usain Bolt is a very well known ‘hello’ map to the same slot in the embedding table and will learn the same embedding. Figure 4-13: Hashing trick as an option for reducing the vocabulary size. In the figure, ‘bar’ and ‘hello’ map to the
    0 码力 | 53 页 | 3.92 MB | 1 年前
    3
  • pdf文档 机器学习课程-温州大学-09深度学习-目标检测

    2.目标检测算法 mAP(Mean Average Precision) 多个类别的目标检测中,每一个类别都 可以绘制一条P-R曲线,各类别AP的均 值(即所有类别的AP和/类别数目)即 是mAP,mAP衡量的是训练出来的模型 在所有类别上的检测能力的好坏。 AP衡量的是学出来的模型在每个类别上的好坏,mAP衡量的是学出的模型在所有 类别上的好坏,得到AP后mAP的计算就变得很简单了,就是取所有AP的平均值。
    0 码力 | 43 页 | 4.12 MB | 1 年前
    3
  • pdf文档 全连接神经网络实战. pytorch 版

    测 试 的 数 据 download=True , #如 果 根 目 录 没 有 就 下 载 transform=ToTensor () ) #把 数 据 显 示 一 下 labels_map = { 0: ”T−Shirt ” , 1: ” Trouser ” , 2: ” Pullover ” , 3: ” Dress ” , 4: ”Coat” , 5: ” Sandal ” # 抽 取 索 引 为 100 的 数 据 来 显 示 img , l a b e l = training_data [ 1 0 0 ] plt . t i t l e ( labels_map [ l a b e l ] ) #squeeze 函 数 把 为1 的 维 度 去 掉 plt . imshow ( img . squeeze () , cmap=” gray ” ) plt 为numpy, 然 后 再 取 索 引 23 的 标 签 l a b e l = train_labels . numpy() [ 2 3 ] plt . t i t l e ( labels_map [ l a b e l ] ) plt . imshow (img , cmap=” gray ” ) plt . show () 程序得到显示结果: 数据有时候并不适合直接丢进网络进行训练,因此我们需要把数据进行转换。由于
    0 码力 | 29 页 | 1.40 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 6 - Advanced Learning Techniques - Technical Review

    model in this fine-tuning stage is not being used for learning rudimentary features, but rather how to map the high-level representations it learned in the pretraining stage to solving our new task. Thus, the model will work on. train_ds = batched_train.map( lambda x: (preprocessor(x['description']), tf.expand_dims(x['label'], axis=-1))) test_ds = batched_test.map( lambda x: (preprocessor(x['description'])
    0 码力 | 31 页 | 4.03 MB | 1 年前
    3
  • pdf文档 机器学习课程-温州大学-14深度学习-Vision Transformer (ViT)

    作为原始图像块的替代方法,输入序列可以由CNN的特征图形成。 在该混合模型中,将patch嵌入投影E应用于从CNN feature map中提取的patch。 作为一种特殊情况,patches的空间大小可以是1x1,这意味着输入序列是通过简单地打平 feature map的空间维度并投射到Transformer维度来获得的。如前所述,增加分类输入嵌入和 位置嵌入。 4.模型缺点与改进 30
    0 码力 | 34 页 | 2.78 MB | 1 年前
    3
共 25 条
  • 1
  • 2
  • 3
前往
页
相关搜索词
LectureRegularizationandBayesianStatisticsEfficientDeepLearningBookEDLChapterCompressionTechniquesAutomationAI模型千问qwen中文文档Architectures机器学习课程温州大学09深度目标检测连接神经网络神经网神经网络实战pytorchAdvancedTechnicalReview14VisionTransformerViT
IT文库
关于我们 文库协议 联系我们 意见反馈 免责声明
本站文档数据由用户上传或本站整理自互联网,不以营利为目的,供所有人免费下载和学习使用。如侵犯您的权益,请联系我们进行删除。
IT文库 ©1024 - 2025 | 站点地图
Powered By MOREDOC AI v3.3.0-beta.70
  • 关注我们的公众号【刻舟求荐】,给您不一样的精彩
    关注我们的公众号【刻舟求荐】,给您不一样的精彩