当前位置: 首页 > news >正文

网页设计网站概述怎么写seo推广网络

网页设计网站概述怎么写,seo推广网络,wordpress字体哪个好看,株洲关键词优化公司文章目录机器学习三大分支:有监督学习,无监督学习,强化学习无监督学习对输入的无标签数据的先验概率密度进行建模有监督学习试图推断出条件概率密度PX(x∣y)P_X(x|y)PX​(x∣y)无监督学习试图推断输入数据的先验概率分布PX(x)P_X(x)PX​(x)无…

文章目录

机器学习三大分支:有监督学习,无监督学习,强化学习

机器学习的三个主要分类是:有监督学习,无监督学习,强化学习。(并不是有监督,半监督,无监督!半监督只是综合利用了有监督和无监督技术,算是无监督和有监督的变体)

Unsupervised learning is a type of machine learning that looks for previously undetected patterns in a data set with no pre-existing labels and with a minimum of human supervision. In contrast to supervised learning that usually makes use of human-labeled data, unsupervised learning, also known as self-organization allows for modeling of probability densities over inputs.[1] It forms one of the three main categories of machine learning, along with supervised and reinforcement learning. Semi-supervised learning, a related variant, makes use of supervised and unsupervised techniques.

无监督学习对输入的无标签数据的先验概率密度进行建模

无监督学习,也叫做自组织,self-organization,对输入的概率密度进行建模。这个概率密度更精准的说,就是先验概率。

有监督学习试图推断出条件概率密度PX(x∣y)P_X(x|y)PX(xy)

即以输入数据的标签yyy为条件,输入数据xxx的概率密度。

用最简单的二分类举例,y取0或1, 假设数据也只有三种可能的离散情况x0,x1,x2x_0,x_1,x_2x0,x1,x2,有监督学习需要对数据集的每一个数据点的P(xi∣y=0),P(xi∣y=1),i=0,1,2P(x_i|y = 0), P(x_i|y = 1), i = 0,1,2P(xiy=0),P(xiy=1),i=0,1,2进行计算,比如数据集是{(x0,1),(x1,1),(x0,0),(x2,0),(x0,1),(x1,0),(x2,1),(x0,1),(x2,0),(x1,1)}\{(x_0, 1), (x_1, 1), (x_0, 0), (x_2, 0), (x_0, 1), (x_1, 0), (x_2, 1), (x_0, 1), (x_2, 0), (x_1, 1)\}{(x0,1),(x1,1),(x0,0),(x2,0),(x0,1),(x1,0),(x2,1),(x0,1),(x2,0),(x1,1)}
P(x0∣y=0)=0.25,P(x0∣y=1)=0.5P(x_0|y = 0) = 0.25, P(x_0|y = 1) = 0.5P(x0y=0)=0.25,P(x0y=1)=0.5
P(x1∣y=0)=0.25,P(x1∣y=1)=0.33P(x_1|y = 0) = 0.25, P(x_1|y = 1) = 0.33P(x1y=0)=0.25,P(x1y=1)=0.33
P(x2∣y=0)=0.5,P(x2∣y=1)=0.16P(x_2|y = 0) = 0.5, P(x_2|y = 1) = 0.16P(x2y=0)=0.5,P(x2y=1)=0.16

这是根据当前数据集所计算出来的精确的条件概率值,但是这是最简单的情况,即输入离散且可能情况少,标签也少。并且还有两个问题需要考虑:

一,这个数据集太小(是我随便编的),小数据集很有可能无法真实反映源域的实际概率分布情况,所以这样训练出来的模型泛化能力极差,这可以通过增大数据集和提高数据集质量来缓解,这就是为什么有监督学习领域数据集的制作也是个重要的精细的体力活儿;

二,这里做的假设比较简单,输入数据只有三种离散情况,但是实际中,输入数据要么是离散的但有更多可能情况,这就需要更多数据,以防止数据集中某种情况没出现导致模型以后都不认识这种情况;要么输入数据是连续取值的,这时候有监督学习的分类器要学的就是条件概率分布ccdf了,或者说条件概率密度cpdf(cpdf的积分是ccdf),而不是仅仅计算条件概率的值,这是更常见更普通的情况。

无监督学习试图推断输入数据的先验概率分布PX(x)P_X(x)PX(x)

It could be contrasted with supervised learning by saying that whereas supervised learning intends to infer a conditional probability distribution PX(x∣y)P_X(x|y)PX(xy) conditioned on the label yyy of input data; unsupervised learning intends to infer an a priori probability distribution PX(x)P_X(x)PX(x).

比如数据集是{x0,x1,x0,x2,x0,x1,x2,x0,x2,x1}\{x_0, x_1, x_0, x_2, x_0, x_1, x_2, x_0, x_2, x_1\}{x0,x1,x0,x2,x0,x1,x2,x0,x2,x1}

PX(x0)=0.4,PX(x1)=0.3,PX(x2)=0.3P_X(x_0) = 0.4, P_X(x_1) = 0.3, P_X(x_2) = 0.3PX(x0)=0.4,PX(x1)=0.3,PX(x2)=0.3

无监督学习的两种主要方法:PCA和聚类

PCA(主成分分析)

提取主要成分,以降低维度,是完全线性的,即每一个主成分都是独立的,对应一个特征值。所以每一个主成分都在不同的方向上,彼此完全正交,因此相互独立。

PCA计算数据的协方差矩阵的特征值,保留前n个特征值(以保证累计贡献率达到80%以上),前n个特征值的特征想向量们构建了一个n维空间,PCA就相当于把数据放到了这个n维的高维度的空间去审视,每一个主成分在一个独立的维度(方向上)。

自编码器:以无监督方式学习数据编码的神经网络,被称为非线性PCA

Autoencoder,自编码器的目标是为数据学出一组表示,representation或者叫做encodings,这样做主要的好处是降维。
自编码器的变体有正则化自编码器regularized autoencoders (比如稀疏自编码器,去噪自编码器,收缩自编码器,这些正则化自编码器可以学到很好的特征表示,对后续的有监督分类器很有帮助)和变分自编码器(主要被用作为生成模型)。

An autoencoder is a type of artificial neural network used to learn efficient data codings in an unsupervised manner.[1] The aim of an autoencoder is to learn a representation (encoding) for a set of data, typically for dimensionality reduction, by training the network to ignore signal “noise”. Along with the reduction side, a reconstructing side is learnt, where the autoencoder tries to generate from the reduced encoding a representation as close as possible to its original input, hence its name. Several variants exist to the basic model, with the aim of forcing the learned representations of the input to assume useful properties.[2] Examples are the regularized autoencoders (Sparse, Denoising and Contractive autoencoders), proven effective in learning representations for subsequent classification tasks,[3] and Variational autoencoders, with their recent applications as generative models.[4] Autoencoders are effectively used for solving many applied problems, from face recognition[5] to acquiring the semantic meaning of words。

自编码器80年代就火了,至今用的最多的是降维和特征学习,但是近些年也被用来学习数据的生成(比如由文本描述生成图像,一般都是用的变分自编码器)。

An autoencoder is a neural network that learns to copy its input to its output. It has an internal (hidden) layer that describes a code used to represent the input, and it is constituted by two main parts: an encoder that maps the input into the code, and a decoder that maps the code to a reconstruction of the original input.
Performing the copying task perfectly would simply duplicate the signal, and this is why autoencoders usually are restricted in ways that force them to reconstruct the input approximately, preserving only the most relevant aspects of the data in the copy.
The idea of autoencoders has been popular in the field of neural networks for decades, and the first applications date back to the '80s. Their most traditional application was dimensionality reduction or feature learning, but more recently the autoencoder concept has become more widely used for learning generative models of data. Some of the most powerful AIs in the 2010s involved sparse autoencoders stacked inside of deep neural networks.

聚类分析

聚类是机器学习的一个分支,他把没有标签的数据聚集起来,聚类不是对反馈做出反应(有监督学习,分类任务),而是识别数据中的共性,然后基于这些共性在每一份数据中的出现与否做出反应。聚类很适合检测出不适合任何组团的异常数据点。

Two of the main methods used in unsupervised learning are principal component and cluster analysis. Cluster analysis is used in unsupervised learning to group, or segment, datasets with shared attributes in order to extrapolate algorithmic relationships.[2] Cluster analysis is a branch of machine learning that groups the data that has not been labelled, classified or categorized. Instead of responding to feedback, cluster analysis identifies commonalities in the data and reacts based on the presence or absence of such commonalities in each new piece of data. This approach helps detect anomalous data points that do not fit into either group.\

无监督学习的中心应用:统计学中的密度估计

无监督学习的中心应用是是统计学的密度估计领域。

A central application of unsupervised learning is in the field of density estimation in statistics,

http://www.lbrq.cn/news/2531971.html

相关文章:

  • 新北网站建设引擎优化是什么意思
  • 外语网站制作如何做自己的网站
  • 仪征网站建设seo引擎优化外包公司
  • 外贸商城网站资质seo优化排名技术百度教程
  • 做网站什么职业廊坊网站推广公司
  • 做直播平台网站赚钱吗互联网销售平台
  • 假冒网站能通过备案登记吗电商网站建设公司哪家好
  • 昆明房地产网站建设百度引擎
  • 优秀国内个人网站网址sem推广软件选哪家
  • 宝鸡英文网站建设短视频矩阵seo系统源码
  • to a wordpress上海seo有哪些公司
  • 做金融网站看那些素材宁波优化网页基本流程
  • 杭州网站建设服务企业网站开发费用
  • 在哪个网站找婚照公司网络代运营推广
  • 网站怎么做更新吗如何优化网站
  • 有什么网站可以做宣传网络推广十大平台
  • 做设计需要知道的几个网站吗中国网民博客 seo
  • 北京做网站的公司上海网站推广服务公司
  • 做网站上的在线支付怎么做长沙企业seo服务
  • js网站源码已到期信息流推广主要具有哪两大优势
  • 做网站店铺图片用什么软件搜索引擎营销策划方案
  • 东丰在线网站建设成都移动seo
  • 西宁网站制作宁波谷歌优化
  • 做局域网网站教程网站设计说明
  • 福田网站建设龙岗网站建设龙岗网站建设推销产品的万能句子
  • tp框架做展示网站重庆seo
  • 2017网站趋势百度推广一级代理商名单
  • 常德网站优化想开个网站怎样开
  • 网站建设装什么系统湖南企业竞价优化公司
  • 做网站域名需哪些百度手机下载安装
  • java8学习笔记-Stream流
  • 嵌入式教学的云端革命:高精度仿真如何重塑倒车雷达实验与工程教育——深圳航天科技创新研究院赋能新一代虚实融合实训平台
  • Codeforces Round 1040 (Div. 2) E1 - E3 交互题 | 思维
  • 无监督MVSNet系列网络概述
  • p5.js 从零开始创建 3D 模型,createModel入门指南
  • ESP32学习-1.第一个程序helloworld