日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問(wèn) 生活随笔!

生活随笔

當(dāng)前位置: 首頁(yè) > 编程资源 > 编程问答 >内容正文

编程问答

(转)Predictive learning vs. representation learning 预测学习 与 表示学习

發(fā)布時(shí)間:2023/12/31 编程问答 34 豆豆
生活随笔 收集整理的這篇文章主要介紹了 (转)Predictive learning vs. representation learning 预测学习 与 表示学习 小編覺(jué)得挺不錯(cuò)的,現(xiàn)在分享給大家,幫大家做個(gè)參考.

?

Predictive learning vs. representation learning ?預(yù)測(cè)學(xué)習(xí) 與 表示學(xué)習(xí)

?

When you take?a machine learning class, there’s a good chance it’s divided into a unit on supervised learning and a unit on unsupervised learning. We certainly care about this distinction for a practical reason: often there’s orders of magnitude more data available if we don’t need to collect ground-truth labels. But we also tend to think it matters for more fundamental reasons. In particular, the following are some common intuitions:

  • In supervised learning, the particular algorithm is usually less important than engineering and tuning it really well. In unsupervised learning, we’d think carefully about the structure of the data and build a model which reflects that structure.
  • In supervised learning, except in small-data settings, we throw whatever features we can think of at the problem. In unsupervised learning, we carefully pick the features we think best represent the aspects of the data we care about.
  • Supervised learning seems to have many algorithms with strong theoretical guarantees, and unsupervised learning very few.
  • Off-the-shelf algorithms perform very well on a wide variety of supervised tasks, but unsupervised learning requires more care and expertise to come up with an appropriate model.

I’d argue that this is deceptive. I think real division in machine learning isn’t between supervised and unsupervised, but what I’ll term?predictive learning?and?representation learning. I haven’t heard it described in precisely this way before, but I think this distinction reflects a lot of our intuitions about how to approach a given machine learning problem.

In predictive learning, we observe data drawn from some distribution, and we are interested in predicting some aspect of this distribution. In textbook supervised learning, for instance, we observe a bunch of pairs?, and given some new example?, we’re interested in predicting something about the corresponding?. In density modeling (a form of unsupervised learning), we observe unlabeled data?, and we are interested in modeling the distribution the data comes from, perhaps so we can perform inference in that distribution. In each of these cases, there is a well-defined predictive task where we try to predict some aspect of the observable values possibly given some other aspect.

In representation learning, our goal isn’t to predict observables, but to learn something about the underlying structure. In cognitive science and AI, a representation is a formal system which maps to some domain of interest in systematic ways. A good representation allows us to answer queries about the domain by manipulating that system. In machine learning, representations often take the form of vectors, either real- or binary-valued, and we can manipulate these representations with operations like Euclidean distance and matrix multiplication. For instance, PCA learns representations of data points as vectors. We can ask how similar two data points are by checking the Euclidean distance between them.

In representation learning, the goal isn’t to make predictions about observables, but to learn a representation which would later help us to answer various queries.?Sometimes the representations are meant for people, such as when we visualize data as a two-dimensional embedding. Sometimes they’re meant for machines, such as when the binary vector representations learned by deep Boltzmann machines are fed into a supervised classifier. In either case, what’s important is that mathematical operations map to the underlying relationships in the data in systematic ways.

Whether your goal is prediction or representation learning influences the sorts of techniques you’ll use to solve the problem. If you’re doing predictive learning, you’ll probably try to engineer a system which exploits as much information as possible about the data, carefully using a validation set to tune parameters and monitor overfitting. If you’re doing representation learning, there’s no good quantitative criterion, so you’ll more likely build a model based on your intuitions about the domain, and then keep staring at the learned representations to see if they make intuitive sense.

In other words, it parallels the differences I listed above between supervised and unsupervised learning. This shouldn’t be surprising, because the two dimensions are strongly correlated: most supervised learning is predictive learning, and most unsupervised learning is representation learning. So to see which of these dimensions is really the crux of the issue, let’s look at cases where the two differ.

Language modeling is a perfect example of an application which is unsupervised but predictive. The goal is to take a large corpus of unlabeled text (such as Wikipedia) and learn a distribution over English sentences. The problem is motivated by Bayesian models for speech recognition: a distribution over sentences can be used as a prior for what a person is likely to say. The goal, then, is to model the distribution, and any additional structure is unnecessary. Log-linear models, such as that of Mnih et al. [1], are very good at this, and recurrent neural nets [2] are even better. These are the sorts of approaches we’d normally apply in a supervised setting: very good at making predictions, but often hard to interpret. One state-of-the-art algorithm for density modeling of text is PAQ [3], which is a heavily engineered ensemble of sequential predictors, somewhat reminiscent of the winning entries of the Netflix competition.

On the flip side, supervised neural nets are often used to learn representations. One example is Collobert-Weston networks [4], which attempt to solve a number of supervised NLP tasks by learning representations which are shared between them. Some of the tasks are fairly simple and have a large amount of labeled data, such as predicting which of two words should be used to fill in the blank. Others are harder and have less data available, such as semantic role labeling. The simpler tasks are artificial, and they are there to help learn a representation of words and phrases as vectors, where similar words and phrases map to nearby vectors; this representation should then help performance on the harder tasks. We don’t care about the performance on those tasks per se; we care whether the learned embeddings reflect the underlying structure. To debug and tune the algorithm, we’d focus on whether the representations make intuitive sense, rather than on the quantitative performance. There are no theoretical guarantees that such an approach would work — it all depends on our intuitions of how the different tasks are related.

Based on these two examples, it seems like it’s the predictive/representation dimension which determines how we should approach the problem, rather than supervised/unsupervised.

In machine learning, we tend to think there’s no solid theoretical framework for unsupervised learning. But really, the problem is that we haven’t begun to formally characterize the problem of representation learning. If you just want to build a density modeler, that’s about as well understood as the supervised case. But if the goal is to learn representations which capture the underlying structure, that’s much harder to formalize. In my next post, I’ll try to take a stab at characterizing what representation learning is actually about.

[1] Mnih, A., and Hinton, G. E. Three new graphical models for statistical language modeling. NIPS 2009

[2] Sutskever, I., Martens, J., and Hinton, G. E. Generating text with recurrent neural networks. ICML 2011

[3] Mahoney, M. Adaptive weighting of context models for lossless data compression. Florida Institute of Technology Tech report, 2005

[4] Collobert, R., and Weston, J. A unified architecture for natural language processing: deep neural networks with multitask learning. ICML 2008

?

Posted in?Machine Learning.

總結(jié)

以上是生活随笔為你收集整理的(转)Predictive learning vs. representation learning 预测学习 与 表示学习的全部?jī)?nèi)容,希望文章能夠幫你解決所遇到的問(wèn)題。

如果覺(jué)得生活随笔網(wǎng)站內(nèi)容還不錯(cuò),歡迎將生活随笔推薦給好友。

主站蜘蛛池模板: 97精品久久| 少妇搡bbbb搡bbb搡澳门 | 亚洲激情啪啪 | 欧美日韩高清在线观看 | 五十路av| 一眉道姑 电影 | 欧美高清 | 精品无码在线观看 | 国产香蕉视频在线播放 | 羞羞涩| 91禁动漫在线 | www.色欧美 | 乳孔很大能进去的av番号 | 丁香六月五月婷婷 | 天堂网va | 久一国产 | 国产精品极品 | 色呦呦精品 | 日本成人黄色片 | 亚洲最大的黄色网 | 国产91看片 | 麻豆亚洲精品 | 成人a毛片| 欧美亚洲精品天堂 | √资源天堂中文在线视频 | 亚洲av色香蕉一区二区三区 | 四虎永久免费在线观看 | 黄色av免费在线看 | 在线青草 | 欧美黑人xxx | 天天躁夜夜躁狠狠躁 | 国产影音先锋 | 91精品国产综合久久久久久 | 亚洲乱码中文字幕久久孕妇黑人 | 国产精品高潮呻吟久久久 | 黄色a免费 | 亚洲深夜福利视频 | 激情黄色小说网站 | 又色又爽又高潮免费视频国产 | 国产一级二级三级在线观看 | 麻豆视频在线观看免费网站 | 精品久久久久久久久久久久久久久久久 | 极品美女扒开粉嫩小泬 | 国产女上位 | www成人 | 国产精品成人久久久 | 波多野结衣电车痴汉 | 日韩极品少妇 | 男女拍拍拍网站 | 91一区| 中文字幕人妻丝袜乱一区三区 | 亚洲第一淫片 | 美女被草视频 | 视频二区欧美 | 丝袜五月天 | 黄色大片在线播放 | 国产成人无码一区二区在线观看 | 在线免费小电影 | 日本一二三区视频在线 | 无套暴操 | 国产乱淫片视频 | 日韩a级在线观看 | 在线电影一区二区 | 涩色视频 | 成年人看的视频网站 | 久久精品123| 另类老妇性bbwbbw图片 | 国产成人av在线 | www.中文字幕av | 国产免费av电影 | 最新国产网址 | 婷婷色一区二区三区 | 一级片免费在线观看 | 精品视频免费观看 | 偷拍青青草 | 无码人妻一区二区三区一 | 人妻夜夜爽天天爽 | 亚色成人 | 一本一道久久 | 欧美sm凌虐视频网站 | 国产精品伦一区二区三级古装电影 | 国产日韩欧美91 | 国产亚洲小视频 | 伊人久久麻豆 | 欧美做爰爽爽爽爽爽爽 | 九九看片 | 影音先锋男人天堂 | www.夜夜爱 | 91精品国产99 | 欧美一区二区福利视频 | 日本免费成人 | 日韩视频一区二区三区 | 99精品视频在线 | 另类一区二区三区 | 日本深夜福利 | 91超碰在线播放 | 91网站在线观看视频 | 免费看黄网站在线 | 亚洲乱码久久 |