文档章节

人工智能资料库:第12辑(20170120)

AllenOR灵感
 AllenOR灵感
发布于 2017/09/10 01:27
字数 1029
阅读 3
收藏 0

  1. 【博客】Understanding the new Google Translate

简介:


Google launched a new version of the Translate in September 2016. Since then, there have been a few interesting developments in the project, and this post attempts to explain it all in as simple terms as possible.

The earlier version of the Translate used Phrase-based Machine Translation, or PBMT. What PBMT does is break up an input sentence into a set of words/phrases and translate each one individually. This is obviously not an optimal strategy, since it completely misses out on the context of the overall sentence. The new Translate uses what Google calls *Google Neural Machine Translation (*GNMT**), an improvement over a traditional version of NMT. Lets see how GNMT works on a high-level:

原文链接:https://codesachin.wordpress.com/2017/01/18/understanding-the-new-google-translate/


2.【博客 & 代码】Self-Organizing Maps with Google’s TensorFlow

简介:

A Self-Organizing Map, or SOM, falls under the rare domain of unsupervised learning in Neural Networks. Its essentially a grid of neurons, each denoting one cluster learned during training. Traditionally speaking, there is no concept of neuron ‘locations’ in ANNs. However, in an SOM, each neuron has a location, and neurons that lie close to each other represent clusters with similar properties. Each neuron has a weightage vector, which is equal to the centroid of its particular cluster.

原文链接:https://codesachin.wordpress.com/2015/11/28/self-organizing-maps-with-googles-tensorflow/

原理链接:http://www.ai-junkie.com/ann/som/som1.html


3.【博客】Simple Beginner’s guide to Reinforcement Learning & its implementation

简介:

这篇博客,我近期会把它翻译成中文,并且做一个学习笔记。

One of the most fundamental question for scientists across the globe has been – “How to learn a new skill?”. The desire to understand the answer is obvious – if we can understand this, we can enable human species to do things we might not have thought before. Alternately, we can train machines to do more “human” tasks and create true artificial intelligence.

While we don’t have a complete answer to the above question yet, there are a few things which are clear. Irrespective of the skill, we first learn by interacting with the environment. Whether we are learning to drive a car or whether it an infant learning to walk, the learning is based on the interaction with the environment. Learning from interaction is the foundational underlying concept for all theories of learning and intelligence.

原文链接:https://www.analyticsvidhya.com/blog/2017/01/introduction-to-reinforcement-learning-implementation/


4.【论文】Revisiting Visual Question Answering Baselines

简介:

Visual question answering (VQA) is an interesting learning setting for evaluating the abilities and shortcomings of current systems for image understanding. Many of the recently proposed VQA systems include attention or memory mechanisms designed to support “reasoning”. For multiple-choice VQA, nearly all of these systems train a multi-class classifier on image and question features to predict an answer. This paper questions the value of these common practices and develops a simple alternative model based on binary classification. Instead of treating answers as competing choices, our model receives the answer as input and predicts whether or not an image-question-answer triplet is correct. We evaluate our model on the Visual7W Telling and the VQA Real Multiple Choice tasks, and find that even simple versions of our model perform competitively. Our best model achieves state-of-the-art performance on the Visual7W Telling task and compares surprisingly well with the most complex systems proposed for the VQA Real Multiple Choice task. We explore variants of the model and study its transferability between both datasets. We also present an error analysis of our model that suggests a key problem of current VQA systems lies in the lack of visual grounding of concepts that occur in the questions and answers. Overall, our results suggest that the performance of current VQA systems is not significantly better than that of systems designed to exploit dataset biases.

原文链接:https://arxiv.org/pdf/1606.08390v2.pdf


5.【Tutorial & 代码】Introduction to Natural Language Processing with fastText

简介:

这篇博客,我近期会把它翻译成中文,并且做一个学习笔记。

Natural Language Processing (NLP) is one of the hottest areas in machine learning. Its global purpose is to understand language the way humans do. NLP subareas include machine translation, text classification, speech recognition, sentiment analysis, question answering, text-to-speech, etc.

As in most areas of Machine Learning, NLP accuracy has improved considerably thanks to deep learning. Just to highlight the most recent and impressive achievement, in October 2016 Microsoft Research reached human parity in speech recognition. For that milestone, they used a combination of Convolutional Neural Networks and LSTM networks.

However, not all machine learning is deep learning, and in this notebook I would like to highlight a great example. In the summer of 2016, two interesting NLP papers were published by Facebook Research, Bojanowski et al., 2016 and Joulin et al., 2016. The first one proposed a new method for word embedding and the second one a method for text classification. The authors also opensourced a C++ library with the implementation of these methods, fastText, that rapidly attracted a lot of interest.

The reason for this interest is that fastText obtains an accuracy in text classification almost as good as the state of the art in deep learning, but it is several orders of magnitude faster. In their paper, the authors compare the accuracy and computation time of several datasets with deep nets. As an example, in the Amazon Polarity dataset, fastText achieves an accuracy of 94.6% in 10s. In the same dataset, the crepe CNN model of Zhang and LeCun, 2016 achieves 94.5% in 5 days, while the Very Deep CNN model of Conneau et al., 2016 achieves 95.7% in 7h. The comparison is not even fair, because while fastText's time is computed with CPUs, the CNN models are computed using Tesla K40 GPUs.

原文链接:https://github.com/miguelgfierro/sciblog_support/blob/master/Intro_to_NLP_with_fastText/Intro_to_NLP.ipynb


本文转载自:http://www.jianshu.com/p/ec20a33aa2f2

共有 人打赏支持
AllenOR灵感
粉丝 11
博文 2635
码字总数 83001
作品 0
程序员
私信 提问
预测流行偏好,时尚 AI 未来可望取代造型师

【Technews科技新报】预测时尚潮流是一项需要天分的工作,还得仰赖一个庞大的系统让少数人追捧的时尚进入大众流行市场,进而让业者赚取大笔钞票。现在预测工作也可以交给人工智能,让服饰业者...

黄 嬿
2017/12/26
0
0
【Java每日一题】20170120

20170119问题解析请点击今日问题下方的“【Java每日一题】20170120”查看(问题解析在公众号首发,公众号ID:weknow619) 今日问题: 请问主程序输出结果是什么?(点击以下“【Java每日一题...

weknow
2017/01/22
0
0
人工智能资料库:第72辑(20171203)

1.【会议】Bayesian Deep Learning 简介: While deep learning has been revolutionary for machine learning, most modern deep learning models cannot represent their uncertainty nor......

chen_h
2017/12/03
0
0
【Java每日一题】20170206

20170120问题解析请点击今日问题下方的“【Java每日一题】20170206”查看(问题解析在公众号首发,公众号ID:weknow619) 今日问题: 请问主程序有无编译错误?(点击以下“【Java每日一题】...

weknow
2017/02/06
0
0
人工智能知识整理-第1辑(20170603)-机器学习入门资源汇总

有一天我忽然忘记了一个函数的用法,于是就上谷歌搜,结果搜出来的竟然是自己写的一篇笔记,上面有很详细的回答。当时感觉是跟另外一个自己进行交流,那一个是刚学完知识,印象还非常深的自己...

人工智豪
2017/06/03
0
0

没有更多内容

加载失败,请刷新页面

加载更多

微服务分布式事务实现

https://www.processon.com/view/link/5b2144d7e4b001a14d3d2d30

WALK_MAN
今天
2
0
《大漠烟尘》读书笔记及读后感文章3700字

《大漠烟尘》读书笔记及读后感文章3700字: 在这个浮躁的社会里,你有多久没有好好读完一本书了? 我们总觉得自己和别人不一样,所以当看到别人身上的问题时,很少有“反求诸己”,反思自己。...

原创小博客
今天
3
0
大数据教程(9.5)用MR实现sql中的jion逻辑

上一篇博客讲解了使用jar -jar的方式来运行提交MR程序,以及通过修改YarnRunner的源码来实现MR的windows开发环境提交到集群的方式。本篇博主将分享sql中常见的join操作。 一、需求 订单数据表...

em_aaron
今天
3
0
十万个为什么之什么是resultful规范

起源 越来越多的人开始意识到,网站即软件,而且是一种新型的软件。这种"互联网软件"采用客户端/服务器模式,建立在分布式体系上,通过互联网通信,具有高延时(high latency)、高并发等特点...

尾生
今天
3
0
Terraform配置文件(Terraform configuration)

Terraform配置文件 翻译自Terraform Configuration Terraform用文本文件来描述设备、设置变量。这些文件被称为Terraform配置文件,以.tf结尾。这一部分将讲述Terraform配置文件的加载与格式。...

buddie
今天
5
0

没有更多内容

加载失败,请刷新页面

加载更多

返回顶部
顶部