Retrofitting Analysis
Retrofitting Analysis
To figure out the process of retrofitting[1] objective updating, we do the following math.
Forward Derivation
\[
\psi(Q) = \sum_{i=1}^{n}\left[ \alpha_i||q_i-\hat{q_i}||^2 + \sum\beta||q_i-q_j||^2 \right] \\
\frac{\partial \psi(Q)}{\partial q_i} = \alpha_i(q_i-\hat{q_i}) + \sum\beta(q_i-q_j) = 0 \\
(\alpha_i+\sum\beta_{ij})q_i -\alpha_i\hat{q_i} -\sum\beta_{ij}q_j = 0 \\
q_i = \frac{\sum\beta_{ij}q_j+\alpha_i\hat{q_i}}{\sum\beta_{ij}+\alpha_i}
\]
Backward Derivation
This is how I understood this updating equation.
In the paper[1], it has mentioned "We take the first derivative of \(\psi\) with respect to one qi vector, and by equating it to zero", hence we get follow idea:
\[
\frac{\partial\psi(Q)}{\partial q_i} = 0
\]
And,
\[
q_i = \frac{\sum\beta_{ij}q_j+\alpha_i\hat{q_i}}{\sum\beta_{ij}+\alpha_i} \\
\alpha_iq_i - \alpha_i\hat{q_j} + \sum\beta_{ij}q_i - \sum\beta q_j = 0 \\
\alpha_i(q_i-\hat{q_j})+ \sum\beta_{ij}(q_i-q_j) = 0
\]
Apparently,
\[
\frac{\partial\psi(Q)}{\partial q_i} = \alpha_i(q_i-\hat{q_j})+ \sum\beta_{ij}(q_i-q_j) = 0
\]
Reference
Faruqui M, Dodge J, Jauhar S K, et al. Retrofitting Word Vectors to Semantic Lexicons[J]. ACL, 2015.
最新文章
- CentOS 搭建openVPN
- MySQL数据库命名及设计规范
- HDFS分布式文件系统资源管理器开发总结
- 团队开发——冲刺2.d
- 【py网页】urlopen的补充,完美
- SCCM日志存放路径
- QT事件研究的文章
- 使用hibernate在5秒内插入11万条数据,你觉得可能吗?
- h5面试题
- PhantomJS报错warnings.warn('Selenium support for PhantomJS has been deprecated, please use headless '
- 获取input标签中file的内容
- JS模拟滚动条(有demo和源码下载,支持拖动 滚轮 点击事件)
- IIS通过HTML5实现应用程序缓存的离线浏览
- psql: 致命错误: 对用户";user1";的对等认证失败
- Disruptor时序调用
- Web 端屏幕适配方案
- 使用 Rational AppScan 保证 Web 应用的安全性,第 1 部分: Web 安全与 Rational AppScan 入门
- 197. Rising Temperature
- PCU
- 记一次MySQL手工注入
热门文章
- 让图表的Y轴 产生几个刻度距离
- 微信小程序刮刮乐
- dedecms 公共模板写法 提高生成速度
- 第二篇.1、python基础之数据类型与变量
- deep_learning_Function_tf.identity()
- Linux查看系统开机和关机时间
- 修改linux系统TCP连接数
- 完美解决Mysql的Access denied for user 'root'@'%的'问题
- docker查看容器IP地址
- 关于深度学习之TensorFlow简单实例