Deeplearning Algorithms tutorial
谷歌的人工智能位于全球前列,在图像识别、语音识别、无人驾驶等技术上都已经落地。而百度实质意义上扛起了国内的人工智能的大旗,覆盖无人驾驶、智能助手、图像识别等许多层面。苹果业已开始全面拥抱机器学习,新产品进军家庭智能音箱并打造工作站级别Mac。另外,腾讯的深度学习平台Mariana已支持了微信语音识别的语音输入法、语音开放平台、长按语音消息转文本等产品,在微信图像识别中开始应用。全球前十大科技公司全部发力人工智能理论研究和应用的实现,虽然入门艰难,但是一旦入门,高手也就在你的不远处! AI的开发离不开算法那我们就接下来开始学习算法吧!
局部线性嵌入(Locally Linear Embedding)
数据降维是指通过线性的或非线性的映射关系将高维数据转换成低维数据的过程。一般情况下,该低维数据代表了原始高维数据的主要成分(图1),并描述了原始高维数据的空间分布结构。由于经过降维后的数据更易于被分类、识别、可视化以及存储等,故数据降维技术在诸多科研领域受到了越来越多地关注。
从数据本身的的性质特征来看,数据降维可以大致分为线性降维和非线性降维两种技术方法。其中,线性降维技术仅对于数据维数相对较低、且具有全局线性结构的数据有着很好的降维效果。然而,在实际的科学研究中,科研工作者却需要面对海量的非线性高维数据。因此,能够有效处理高维非线性数据的方法亟待被提出,本文将介绍一种用于处理高维非线性数据的降维方法。
局部线性嵌入(Locally linear embedding, LLE)是一种非线性的降维方法,该算法由 Sam T.Roweis等人于2000年提出并发表在《Science》杂志上。LLE试图保留原始高维数据的局部性质,通过假设局部原始数据近似位于一张超平面上,从而使得该局部的某一个数据可以由其邻域数据线性表示。
Sam T.Roweis 和 Lawrence K.Saul提出局部线性嵌入(Locally linear embedding, LLE)算法,它是针对非线性数据的一种新的降维技术,并且能够使降维后的数据保持原有的拓扑结构。 LLE算法可以广泛的应用于非线性数据的降维、聚类以及图像分割等领域。
局部线性嵌入(Locally linear embedding, LLE)是最新提出的非线性降维方法。该算法即具有处理非线性数据的优点又有线性降维方法计算性能的优越性。 简单的讲,该方法是将高维流型用剪刀剪成很多的小块,每一小块可以用平面代替,然后再低维中重新拼合出来, 且要求保留各点之间的拓扑关系不变。整个问题最后被转化为两个二次规划问题。
局部线性嵌入(Locally linear embedding, LLE)算法可以归结为三步:
- 寻找每个样本点的k个近邻点;
- 由每个样本点的近邻点计算出该样本点的局部重建权值矩阵;
- 由该样本点的局部重建权值矩阵和其近邻点计算出该样本点的输出值。
局部线性嵌入算法的第一步是计算出每个样本点的k个近邻点。把相对于所求样本点距离最近的k个样本点规定为所求样本点的k个近邻点。k是一个预先给定值。Sam T.Roweis 和 Lawrence K.Saul算法采用的是欧氏距离,则减轻复杂的计算。然而本文是假定高维空间中的数据是非线性分布的,采用了diijstra距离。Dijkstra 距离是一种测地距离,它能够保持样本点之间的曲面特性,在ISOMAP算法中有广泛的应用。针对样本点多的情况,普通的dijkstra算法不能满足LLE算法的要求。
应用示例:
import numpy
import sources
from sklearn import manifold
from sklearn.utils.extmath import randomized_svd
from sklearn.neighbors import NearestNeighbors
from wordreps import WordReps
from scipy.sparse import csr_matrix, eye
from scipy.sparse.linalg import eigsh
from scipy.sparse.linalg.eigen.arpack.arpack import ArpackNoConvergence
from scipy.io import savemat, loadmat
import sys
import time
import argparse
import collections
class MetaEmbed():
def __init__(self, wordreps, words):
self.words = words
self.ids = {}
for (i,word) in enumerate(words):
self.ids[word] = i
self.reps = wordreps
self.dims = [x.dim for x in self.reps]
self.embeds = []
N = len(words)
# Create the source embedding matrices
write("Creating source embedding matrices...")
for i in range(len(self.reps)):
M = numpy.zeros((self.reps[i].dim, N), dtype=numpy.float64)
for j in range(N):
M[:,j] = self.reps[i].vects[self.words[j]]
self.embeds.append(M)
write("done\n")
pass
def compute_neighbours(self, nns):
"""
Compute the nearest neighbours for each embedding.
"""
self.NNS = []
for i in range(len(self.embeds)):
start_time = time.clock()
write("Computing nearest neighbours for embedding no = %d ..." % i)
nbrs = NearestNeighbors(n_neighbors=nns, algorithm='ball_tree').fit(self.embeds[i].T)
distances, indices = nbrs.kneighbors(self.embeds[i].T)
self.NNS.append(indices[:,1:])
end_time = time.clock()
write("Done (%s sec.)\n" % str(end_time - start_time))
pass
def show_nns(self, word ,nns):
"""
Print nearest neigbours for a word in different embeddings.
"""
for i in range(len(self.embeds)):
print "Showing nearest neighbours for = %s" % word
print "\nEmbedding no = %d" % i
for s in self.NNS[i][self.ids[word], :][:nns]:
print self.words[s]
pass
def compute_weights(self):
"""
Computes the reconstruction weights.
"""
start_time = time.clock()
T = 10 # no. of iterations.
alpha = 0.01 # learning rate.
N = len(self.words)
self.W = numpy.zeros((N, N), dtype=numpy.float64)
# initialise the weights.
for i in range(N):
nns = set()
for j in range(len(self.embeds)):
for x in self.NNS[j][i,:]:
nns.add(x)
val = 1.0 / float(len(nns))
for j in nns:
self.W[i,j] = val
# iterate
for i in range(N):
write("\x1b[2K\rLearning weights for (%d of %d) = %s" % (i, N, self.words[i]))
for t in range(T):
d = [self.embeds[j][:,i] - numpy.sum([self.W[i,k] * self.embeds[j][:,k] for k in self.NNS[j][i,:]], axis=0) for j in range(len(self.embeds))]
#for j in range(len(self.embeds)):
# d.append(self.embeds[j][:,i] - numpy.sum([self.W[i,k] * self.embeds[j][:,k] for k in self.NNS[j][i,:]], axis=0))
grad = numpy.zeros(N, dtype=numpy.float64)
for j in range(len(self.embeds)):
for k in self.NNS[j][i,:]:
grad[k] += -2.0 * numpy.dot(d[j], self.embeds[j][:,k])
self.W[i,:] -= (alpha * grad)
total = numpy.sum(self.W[i,:])
if total != 0:
self.W[i,:] = self.W[i,:] / total
write("\n")
end_time = time.clock()
write("Done (took %s seconds)\n" % str(end_time - start_time))
pass
def save_weights(self, fname):
"""
Save the weight matrix to a disk file.
"""
savemat(fname, {"W":self.W})
pass
def load_weights(self, fname):
"""
Load the weight matrix from a disk file.
"""
self.W = loadmat(fname)["W"]
pass
def test_compute_weights(self):
"""
Check whether the weights are computed correctly
"""
N = len(self.words)
# Check whether non-neighbours have weights equal to zero.
write("Checking whether non-neighbours have zero weights...\n")
for i in range(N):
pred_nns = set(numpy.where(self.W[i,:] != 0)[0])
nns = set()
for j in range(len(self.embeds)):
nns = nns.union(set(self.NNS[j][i,:]))
assert(pred_nns == nns)
# Check whether reconstruction weights add upto one.
write("Checking whether weights add to 1...\n")
for i in range(N):
assert(numpy.allclose(numpy.sum(self.W[i,:]), 1))
# print nearest neighbours and their weights
nn_file = open("../work/nn.csv", 'w')
for i in range(N):
nn_file.write("%s, " % self.words[i])
L = []
for j in range(N):
if self.W[i,j] != 0:
L.append((self.words[j], self.W[i,j]))
L.sort(lambda x, y: -1 if x[1] > y[1] else 1)
for (w, val) in L:
nn_file.write("%s, %f, " % (w, val))
nn_file.write("\n")
nn_file.close()
pass
def compute_M(self):
"""
Compute the smallest eigenvectors of M = (I - W')\T(I - W').
"""
# Building W'
N = len(self.words)
start_time = time.clock()
write("Computing W'...")
for i in range(N):
z = numpy.zeros(N)
write("Completed %d of %d\r" % (i, N))
for nns in self.NNS:
z[nns[i,:]] += 1
self.W[i,:] = z * self.W[i,:]
end_time = time.clock()
write("Done (took %s seconds)\n" % str(end_time - start_time))
# Computing M.
start_time = time.clock()
write("Computing M....")
self.W = csr_matrix(self.W)
M = eye(N, format=self.W.format) - self.W
M = (M.T * M).tocsr()
end_time = time.clock()
write("Done (took %s seconds)\n" % str(end_time - start_time))
return M
def compute_embeddings(self, k, M, embed_fname):
"""
Perform eigen decomposition.
"""
N = len(self.words)
start_time = time.clock()
write("Computing Eigen decomposition...")
s, V = eigsh(M, k+1, tol=1E-6, which="SA", maxiter=100)
end_time = time.clock()
write("Done (took %s seconds)\n" % str(end_time - start_time))
P = V[:, 1:]
err = numpy.sum(s[1:])
write("Projection error = %f\n" % err)
write("Writing embeddings to file...")
# Write embeddings to file.
with open(embed_fname, 'w') as embed_file:
for i in range(N):
embed_file.write("%s %s\n" % (self.words[i], " ".join([str(x) for x in P[i,:]])))
write("Done\n")
pass
def write(msg):
sys.stdout.write(msg)
sys.stdout.flush()
pass
def meta_embed(embeddings, words, nns, comps, embed_path):
"""
Perform meta-embedding using LLE.
"""
ME = MetaEmbed(embeddings, words)
ME.compute_neighbours(nns)
#ME.show_nns("king", 5)
#ME.compute_weights_parallel()
ME.compute_weights()
#ME.save_weights("../work/weights_%d" % nns)
#ME.load_weights("../work/weights+n=%d.meta" % nns)
#ME.test_compute_weights()
M = ME.compute_M()
for k in comps:
embed_fname = "%s/n=%d+k=%d" % (embed_path, nns, k)
write("Embedding NNS = %d, Components (k) = %d\n" % (nns, k))
try:
ME.compute_embeddings(k, M, embed_fname)
except ArpackNoConvergence as e:
print e
return ME
def baseline_concatenate(embeddings, words, embed_fname):
"""
Concatenate embeddings to create co-embeddings.
"""
dim = sum([x.dim for x in embeddings])
print "Concatenation dimension =", dim
# concatenate the vectors.
with open(embed_fname, 'w') as embed_file:
for (i,word) in enumerate(words):
L = []
for x in embeddings:
w = 8 if x.dim == 300 else 1
#w = 1
L.append(w * x.vects[word])
z = numpy.concatenate(L)
embed_file.write("%s %s\n" % (word, " ".join([str(x) for x in z])))
pass
def get_common_words(embeddings):
words = set(embeddings[0].vocab)
for i in range(1, len(embeddings)):
words = words.intersection(set(embeddings[i].vocab))
return words
def get_selected_words(fname):
words = []
with open(fname) as F:
for line in F:
words.append(line.strip())
return words
def perform_embedding(nns, comps):
print "Neigbourhood size = %d" % nns
#embed_sett
embed_settings = sources.embed_settings
embeddings = []
for (embd_fname, dim) in embed_settings:
start_time = time.clock()
sys.stdout.write("Loading %s -- (%d dim) ..." % (embd_fname, dim))
sys.stdout.flush()
WR = WordReps()
WR.read_model(embd_fname, dim)
end_time = time.clock()
sys.stdout.write("\nDone. took %s seconds\n" % str(end_time - start_time))
sys.stdout.flush()
embeddings.append(WR)
common_words = get_common_words(embeddings)
selected_words = get_selected_words("../work/selected-words")
words = []
for word in selected_words:
if word in common_words and word not in words:
words.append(word)
print "No. of common words =", len(common_words)
print "Vocabulary size =", len(words)
ME = meta_embed(embeddings, words, nns, comps, "../work/meta-embeds")
pass
def save_embedding(words, WR, fname):
F = open(fname, 'w')
for w in words:
if w in WR.vects:
F.write("%s " % w)
F.write("%s\n" % " ".join([str(x) for x in WR.vects[w]]))
# elif w.lower() in WR.vects:
# F.write("%s " % w.lower())
# F.write("%s\n" % " ".join([str(x) for x in WR.vects[w.lower()]]))
F.close()
pass
def main():
parser = argparse.ArgumentParser()
parser.add_argument("-nns", type=int, help="number of nearest neighbours")
parser.add_argument("-comps", type=str, help="components for the projection")
args = parser.parse_args()
comps = [int(x) for x in args.comps.split(',')]
perform_embedding(args.nns, comps)
pass
if __name__ == '__main__':
main()
pass