Classification of "BBC News" and comparison of performance between 3 types of model's architectures. Then 2D word embedding visualization using PCA and 3D word embedding visualisation using T-SNE
We have implemented, expanded and reviewed the paper “Sense2Vec - A Fast and Accurate Method For Word Sense Disambiguation In Neural Word Embeddings" by Andrew Trask, Phil Michalak and John Liu.
This a part of Kaggle Competion, Toxic Comment Classification Challenge by Jigsaw .This was a multilabel classification challenge.This code is a improved version of my submission in the competion.
A word embedding is a learned representation for text where words that have the same meaning have a similar representation.Word embeddings are in fact a class of techniques where individual words are represented as real-valued vectors in a predefined vector space. Each word is mapped to one vector and the vector values are learned in a way that resembles a neural network