RE WORK - Boston Deep Learning Summit

I have recently attended the Deep Learning Summit in Boston. The event was organized by RE WORK. RE WORK was founded in London. The team is all women. The mission of the RE WORK team is to encourage conversations around entrepreneurship, technology, and science to shape the future. 

This is a quick recount of the event from my perspective.
First of all, I have never been to Boston. The public transportation that I took from the airport to the place of the conference was really easy to navigate (In short, I did not get lost). This is probably a result of the effort put in by the local government to make Boston a premier conference venue. Traffic congestion is another story.

Schedule of Talks
The conference schedule is packed.  The speakers are researchers from some of the top tech companies. Facebook, Google, Amazon, Ebay, and Spotify are all represented. I was excited about two topics in the schedule. Here are some of the papers presented. The papers I chose below are some of the research which I thought was significant to some of the efforts we are doing for our business partners:

      The end product of this research is the MovieQA dataset which contains several sources of information  – video clips, subtitles, scripts, plots, and DVS. One of the representations that were used to build the MovieQA data set is SkipThoughts. SkipThoughts uses a recurrent neural network to capture the underlying sentence semantics. The SkipThought model [2] was inspired by word vector learning using the skip-gram model. Instead of just using words to predicting the context around it, SkipThought encodes sentences to predict sentences around it. The model is based on an encoder-decoder architecture. The encoder maps the English sentence into a vector. The decoder then conditions on the encoded vector to generate a translation for the source English sentence.

Domain Separation Networks lets you adapt your model to a target dataset which you did not train on. Domain adaptation's goal is to train on a "source" dataset then apply the model to a "target" data set generated from a different distribution [3]. The motivations behind domain adaptation are the following: 1. Create high-quality datasets for training is costly and time-consuming. 2. Certain annotation tasks are impossible to scale. Domain adaptation allows you to train on synthetic data then apply the model to the real domain.

References
[1] Distributed representations of words and phrases and their Compositionality, https://arxiv.org/abs/1310.4546
[2] Skip-Thought Vectors, http://arxiv.org/abs/1506.06726
[3] Domain Separation Networks, https://arxiv.org/abs/1608.06019 


Comments

Popular posts from this blog

OAuth 1.0a Request Signing and Verification - HMAC-SHA1 - HMAC-SHA256

Spark DataFrame - Array[ByteBuffer] - IllegalAurmentException

Gensim Doc2Vec on Spark - a quest to get the right Vector