Deep Learning: What's hot, what's hype and what can we use?
DL is a subset of machine learning.
DL is not a SVM, random forest, single layer HMM, ...
DL is defined by multiple stacks of layers (deep)
DL is exploding in terms of publications and use-cases.
It works ... image recognition, captioning, robotic control, text synthesis, language generation, audio encoding, ...
Major investment by Google, Facebook, Twitter, Intel, Baidu, Amazon, Adobe, Oracle, IBM Watson, ...
Trained on text, RNN's can reproduce passages, punctuation and style
PANDARUS:
Alas, I think he shall be come approached and the day
When little srain would be attain'd into being never fed,
And who is but a chain and subjects of his death,
I should not sleep.
Second Senator:
They are away this miseries, produced upon my soul,
Breaking and strongly should be buried, when I perish
The earth and thoughts of many states.
Or titles of scientific articles (high-energy physics):
Search for CP-Violation in Right-Handed Neutrinos
Neutron Star Propagation as a Source of the Hadron-Hadron Scattering Measurement at the Planck Scale
Goldstone bosons and scattering modes from two-dimensional Hamiltonians
Baryogenesis with Yukawa Unified SUSY GUTs
Radial scattering and dense quark matter and chiral phase transition of a pseudoscalar meson at finite baryon density
Transition Form Factor Constant in a Quark-Diquark Model
A New Theory of Supersymmetry Breaking
Effective gluon propagator in QCD at zero and finite temperature and and dual gauge QCD in the BCS-BCS limit
Light scalar pair signals in top quark multiple polarizatino
Not covered here
More topics not covered (but very interesting!)
BlackOut: Speeding up Recurrent Neural Network Language Models With Very Large Vocabularies
Guaranteed Non-convex Learning Algorithms through Tensor Factorization
Convergent Learning: Do different neural networks learn the same representations?
Built a corpus of children's books with attention based questions
Since humans need context to solve Q&A, train attention mechanisms!
An example of a positive TE (text entails hypothesis):
text: If you help the needy, God will reward you.
hypothesis: Giving money to a poor man has good consequences.
An example of a negative TE (text contradicts hypothesis):
text: If you help the needy, God will reward you.
hypothesis: Giving money to a poor man has no consequences.
An example of a non-TE (text does not entail nor contradict):
text: If you help the needy, God will reward you.
hypothesis: Giving money to a poor man will make you a better person.
In Electronic Health Records the visit sequences of patients have multiple concepts (diagnosis, procedure, and medication codes) per visit. This structure provides two types of relational information, namely sequential order of visits and co-occurrence of the codes within each visit. Med2Vec learns distributed representations for both medical codes and visits from a large EHR dataset with over 3 million visits, and allows us to interpret the learned representations confirmed positively by clinical experts.
Thank you