Overall, it is going pretty good. Many very good papers, diarization joins with decoding, everything goes to the right direction. RadioTalk: a large-scale corpus of talk radio transcripts Doug Beeferman...
It is interesting how similar ideas raise here and there in seemingly unrelated context. The recent quote from Actionable Book Summary: The Inevitable by Kevin Kelly And what’s next probably...
An important issue with a modern neural networks is their vulnerability to the masked corruption, that is the random corruption of some small amount of samples in the image or...
I've got quite interested in the future prediction these days, one nice idea by Russian writer Sergey Borisovich Pereslegin is that we should build the future based on the theory...
Dear friends, as you know Google+ is shutting down. I considered several alternatives: Facebook, Quora, Linkedin, my old blog, Reddit, Twitter, Telegram. Unfortunately there are things I dislike in all...
In speech recognition we frequently deal with noisy or simply corrupted recordings. For example, in call center recordings you still get error rates like 50% or 60% even with the...
Recently a set of papers were published about "memorization" in neural networks. For example:Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts LayeralsoUnderstanding deep learning requires rethinking generalizationIt seems that large memory...
It seems that people will loose the ability to read, comprehend and remember long texts soon, the question now is - is it possible to deliver very complex messages without...
A recent discussion on kaldi group about OOV words reminded me about this old problem.One of the things that makes modern recognizers so unnatural is probabilistic models behind them. It's...
IWSLT 2015 proceedings recently appeared. This is an important competition in ASR focused on TED talks translation (and, more interesting for us, transcription).Best system from MITLL-AFRL had a nice WER...