Full Bayesian Hidden Markov Model Variational Autoencoder for Acoustic Unit Discovery

Author(s):  
Thomas Glarner ◽  
Patrick Hanebrink ◽  
Janek Ebbers ◽  
Reinhold Haeb-Umbach
Author(s):  
Natsuki Iwano ◽  
Tatsuo Adachi ◽  
Kazuteru Aoki ◽  
Yoshikazu Nakamura ◽  
Michiaki Hamada

AbstractNucleic acid aptamers are generated by an in vitro molecular evolution method known as systematic evolution of ligands by exponential enrichment (SELEX). A variety of candidates is limited by actual sequencing data from an experiment. Here, we developed RaptGen, which is a variational autoencoder for in silico aptamer generation. RaptGen exploits a profile hidden Markov model decoder to represent motif sequences effectively. We showed that RaptGen embedded simulation sequence data into low-dimension latent space dependent on motif information. We also performed sequence embedding using two independent SELEX datasets. RaptGen successfully generated aptamers from the latent space even though they were not included in high-throughput sequencing. RaptGen could also generate a truncated aptamer with a short learning model. We demonstrated that RaptGen could be applied to activity-guided aptamer generation according to Bayesian optimization. We concluded that a generative method by RaptGen and latent representation are useful for aptamer discovery. Codes are available at https://github.com/hmdlab/raptgen.


Author(s):  
Janek Ebbers ◽  
Jahn Heymann ◽  
Lukas Drude ◽  
Thomas Glarner ◽  
Reinhold Haeb-Umbach ◽  
...  

Entropy ◽  
2021 ◽  
Vol 23 (10) ◽  
pp. 1277
Author(s):  
Kun Zhao ◽  
Hongwei Ding ◽  
Kai Ye ◽  
Xiaohui Cui

The Variational AutoEncoder (VAE) has made significant progress in text generation, but it focused on short text (always a sentence). Long texts consist of multiple sentences. There is a particular relationship between each sentence, especially between the latent variables that control the generation of the sentences. The relationships between these latent variables help in generating continuous and logically connected long texts. There exist very few studies on the relationships between these latent variables. We proposed a method for combining the Transformer-Based Hierarchical Variational Autoencoder and Hidden Markov Model (HT-HVAE) to learn multiple hierarchical latent variables and their relationships. This application improves long text generation. We use a hierarchical Transformer encoder to encode the long texts in order to obtain better hierarchical information of the long text. HT-HVAE’s generation network uses HMM to learn the relationship between latent variables. We also proposed a method for calculating the perplexity for the multiple hierarchical latent variable structure. The experimental results show that our model is more effective in the dataset with strong logic, alleviates the notorious posterior collapse problem, and generates more continuous and logically connected long text.


2012 ◽  
Vol 132 (10) ◽  
pp. 1589-1594 ◽  
Author(s):  
Hayato Waki ◽  
Yutaka Suzuki ◽  
Osamu Sakata ◽  
Mizuya Fukasawa ◽  
Hatsuhiro Kato

MIS Quarterly ◽  
2018 ◽  
Vol 42 (1) ◽  
pp. 83-100 ◽  
Author(s):  
Wei Chen ◽  
◽  
Xiahua Wei ◽  
Kevin Xiaoguo Zhu ◽  
◽  
...  

Sign in / Sign up

Export Citation Format

Share Document