default search action
Kazuyoshi Yoshii
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c138]Diego Di Carlo, Aditya Arie Nugraha, Mathieu Fontaine, Yoshiaki Bando, Kazuyoshi Yoshii:
Neural Steerer: Novel Steering Vector Synthesis with a Causal Neural Field over Frequency and Direction. ICASSP Workshops 2024: 740-744 - [c137]Yoshiaki Sumura, Diego Di Carlo, Aditya Arie Nugraha, Yoshiaki Bando, Kazuyoshi Yoshii:
Joint Audio Source Localization and Separation with Distributed Microphone Arrays Based on Spatially-Regularized Multichannel NMF. IWAENC 2024: 145-149 - [i27]Yoto Fujita, Yoshiaki Bando, Keisuke Imoto, Masaki Onishi, Kazuyoshi Yoshii:
DOA-Aware Audio-Visual Self-Supervised Learning for Sound Event Localization and Detection. CoRR abs/2410.22803 (2024) - [i26]Yoto Fujita, Aditya Arie Nugraha, Diego Di Carlo, Yoshiaki Bando, Mathieu Fontaine, Kazuyoshi Yoshii:
Run-Time Adaptation of Neural Beamforming for Robust Speech Dereverberation and Denoising. CoRR abs/2410.22805 (2024) - 2023
- [c136]Daichi Kamakura, Eita Nanamura, Takehisa Oyama, Kazuyoshi Yoshii:
Joint Drum Transcription and Metrical Analysis Based on Periodicity-Aware Multi-Task Learning. APSIPA ASC 2023: 151-157 - [c135]Daichi Kamakura, Eita Nakamura, Kazuyoshi Yoshii:
CTC2: End-to-End Drum Transcription Based on Connectionist Temporal Classification With Constant Tempo Constraint. APSIPA ASC 2023: 158-164 - [c134]Tsung-Ping Chen, Li Su, Kazuyoshi Yoshii:
Learning Multifaceted Self-Similarity for Musical Structure Analysis. APSIPA ASC 2023: 165-172 - [c133]Tengyu Deng, Eita Nakamura, Kazuyoshi Yoshii:
Audio-to-Score Singing Transcription Based on Joint Estimation of Pitches, Onsets, and Metrical Positions With Tatum-Level CTC Loss. APSIPA ASC 2023: 583-590 - [c132]Jiahao Zhao, Kazuyoshi Yoshii:
Multimodal Multifaceted Music Emotion Recognition Based on Self-Attentive Fusion of Psychology-Inspired Symbolic and Acoustic Features. APSIPA ASC 2023: 1641-1645 - [c131]Yoto Fujita, Yoshiaki Bando, Keisuke Imoto, Masaki Onishi, Kazuyoshi Yoshii:
DOA-Aware Audio-Visual Self-Supervised Learning for Sound Event Localization and Detection. APSIPA ASC 2023: 2061-2067 - [c130]Yoshiaki Bando, Yoshiki Masuyama, Aditya Arie Nugraha, Kazuyoshi Yoshii:
Neural Fast Full-Rank Spatial Covariance Analysis for Blind Source Separation. EUSIPCO 2023: 51-55 - [c129]Moyu Terao, Eita Nakamura, Kazuyoshi Yoshii:
Neural Band-to-Piano Score Arrangement with Stepless Difficulty Control. ICASSP 2023: 1-5 - [c128]Aditya Arie Nugraha, Diego Di Carlo, Yoshiaki Bando, Mathieu Fontaine, Kazuyoshi Yoshii:
Time-Domain Audio Source Separation Based on Gaussian Processes with Deep Kernel Learning. WASPAA 2023: 1-5 - [i25]Diego Di Carlo, Aditya Arie Nugraha, Mathieu Fontaine, Yoshiaki Bando, Kazuyoshi Yoshii:
Neural Steerer: Novel Steering Vector Synthesis with a Causal Neural Field over Frequency and Source Positions. CoRR abs/2305.04447 (2023) - [i24]Yoshiaki Bando, Yoshiki Masuyama, Aditya Arie Nugraha, Kazuyoshi Yoshii:
Neural Fast Full-Rank Spatial Covariance Analysis for Blind Source Separation. CoRR abs/2306.10240 (2023) - 2022
- [j33]Yicheng Du, Robin Scheibler, Masahito Togami, Kazuyoshi Yoshii, Tatsuya Kawahara:
Computationally-Efficient Overdetermined Blind Source Separation Based on Iterative Source Steering. IEEE Signal Process. Lett. 29: 927-931 (2022) - [j32]Mathieu Fontaine, Kouhei Sekiguchi, Aditya Arie Nugraha, Yoshiaki Bando, Kazuyoshi Yoshii:
Generalized Fast Multichannel Nonnegative Matrix Factorization Based on Gaussian Scale Mixtures for Blind Source Separation. IEEE ACM Trans. Audio Speech Lang. Process. 30: 1734-1748 (2022) - [j31]Kouhei Sekiguchi, Yoshiaki Bando, Aditya Arie Nugraha, Mathieu Fontaine, Kazuyoshi Yoshii, Tatsuya Kawahara:
Autoregressive Moving Average Jointly-Diagonalizable Spatial Covariance Analysis for Joint Source Separation and Dereverberation. IEEE ACM Trans. Audio Speech Lang. Process. 30: 2368-2382 (2022) - [c127]Mathieu Fontaine, Diego Di Carlo, Kouhei Sekiguchi, Aditya Arie Nugraha, Yoshiaki Bando, Kazuyoshi Yoshii:
Elliptically Contoured Alpha-Stable Representation for MUSIC-Based Sound Source Localization. EUSIPCO 2022: 26-30 - [c126]Moyu Terao, Yuki Hiramatsu, Ryoto Ishizuka, Yiming Wu, Kazuyoshi Yoshii:
Difficulty-Aware Neural Band-to-Piano Score Arrangement based on Note- and Statistic-Level Criteria. ICASSP 2022: 196-200 - [c125]Aditya Arie Nugraha, Kouhei Sekiguchi, Mathieu Fontaine, Yoshiaki Bando, Kazuyoshi Yoshii:
Flow-Based Fast Multichannel Nonnegative Matrix Factorization for Blind Source Separation. ICASSP 2022: 501-505 - [c124]Yicheng Du, Aditya Arie Nugraha, Kouhei Sekiguchi, Yoshiaki Bando, Mathieu Fontaine, Kazuyoshi Yoshii:
Direction-Aware Joint Adaptation of Neural Speech Enhancement and Recognition in Real Multiparty Conversational Environments. INTERSPEECH 2022: 2918-2922 - [c123]Kouhei Sekiguchi, Aditya Arie Nugraha, Yicheng Du, Yoshiaki Bando, Mathieu Fontaine, Kazuyoshi Yoshii:
Direction-Aware Adaptive Online Neural Speech Enhancement with an Augmented Reality Headset in Real Noisy Conversational Environments. IROS 2022: 9266-9273 - [c122]Tengyu Deng, Eita Nakamura, Kazuyoshi Yoshii:
End-to-End Lyrics Transcription Informed by Pitch and Onset Estimation. ISMIR 2022: 633-639 - [c121]Florian Thalmann, Eita Nakamura, Kazuyoshi Yoshii:
Tracking the Evolution of a Band's Live Performances over Decades. ISMIR 2022: 850-857 - [c120]Aditya Arie Nugraha, Kouhei Sekiguchi, Mathieu Fontaine, Yoshiaki Bando, Kazuyoshi Yoshii:
DNN-free Low-Latency Adaptive Speech Enhancement Based on Frame-Online Beamforming Powered by Block-Online Fastmnmf. IWAENC 2022: 1-5 - [c119]Yoshiaki Sumura, Kouhei Sekiguchi, Yoshiaki Bando, Aditya Arie Nugraha, Kazuyoshi Yoshii:
Joint Localization and Synchronization of Distributed Camera-Attached Microphone Arrays for Indoor Scene Analysis. IWAENC 2022: 1-5 - [i23]Mathieu Fontaine, Kouhei Sekiguchi, Aditya Arie Nugraha, Yoshiaki Bando, Kazuyoshi Yoshii:
Generalized Fast Multichannel Nonnegative Matrix Factorization Based on Gaussian Scale Mixtures for Blind Source Separation. CoRR abs/2205.05330 (2022) - [i22]Yicheng Du, Aditya Arie Nugraha, Kouhei Sekiguchi, Yoshiaki Bando, Mathieu Fontaine, Kazuyoshi Yoshii:
Direction-Aware Joint Adaptation of Neural Speech Enhancement and Recognition in Real Multiparty Conversational Environments. CoRR abs/2207.07273 (2022) - [i21]Kouhei Sekiguchi, Aditya Arie Nugraha, Yicheng Du, Yoshiaki Bando, Mathieu Fontaine, Kazuyoshi Yoshii:
Direction-Aware Adaptive Online Neural Speech Enhancement with an Augmented Reality Headset in Real Noisy Conversational Environments. CoRR abs/2207.07296 (2022) - [i20]Aditya Arie Nugraha, Kouhei Sekiguchi, Mathieu Fontaine, Yoshiaki Bando, Kazuyoshi Yoshii:
DNN-Free Low-Latency Adaptive Speech Enhancement Based on Frame-Online Beamforming Powered by Block-Online FastMNMF. CoRR abs/2207.10934 (2022) - 2021
- [j30]Kentaro Shibata, Eita Nakamura, Kazuyoshi Yoshii:
Non-local musical statistics as guides for audio-to-score piano transcription. Inf. Sci. 566: 262-280 (2021) - [j29]Eita Nakamura, Kazuyoshi Yoshii:
Musical rhythm transcription based on Bayesian piece-specific score models capturing repetitions. Inf. Sci. 572: 482-500 (2021) - [j28]Takayuki Nakatsuka, Kazuyoshi Yoshii, Yuki Koyama, Satoru Fukayama, Masataka Goto, Shigeo Morishima:
MirrorNet: A Deep Reflective Approach to 2D Pose Estimation for Single-Person Images. J. Inf. Process. 29: 406-423 (2021) - [j27]Yoshiaki Bando, Kouhei Sekiguchi, Yoshiki Masuyama, Aditya Arie Nugraha, Mathieu Fontaine, Kazuyoshi Yoshii:
Neural Full-Rank Spatial Covariance Analysis for Blind Source Separation. IEEE Signal Process. Lett. 28: 1670-1674 (2021) - [c118]Yoshiaki Bando, Kouhei Sekiguchi, Kazuyoshi Yoshii:
Gamma Process FastMNMF for Separating an Unknown Number of Sound Sources. EUSIPCO 2021: 291-295 - [c117]Keitaro Tanaka, Ryo Nishikimi, Yoshiaki Bando, Kazuyoshi Yoshii, Shigeo Morishima:
Pitch-Timbre Disentanglement Of Musical Instrument Sounds Based On Vae-Based Metric Learning. ICASSP 2021: 111-115 - [c116]Yuki Hiramatsu, Go Shibata, Ryo Nishikimi, Eita Nakamura, Kazuyoshi Yoshii:
Statistical Correction of Transcribed Melody Notes Based on Probabilistic Integration of a Music Language Model and a Transcription Error Model. ICASSP 2021: 256-260 - [c115]Kouhei Sekiguchi, Yoshiaki Bando, Aditya Arie Nugraha, Mathieu Fontaine, Kazuyoshi Yoshii:
Autoregressive Fast Multichannel Nonnegative Matrix Factorization For Joint Blind Source Separation And Dereverberation. ICASSP 2021: 511-515 - [c114]Mathieu Fontaine, Kouhei Sekiguchi, Aditya Arie Nugraha, Yoshiaki Bando, Kazuyoshi Yoshii:
Alpha-Stable Autoregressive Fast Multichannel Nonnegative Matrix Factorization for Joint Speech Enhancement and Dereverberation. Interspeech 2021: 661-665 - [c113]Yuki Hiramatsu, Eita Nakamura, Kazuyoshi Yoshii:
Joint Estimation of Note Values and Voices for Audio-to-Score Piano Transcription. ISMIR 2021: 278-284 - [c112]Takehisa Oyama, Ryoto Ishizuka, Kazuyoshi Yoshii:
Phase-Aware Joint Beat and Downbeat Estimation Based on Periodicity of Metrical Structure. ISMIR 2021: 493-499 - [c111]Mitsuki Hosoya, Masanori Morise, Satoshi Nakamura, Kazuyoshi Yoshii:
A Real-Time Drum-Wise Volume Visualization System for Learning Volume-Balanced Drum Performance. ICEC 2021: 154-166 - [i19]Ryoto Ishizuka, Ryo Nishikimi, Kazuyoshi Yoshii:
Global Structure-Aware Drum Transcription Based on Self-Attention Mechanisms. CoRR abs/2105.05791 (2021) - 2020
- [j26]Eita Nakamura, Yasuyuki Saito, Kazuyoshi Yoshii:
Statistical learning and estimation of piano fingering. Inf. Sci. 517: 68-85 (2020) - [j25]Aditya Arie Nugraha, Kouhei Sekiguchi, Mathieu Fontaine, Yoshiaki Bando, Kazuyoshi Yoshii:
Flow-Based Independent Vector Analysis for Blind Source Separation. IEEE Signal Process. Lett. 27: 2173-2177 (2020) - [j24]Aditya Arie Nugraha, Kouhei Sekiguchi, Kazuyoshi Yoshii:
A Flow-Based Deep Latent Variable Model for Speech Spectrogram Modeling and Enhancement. IEEE ACM Trans. Audio Speech Lang. Process. 28: 1104-1117 (2020) - [j23]Hiroaki Tsushima, Eita Nakamura, Kazuyoshi Yoshii:
Bayesian Melody Harmonization Based on a Tree-Structured Generative Model of Chord Sequences and Melodies. IEEE ACM Trans. Audio Speech Lang. Process. 28: 1644-1655 (2020) - [j22]Ryo Nishikimi, Eita Nakamura, Masataka Goto, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Bayesian Singing Transcription Based on a Hierarchical Generative Model of Keys, Musical Notes, and F0 Trajectories. IEEE ACM Trans. Audio Speech Lang. Process. 28: 1678-1691 (2020) - [j21]Kouhei Sekiguchi, Yoshiaki Bando, Aditya Arie Nugraha, Kazuyoshi Yoshii, Tatsuya Kawahara:
Fast Multichannel Nonnegative Matrix Factorization With Directivity-Aware Jointly-Diagonalizable Spatial Covariance Matrices for Blind Source Separation. IEEE ACM Trans. Audio Speech Lang. Process. 28: 2610-2625 (2020) - [j20]Yiming Wu, Tristan Carsault, Eita Nakamura, Kazuyoshi Yoshii:
Semi-Supervised Neural Chord Estimation Based on a Variational Autoencoder With Latent Chord Labels and Features. IEEE ACM Trans. Audio Speech Lang. Process. 28: 2956-2966 (2020) - [c110]Ryoto Ishizuka, Ryo Nishikimi, Eita Nakamura, Kazuyoshi Yoshii:
Tatum-Level Drum Transcription Based on a Convolutional Recurrent Neural Network with Language Model-Based Regularized Training. APSIPA 2020: 359-364 - [c109]Yiming Wu, Eita Nakamura, Kazuyoshi Yoshii:
A Variational Autoencoder for Joint Chord and Key Estimation from Audio Chromagrams. APSIPA 2020: 500-506 - [c108]Masaya Wake, Masahito Togami, Kazuyoshi Yoshii, Tatsuya Kawahara:
Integration of Semi-Blind Speech Source Separation and Voice Activity Detection for Flexible Spoken Dialogue. APSIPA 2020: 775-780 - [c107]Masahito Togami, Yoshiki Masuyama, Tatsuya Komatsu, Kazuyoshi Yoshii, Tatsuya Kawahara:
Computer-Resource-Aware Deep Speech Separation with a Run-Time-Specified Number of BLSTM Layers. APSIPA 2020: 788-793 - [c106]Jeongwoo Woo, Masato Mimura, Kazuyoshi Yoshii, Tatsuya Kawahara:
End-to-end Music-mixed Speech Recognition. APSIPA 2020: 800-804 - [c105]Kazuyoshi Yoshii, Kouhei Sekiguchi, Yoshiaki Bando, Mathieu Fontaine, Aditya Arie Nugraha:
Fast Multichannel Correlated Tensor Factorization for Blind Source Separation. EUSIPCO 2020: 306-310 - [c104]Yicheng Du, Kouhei Sekiguchi, Yoshiaki Bando, Aditya Arie Nugraha, Mathieu Fontaine, Kazuyoshi Yoshii, Tatsuya Kawahara:
Semi-supervised Multichannel Speech Separation Based on a Phone- and Speaker-Aware Deep Generative Model of Speech Spectrograms. EUSIPCO 2020: 870-874 - [c103]Yoshiaki Bando, Kouhei Sekiguchi, Kazuyoshi Yoshii:
Adaptive Neural Speech Enhancement with a Denoising Variational Autoencoder. INTERSPEECH 2020: 2437-2441 - [c102]Mathieu Fontaine, Kouhei Sekiguchi, Aditya Arie Nugraha, Kazuyoshi Yoshii:
Unsupervised Robust Speech Enhancement Based on Alpha-Stable Fast Multichannel Nonnegative Matrix Factorization. INTERSPEECH 2020: 4541-4545 - [c101]Go Shibata, Ryo Nishikimi, Kazuyoshi Yoshii:
Music Structure Analysis Based on an LSTM-HSMM Hybrid Model. ISMIR 2020: 23-29 - [c100]Keitaro Tanaka, Takayuki Nakatsuka, Ryo Nishikimi, Kazuyoshi Yoshii, Shigeo Morishima:
Multi-Instrument Music Transcription Based on Deep Spherical Clustering of Spectrograms and Pitchgrams. ISMIR 2020: 327-334 - [c99]Florian Thalmann, Kazuyoshi Yoshii, Thomas Wilmering, Geraint A. Wiggins, Mark B. Sandler:
A Method for Analysis of Shared Structure in Large Music Collections using Techniques from Genetic Sequencing and Graph Theory. ISMIR 2020: 343-350 - [c98]Andrew McLeod, James Owers, Kazuyoshi Yoshii:
The MIDI Degradation Toolkit: Symbolic Music Augmentation and Correction. ISMIR 2020: 846-852 - [i18]Takayuki Nakatsuka, Kazuyoshi Yoshii, Yuki Koyama, Satoru Fukayama, Masataka Goto, Shigeo Morishima:
MirrorNet: A Deep Bayesian Approach to Reflective 2D Pose Estimation from Human Images. CoRR abs/2004.03811 (2020) - [i17]Yiming Wu, Tristan Carsault, Eita Nakamura, Kazuyoshi Yoshii:
Semi-supervised Neural Chord Estimation Based on a Variational Autoencoder with Discrete Labels and Continuous Textures of Chords. CoRR abs/2005.07091 (2020) - [i16]Kentaro Shibata, Eita Nakamura, Kazuyoshi Yoshii:
Non-Local Musical Statistics as Guides for Audio-to-Score Piano Transcription. CoRR abs/2008.12710 (2020) - [i15]Andrew McLeod, James Owers, Kazuyoshi Yoshii:
The MIDI Degradation Toolkit: Symbolic Music Augmentation and Correction. CoRR abs/2010.00059 (2020) - [i14]Ryoto Ishizuka, Ryo Nishikimi, Eita Nakamura, Kazuyoshi Yoshii:
Tatum-Level Drum Transcription Based on a Convolutional Recurrent Neural Network with Language Model-Based Regularized Training. CoRR abs/2010.03749 (2020)
2010 – 2019
- 2019
- [j19]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Unsupervised Speech Enhancement Based on Multichannel NMF-Informed Beamforming for Noise-Robust Automatic Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 27(5): 960-971 (2019) - [j18]Kouhei Sekiguchi, Yoshiaki Bando, Aditya Arie Nugraha, Kazuyoshi Yoshii, Tatsuya Kawahara:
Semi-Supervised Multichannel Speech Enhancement With a Deep Speech Prior. IEEE ACM Trans. Audio Speech Lang. Process. 27(12): 2197-2212 (2019) - [c97]Mathieu Fontaine, Aditya Arie Nugraha, Roland Badeau, Kazuyoshi Yoshii, Antoine Liutkus:
Cauchy Multichannel Speech Enhancement with a Deep Speech Prior. EUSIPCO 2019: 1-5 - [c96]Kouhei Sekiguchi, Aditya Arie Nugraha, Yoshiaki Bando, Kazuyoshi Yoshii:
Fast Multichannel Source Separation Based on Jointly Diagonalizable Spatial Covariance Matrices. EUSIPCO 2019: 1-5 - [c95]Yiming Wu, Tristan Carsault, Kazuyoshi Yoshii:
Automatic Chord Estimation Based on a Frame-wise Convolutional Recurrent Neural Network with Non-Aligned Annotations. EUSIPCO 2019: 1-5 - [c94]Ryo Nishikimi, Eita Nakamura, Satoru Fukayama, Masataka Goto, Kazuyoshi Yoshii:
Automatic Singing Transcription Based on Encoder-decoder Recurrent Neural Networks with a Weakly-supervised Attention Mechanism. ICASSP 2019: 161-165 - [c93]Andrew McLeod, Eita Nakamura, Kazuyoshi Yoshii:
Improved Metrical Alignment of Midi Performance Based on a Repetition-aware Online-adapted Grammar. ICASSP 2019: 186-190 - [c92]Eita Nakamura, Kentaro Shibata, Ryo Nishikimi, Kazuyoshi Yoshii:
Unsupervised Melody Style Conversion. ICASSP 2019: 196-200 - [c91]Kentaro Shibata, Ryo Nishikimi, Satoru Fukayama, Masataka Goto, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Joint Transcription of Lead, Bass, and Rhythm Guitars Based on a Factorial Hidden Semi-Markov Model. ICASSP 2019: 236-240 - [c90]Shun Ueda, Kentaro Shibata, Yusuke Wada, Ryo Nishikimi, Eita Nakamura, Kazuyoshi Yoshii:
Bayesian Drum Transcription Based on Nonnegative Matrix Factor Decomposition with a Deep Score Prior. ICASSP 2019: 456-460 - [c89]Aditya Arie Nugraha, Kouhei Sekiguchi, Kazuyoshi Yoshii:
A Deep Generative Model of Speech Complex Spectrograms. ICASSP 2019: 905-909 - [c88]Go Shibata, Ryo Nishikimi, Eita Nakamura, Kazuyoshi Yoshii:
Statistical Music Structure Analysis Based on a Homogeneity-, Repetitiveness-, and Regularity-Aware Hierarchical Hidden Semi-Markov Model. ISMIR 2019: 268-275 - [c87]Adrien Ycart, Andrew McLeod, Emmanouil Benetos, Kazuyoshi Yoshii:
Blending Acoustic and Language Model Predictions for Automatic Music Transcription. ISMIR 2019: 454-461 - [c86]Yoshiaki Bando, Yoko Sasaki, Kazuyoshi Yoshii:
Deep Bayesian Unsupervised Source Separation Based On A Complex Gaussian Mixture Model. MLSP 2019: 1-6 - [c85]Tristan Carsault, Andrew McLeod, Philippe Esling, Jérôme Nika, Eita Nakamura, Kazuyoshi Yoshii:
Multi-Step Chord Sequence Prediction Based On Aggregated Multi-Scale Encoder-Decoder Networks. MLSP 2019: 1-6 - [c84]Aaron Chau, Kouhei Sekiguchi, Aditya Arie Nugraha, Kazuyoshi Yoshii, Kotaro Funakoshi:
Audio-Visual SLAM towards Human Tracking and Human-Robot Interaction in Indoor Environments. RO-MAN 2019: 1-8 - [c83]Ryo Nishikimi, Eita Nakamura, Masataka Goto, Kazuyoshi Yoshii:
End-To-End Melody Note Transcription Based on a Beat-Synchronous Attention Mechanism. WASPAA 2019: 26-30 - [c82]Tomoyasu Nakano, Kazuyoshi Yoshii, Yiming Wu, Ryo Nishikimi, Kin Wah Edward Lin, Masataka Goto:
Joint Singing Pitch Estimation and Voice Separation Based on a Neural Harmonic Structure Renderer. WASPAA 2019: 160-164 - [i13]Kouhei Sekiguchi, Aditya Arie Nugraha, Yoshiaki Bando, Kazuyoshi Yoshii:
Fast Multichannel Source Separation Based on Jointly Diagonalizable Spatial Covariance Matrices. CoRR abs/1903.03237 (2019) - [i12]Aditya Arie Nugraha, Kouhei Sekiguchi, Kazuyoshi Yoshii:
A Deep Generative Model of Speech Complex Spectrograms. CoRR abs/1903.03269 (2019) - [i11]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Unsupervised Speech Enhancement Based on Multichannel NMF-Informed Beamforming for Noise-Robust Automatic Speech Recognition. CoRR abs/1903.09341 (2019) - [i10]Eita Nakamura, Yasuyuki Saito, Kazuyoshi Yoshii:
Statistical Learning and Estimation of Piano Fingering. CoRR abs/1904.10237 (2019) - [i9]Eita Nakamura, Kazuyoshi Yoshii:
Music Transcription Based on Bayesian Piece-Specific Score Models Capturing Repetitions. CoRR abs/1908.06969 (2019) - [i8]Yoshiaki Bando, Yoko Sasaki, Kazuyoshi Yoshii:
Deep Bayesian Unsupervised Source Separation Based on a Complex Gaussian Mixture Model. CoRR abs/1908.11307 (2019) - [i7]Tristan Carsault, Andrew McLeod, Philippe Esling, Jérôme Nika, Eita Nakamura, Kazuyoshi Yoshii:
Multi-Step Chord Sequence Prediction Based on Aggregated Multi-Scale Encoder-Decoder Network. CoRR abs/1911.04972 (2019) - 2018
- [j17]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Tatsuya Kawahara, Hiroshi G. Okuno:
Speech Enhancement Based on Bayesian Low-Rank and Sparse Decomposition of Multichannel Magnitude Spectrograms. IEEE ACM Trans. Audio Speech Lang. Process. 26(2): 215-230 (2018) - [j16]Kousuke Itakura, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Bayesian Multichannel Audio Source Separation Based on Integrated Source and Spatial Models. IEEE ACM Trans. Audio Speech Lang. Process. 26(4): 831-846 (2018) - [c81]Yusuke Wada, Ryo Nishikimi, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Sequential Generation of Singing F0 Contours from Musical Note Sequences Based on WaveNet. APSIPA 2018: 983-989 - [c80]Kouhei Sekiguchi, Yoshiaki Bando, Kazuyoshi Yoshii, Tatsuya Kawahara:
Bayesian Multichannel Speech Enhancement with a Deep Speech Prior. APSIPA 2018: 1233-1239 - [c79]Eita Nakamura, Ryo Nishikimi, Simon Dixon, Kazuyoshi Yoshii:
Probabilistic Sequential Patterns for Singing Transcription. APSIPA 2018: 1905-1912 - [c78]Kazuyoshi Yoshii, Koichi Kitamura, Yoshiaki Bando, Eita Nakamura, Tatsuya Kawahara:
Independent Low-Rank Tensor Analysis for Audio Source Separation. EUSIPCO 2018: 1657-1661 - [c77]Eita Nakamura, Emmanouil Benetos, Kazuyoshi Yoshii, Simon Dixon:
Towards Complete Polyphonic Music Transcription: Integrating Multi-Pitch Detection and Rhythm Quantization. ICASSP 2018: 101-105 - [c76]Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Statistical Speech Enhancement Based on Probabilistic Integration of Variational Autoencoder and Non-Negative Matrix Factorization. ICASSP 2018: 716-720 - [c75]Kazuyoshi Yoshii:
Correlated Tensor Factorization for Audio Source Separation. ICASSP 2018: 731-735 - [c74]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Unsupervised Beamforming Based on Multichannel Nonnegative Matrix Factorization for Noisy Speech Recognition. ICASSP 2018: 5734-5738 - [c73]Hirofumi Inaguma, Masato Mimura, Koji Inoue, Kazuyoshi Yoshii, Tatsuya Kawahara:
An End-to-End Approach to Joint Social Signal Detection and Automatic Speech Recognition. ICASSP 2018: 6214-6218 - [c72]Hiroaki Tsushima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Interactive Arrangement of Chords and Melodies Based on a Tree-Structured Generative Model. ISMIR 2018: 145-151 - [i6]Eita Nakamura, Kazuyoshi Yoshii:
Statistical Piano Reduction Controlling Performance Difficulty. CoRR abs/1808.05006 (2018) - 2017
- [j15]Karim Youssef, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Simultaneous Identification and Localization of Still and Mobile Speakers Based on Binaural Robot Audition. J. Robotics Mechatronics 29(1): 59-71 (2017) - [j14]Kouhei Sekiguchi, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Layout Optimization of Cooperative Distributed Microphone Arrays Based on Estimation of Source Separation Performance. J. Robotics Mechatronics 29(1): 83-93 (2017) - [j13]Misato Ohkita, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Audio-Visual Beat Tracking Based on a State-Space Model for a Robot Dancer Performing with a Human Dancer. J. Robotics Mechatronics 29(1): 125-136 (2017) - [j12]Yoshiaki Bando, Hiroshi Saruwatari, Nobutaka Ono, Shoji Makino, Katsutoshi Itoyama, Daichi Kitamura, Masaru Ishimura, Moe Takakusaki, Narumi Mae, Kouei Yamaoka, Yutaro Matsui, Yuichi Ambe, Masashi Konyo, Satoshi Tadokoro, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Low Latency and High Quality Two-Stage Human-Voice-Enhancement System for a Hose-Shaped Rescue Robot. J. Robotics Mechatronics 29(1): 198-212 (2017) - [j11]Eita Nakamura, Kazuyoshi Yoshii, Shigeki Sagayama:
Rhythm Transcription of Polyphonic Piano Music Based on Merged-Output HMM for Multiple Voices. IEEE ACM Trans. Audio Speech Lang. Process. 25(4): 794-806 (2017) - [j10]Eita Nakamura, Kazuyoshi Yoshii, Simon Dixon:
Note Value Recognition for Piano Transcription Using Markov Random Fields. IEEE ACM Trans. Audio Speech Lang. Process. 25(9): 1846-1858 (2017) - [c71]Kousuke Itakura, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Bayesian multichannel nonnegative matrix factorization for audio source separation and localization. ICASSP 2017: 551-555 - [c70]Masato Mimura, Yoshiaki Bando, Kazuki Shimada, Shinsuke Sakai, Kazuyoshi Yoshii, Tatsuya Kawahara:
Combined Multi-Channel NMF-Based Robust Beamforming for Noisy Speech Recognition. INTERSPEECH 2017: 2451-2455 - [c69]Eita Nakamura, Kazuyoshi Yoshii, Haruhiro Katayose:
Performance Error Detection and Post-Processing for Fast and Accurate Symbolic Music Alignment. ISMIR 2017: 347-353 - [c68]Ryo Nishikimi, Eita Nakamura, Masataka Goto, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Scale- and Rhythm-Aware Musical Note Estimation for Vocal F0 Trajectories Based on a Semi-Tatum-Synchronous Hierarchical Hidden Semi-Markov Model. ISMIR 2017: 376-382 - [c67]Hiroaki Tsushima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Function- and Rhythm-Aware Melody Harmonization Based on Tree-Structured Parsing and Split-Merge Sampling of Chord Sequences. ISMIR 2017: 502-508 - [c66]Antoine Liutkus, Kazuyoshi Yoshii:
A diagonal plus low-rank covariance model for computationally efficient source separation. MLSP 2017: 1-6 - [c65]Masaya Wake, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Semi-Blind speech enhancement basedon recurrent neural network for source separation and dereverberation. MLSP 2017: 1-6 - [c64]Kazuyoshi Yoshii, Eita Nakamura, Katsutoshi Itoyama, Masataka Goto:
Infinite probabilistic latent component analysis for audio source separation. MLSP 2017: 1-6 - [i5]Eita Nakamura, Kazuyoshi Yoshii, Shigeki Sagayama:
Rhythm Transcription of Polyphonic Piano Music Based on Merged-Output HMM for Multiple Voices. CoRR abs/1701.08343 (2017) - [i4]Eita Nakamura, Kazuyoshi Yoshii, Simon Dixon:
Note Value Recognition for Rhythm Transcription Using a Markov Random Field Model for Musical Scores and Performances of Piano Music. CoRR abs/1703.08144 (2017) - [i3]Hiroaki Tsushima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Generative Statistical Models with Self-Emergent Grammar of Chord Sequences. CoRR abs/1708.02255 (2017) - [i2]Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Statistical Speech Enhancement Based on Probabilistic Integration of Variational Autoencoder and Non-Negative Matrix Factorization. CoRR abs/1710.11439 (2017) - 2016
- [j9]Tomoyasu Nakano, Kazuyoshi Yoshii, Masataka Goto:
Musical Similarity and Commonness Estimation Based on Probabilistic Generative Models of Musical Elements. Int. J. Semantic Comput. 10(1): 27-52 (2016) - [j8]Yukara Ikemiya, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Singing Voice Separation and Vocal F0 Estimation Based on Mutual Combination of Robust Principal Component Analysis and Subharmonic Summation. IEEE ACM Trans. Audio Speech Lang. Process. 24(11): 2084-2095 (2016) - [c63]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Variational Bayesian multi-channel robust NMF for human-voice enhancement with a deformable and partially-occluded microphone array. EUSIPCO 2016: 1018-1022 - [c62]Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Rhythm transcription of MIDI performances based on hierarchical Bayesian modelling of repetition and modification of musical note patterns. EUSIPCO 2016: 1946-1950 - [c61]Kousuke Itakura, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
A unified Bayesian model of time-frequency clustering and low-rank approximation for multi-channel source separation. EUSIPCO 2016: 2280-2284 - [c60]Kazuyoshi Yoshii, Katsutoshi Itoyama, Masataka Goto:
Student's T nonnegative matrix factorization and positive semidefinite tensor factorization for single-channel audio source separation. ICASSP 2016: 51-55 - [c59]Eita Nakamura, Masatoshi Hamanaka, Keiji Hirata, Kazuyoshi Yoshii:
Tree-structured probabilistic model of monophonic written music based on the generative theory of tonal music. ICASSP 2016: 276-280 - [c58]Kouhei Sekiguchi, Yoshiaki Bando, Keisuke Nakamura, Kazuhiro Nakadai, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Online simultaneous localization and mapping of multiple sound sources and asynchronous microphone arrays. IROS 2016: 1973-1979 - [c57]Yuta Ojima, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
A Hierarchical Bayesian Model of Chords, Pitches, and Spectrograms for Multipitch Analysis. ISMIR 2016: 309-315 - [c56]Ryo Nishikimi, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Musical Note Estimation for F0 Trajectories of Singing Voices Based on a Bayesian Semi-Beat-Synchronous HMM. ISMIR 2016: 461-467 - [c55]Tomoyasu Nakano, Daichi Mochihashi, Kazuyoshi Yoshii, Masataka Goto:
Musical Typicality: How Many Similar Songs Exist?. ISMIR 2016: 695-701 - [c54]Koichi Kitamura, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Student's t multichannel nonnegative matrix factorization for blind source separation. IWAENC 2016: 1-5 - [c53]Yoshiaki Bando, Hiroki Suhara, Motoyasu Tanaka, Tetsushi Kamegawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Fumitoshi Matsuno, Hiroshi G. Okuno:
Sound-based online localization for an in-pipe snake robot. SSRR 2016: 207-213 - [i1]Yukara Ikemiya, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Singing Voice Separation and Vocal F0 Estimation based on Mutual Combination of Robust Principal Component Analysis and Subharmonic Summation. CoRR abs/1604.00192 (2016) - 2015
- [j7]Izaya Nishimuta, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Toward a quizmaster robot for speech-based multiparty interaction. Adv. Robotics 29(18): 1205-1219 (2015) - [c52]Yoshiaki Bando, Takuma Otsuka, Ikkyu Aihara, Hiromitsu Awano, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi Gitchang Okuno:
Recognition of In-Field Frog Chorusing Using Bayesian Nonparametric Microphone Array Processing. AAAI Workshop: Computational Sustainability 2015 - [c51]Satoshi Maruo, Kazuyoshi Yoshii, Katsutoshi Itoyama, Matthias Mauch, Masataka Goto:
A feedback framework for improved chord recognition based on NMF-based approximate note transcription. ICASSP 2015: 196-200 - [c50]Yukara Ikemiya, Kazuyoshi Yoshii, Katsutoshi Itoyama:
Singing voice analysis and editing based on mutually dependent F0 estimation and source separation. ICASSP 2015: 574-578 - [c49]Yoshiaki Bando, Takuma Otsuka, Katsutoshi Itoyama, Kazuyoshi Yoshii, Yoko Sasaki, Satoshi Kagami, Hiroshi G. Okuno:
Challenges in deploying a microphone array to localize and separate sound sources in real auditory scenes. ICASSP 2015: 723-727 - [c48]Kousuke Itakura, Izaya Nishimuta, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Bayesian integration of sound source separation and speech recognition: a new approach to simultaneous speech recognition. INTERSPEECH 2015: 736-740 - [c47]Kouhei Sekiguchi, Yoshiaki Bando, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Optimizing the layout of multiple mobile robots for cooperative sound source separation. IROS 2015: 5548-5554 - [c46]Misato Ohkita, Yoshiaki Bando, Yukara Ikemiya, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Audio-visual beat tracking based on a state-space model for a music robot dancing with humans. IROS 2015: 5555-5560 - [c45]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Microphone-accelerometer based 3D posture estimation for a hose-shaped rescue robot. IROS 2015: 5580-5586 - [c44]Masataka Goto, Kazuyoshi Yoshii, Tomoyasu Nakano:
Songle Widget: Making Animation and Physical Devices Synchronized with Music Videos on the Web. ISM 2015: 85-88 - [c43]Tomoyasu Nakano, Kazuyoshi Yoshii, Masataka Goto:
Musical Similarity and Commonness Estimation Based on Probabilistic Generative Models. ISM 2015: 197-204 - [c42]Kazuyoshi Yoshii, Katsutoshi Itoyama, Masataka Goto:
Infinite Superimposed Discrete All-Pole Modeling for Multipitch Analysis of Wavelet Spectrograms. ISMIR 2015: 86-92 - [c41]Karim Youssef, Katsutoshi Itoyama, Kazuyoshi Yoshii:
Identification and Localization of One or Two Concurrent Speakers in a Binaural Robotic Context. SMC 2015: 407-412 - [c40]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Human-voice enhancement based on online RPCA for a hose-shaped rescue robot with a microphone array. SSRR 2015: 1-6 - [c39]Akira Maezawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Unified inter- and intra-recording duration model for multiple music audio alignment. WASPAA 2015: 1-5 - 2014
- [j6]Matthew E. P. Davies, Philippe Hamel, Kazuyoshi Yoshii, Masataka Goto:
AutoMashUpper: automatic creation of multi-song music mashups. IEEE ACM Trans. Audio Speech Lang. Process. 22(12): 1726-1737 (2014) - [j5]Akira Maezawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Nonparametric Bayesian dereverberation of power spectrograms based on infinite-order autoregressive processes. IEEE ACM Trans. Audio Speech Lang. Process. 22(12): 1918-1930 (2014) - [c38]Izaya Nishimuta, Naoki Hirayama, Kazuyoshi Yoshii, Katsutoshi Itoyama, Hiroshi G. Okuno:
A robot quizmaster that can localize, separate, and recognize simultaneous utterances for a fastest-voice-first quiz game. Humanoids 2014: 967-972 - [c37]Kazuyoshi Yoshii, Hiromasa Fujihara, Tomoyasu Nakano, Masataka Goto:
Cultivating vocal activity detection for music audio signals in a circulation-type crowdsourcing ecosystem. ICASSP 2014: 624-628 - [c36]Tomoyasu Nakano, Kazuyoshi Yoshii, Masataka Goto:
Vocal timbre analysis using latent Dirichlet allocation and cross-gender vocal timbre similarity. ICASSP 2014: 5202-5206 - [c35]Tomohiko Nakamura, Hirokazu Kameoka, Kazuyoshi Yoshii, Masataka Goto:
Timbre replacement of harmonic and drum components for music audio signals. ICASSP 2014: 7470-7474 - [c34]Taro Masuda, Kazuyoshi Yoshii, Masataka Goto, Shigeo Morishima:
Spotting a Query Phrase from Polyphonic Music Audio Signals Based on Semi-supervised Nonnegative Matrix Factorization. ISMIR 2014: 227-232 - [c33]Akira Maezawa, Katsutoshi Itoyama, Kazuyoshi Yoshii, Hiroshi G. Okuno:
Bayesian Audio Alignment based on a Unified Model of Music Composition and Performance. ISMIR 2014: 233-238 - [c32]Shoto Sasaki, Kazuyoshi Yoshii, Tomoyasu Nakano, Masataka Goto, Shigeo Morishima:
LyricsRadar: A Lyrics Retrieval System Based on Latent Topics of Lyrics. ISMIR 2014: 585-590 - [c31]Yoshiaki Bando, Katsutoshi Itoyama, Masashi Konyo, Satoshi Tadokoro, Kazuhiro Nakadai, Kazuyoshi Yoshii, Hiroshi G. Okuno:
A sound-based online method for estimating the time-varying posture of a hose-shaped robot. SSRR 2014: 1-6 - 2013
- [c30]Kazuyoshi Yoshii, Masataka Goto:
Infinite kernel linear prediction for joint estimation of spectral envelope and fundamental frequency. ICASSP 2013: 463-467 - [c29]Kazuyoshi Yoshii, Ryota Tomioka, Daichi Mochihashi, Masataka Goto:
Infinite Positive Semidefinite Tensor Factorization for Source Separation of Mixture Signals. ICML (3) 2013: 576-584 - [c28]Yoko Sasaki, Naotaka Hatao, Kazuyoshi Yoshii, Satoshi Kagami:
Nested iGMM recognition and multiple hypothesis tracking of moving sound sources for mobile robot audition. IROS 2013: 3930-3936 - [c27]Philippe Hamel, Matthew E. P. Davies, Kazuyoshi Yoshii, Masataka Goto:
Transfer Learning In Mir: Sharing Learned Latent Representations For Music Audio Classification And Similarity. ISMIR 2013: 9-14 - [c26]Kazuyoshi Yoshii, Ryota Tomioka, Daichi Mochihashi, Masataka Goto:
Beyond NMF: Time-Domain Audio Source Separation without Phase Reconstruction. ISMIR 2013: 369-374 - [c25]Satoru Fukayama, Kazuyoshi Yoshii, Masataka Goto:
Chord-Sequence-Factory: A Chord Arrangement System Modifying Factorized Chord Sequence Probabilities. ISMIR 2013: 457-462 - [c24]Matthew E. P. Davies, Philippe Hamel, Kazuyoshi Yoshii, Masataka Goto:
AutoMashUpper: An Automatic Multi-Song Mashup System. ISMIR 2013: 575-580 - [c23]Yoko Sasaki, Kazuyoshi Yoshii, Satoshi Kagami:
A nested infinite Gaussian mixture model for identifying known and unknown audio events. WIAMIS 2013: 1-4 - 2012
- [j4]Kazuyoshi Yoshii, Masataka Goto:
A Nonparametric Bayesian Multipitch Analyzer Based on Infinite Latent Harmonic Allocation. IEEE Trans. Speech Audio Process. 20(3): 717-730 (2012) - [c22]Kazuyoshi Yoshii, Masataka Goto:
Unsupervised music understanding based on nonparametric Bayesian models. ICASSP 2012: 5353-5356 - [c21]Kazuyoshi Yoshii, Masataka Goto:
Infinite Composite Autoregressive Models for Music Signal Analysis. ISMIR 2012: 79-84 - [c20]Masataka Goto, Jun Ogata, Kazuyoshi Yoshii, Hiromasa Fujihara, Matthias Mauch, Tomoyasu Nakano:
PodCastle and songle: Crowdsourcing-based web services for spoken document retrieval and active music listening. ITA 2012: 298-299 - [c19]Masataka Goto, Jun Ogata, Kazuyoshi Yoshii, Hiromasa Fujihara, Matthias Mauch, Tomoyasu Nakano:
PodCastle and songle: crowdsourcing-based web services for spoken content retrieval and active music listening. CrowdMM@ACM Multimedia 2012: 1-2 - [c18]Masataka Goto, Jun Ogata, Kazuyoshi Yoshii, Hiromasa Fujihara, Matthias Mauch, Tomoyasu Nakano:
PodCastle and Songle: Crowdsourcing-Based Web Services for Retrieval and Browsing of Speech and Music Content. CrowdSearch 2012: 36-41 - 2011
- [c17]Matthias Mauch, Hiromasa Fujihara, Kazuyoshi Yoshii, Masataka Goto:
Timbre and Melody Features for the Recognition of Vocal Activity and Instrumental Solos in Polyphonic Music. ISMIR 2011: 233-238 - [c16]Masataka Goto, Kazuyoshi Yoshii, Hiromasa Fujihara, Matthias Mauch, Tomoyasu Nakano:
Songle: A Web Service for Active Music Listening Improved by User Contributions. ISMIR 2011: 311-316 - [c15]Kazuyoshi Yoshii, Masataka Goto:
A Vocabulary-Free Infinity-Gram Model for Nonparametric Bayesian Chord Progression Analysis. ISMIR 2011: 645-650 - 2010
- [c14]Kazuyoshi Yoshii, Masataka Goto:
Infinite Latent Harmonic Allocation: A Nonparametric Bayesian Approach to Multipitch Analysis. ISMIR 2010: 309-314
2000 – 2009
- 2009
- [c13]Kazuyoshi Yoshii, Masataka Goto:
Continuous pLSI and Smoothing Techniques for Hybrid Music Recommendation. ISMIR 2009: 339-344 - [c12]Kazuyoshi Yoshii, Masataka Goto:
MusicCommentator: Generating Comments Synchronized with Musical Audio Signals by a Joint Probabilistic Model of Acoustic and Textual Features. ICEC 2009: 85-97 - [e1]Keiji Hirata, George Tzanetakis, Kazuyoshi Yoshii:
Proceedings of the 10th International Society for Music Information Retrieval Conference, ISMIR 2009, Kobe International Conference Center, Kobe, Japan, October 26-30, 2009. International Society for Music Information Retrieval 2009, ISBN 978-0-9813537-0-8 [contents] - 2008
- [j3]Kazuyoshi Yoshii, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
An Efficient Hybrid Music Recommender System Using an Incrementally Trainable Probabilistic Generative Model. IEEE Trans. Speech Audio Process. 16(2): 435-447 (2008) - [c11]Takeshi Mizumoto, Ryu Takeda, Kazuyoshi Yoshii, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
A robot listens to music and counts its beats aloud by separating music from counting voice. IROS 2008: 1538-1543 - [c10]Kazumasa Murata, Kazuhiro Nakadai, Kazuyoshi Yoshii, Ryu Takeda, Toyotaka Torii, Hiroshi G. Okuno, Yuji Hasegawa, Hiroshi Tsujino:
A robot uses its own microphone to synchronize its steps to musical beats while scatting and singing. IROS 2008: 2459-2464 - [c9]Kouhei Sumi, Katsutoshi Itoyama, Kazuyoshi Yoshii, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Automatic Chord Recognition Based on Probabilistic Integration of Chord Transition and Bass Pitch Estimation. ISMIR 2008: 39-44 - [c8]Kazumasa Murata, Kazuhiro Nakadai, Kazuyoshi Yoshii, Ryu Takeda, Toyotaka Torii, Hiroshi G. Okuno, Yuji Hasegawa, Hiroshi Tsujino:
A Robot Singer with Music Recognition Based on Real-Time Beat Tracking. ISMIR 2008: 199-204 - [c7]Kazuyoshi Yoshii, Masataka Goto:
Music Thumbnailer: Visualizing Musical Pieces in Thumbnail Images Based on Acoustic Features. ISMIR 2008: 211-216 - 2007
- [j2]Kazuyoshi Yoshii, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Drumix: An Audio Player with Real-time Drum-part Rearrangement Functions for Active Music Listening. Inf. Media Technol. 2(2): 601-611 (2007) - [j1]Kazuyoshi Yoshii, Masataka Goto, Hiroshi G. Okuno:
Drum Sound Recognition for Polyphonic Audio Signals by Adaptation and Matching of Spectrogram Templates With Harmonic Structure Suppression. IEEE Trans. Speech Audio Process. 15(1): 333-345 (2007) - [c6]Kazuyoshi Yoshii, Kazuhiro Nakadai, Toyotaka Torii, Yuji Hasegawa, Hiroshi Tsujino, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
A biped robot that keeps steps in time with musical beats while listening to music with its own ears. IROS 2007: 1743-1750 - [c5]Kazuyoshi Yoshii, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Improving Efficiency and Scalability of Model-Based Music Recommender System Based on Incremental Training. ISMIR 2007: 89-94 - 2006
- [c4]Kazuyoshi Yoshii, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
An Error Correction Framework Based on Drum Pattern Periodicity for Improving Drum Sound Detection. ICASSP (5) 2006: 237-240 - [c3]Kazuyoshi Yoshii, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno:
Hybrid Collaborative and Content-based Music Recommendation Using Probabilistic Model with Latent User Preferences. ISMIR 2006: 296-301 - 2004
- [c2]Kazuyoshi Yoshii, Masataka Goto, Hiroshi G. Okuno:
Drum sound identification for polyphonic music using template adaptation and matching methods. SAPA@INTERSPEECH 2004: 51 - [c1]Kazuyoshi Yoshii, Masataka Goto, Hiroshi G. Okuno:
Automatic Drum Sound Description for Real-World Music Using Template Adaptation and Matching Methods. ISMIR 2004
Coauthor Index
aka: Hiroshi Gitchang Okuno
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2025-01-20 22:57 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint