Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
staff:vatolkin:publications [2022-02-01 16:01]
igor.vatolkin [Journal Articles]
staff:vatolkin:publications [2022-03-17 18:37]
igor.vatolkin
Line 40: Line 40:
  
 ===== Peer-Reviewed Conference Proceedings ===== ===== Peer-Reviewed Conference Proceedings =====
 +
 +<​html><​b><​font color=#​006633>​[c38]</​font></​b>​ <i> I. Vatolkin</​i>:<​b><​font color=#​0000FF>​ Identification of the Most Relevant Zygonic Statistics and Semantic Audio Features for Genre Recognition</​font></​b>​. Proceedings of the International Computer Music Conference (ICMC), <font color=#​996600>​2022</​font></​html>​
  
 <​html><​b><​font color=#​006633>​[c37]</​font></​b>​ <i> I. Vatolkin</​i>:<​b><​font color=#​0000FF>​ Improving Interpretable Genre Recognition with Audio Feature Statistics Based on Zygonic Theory</​font></​b>​. Proceedings of the 2nd Nordic Sound and Computing Conference (NordicSMC),​ <font color=#​996600>​2021</​font></​html>​ <​html><​b><​font color=#​006633>​[c37]</​font></​b>​ <i> I. Vatolkin</​i>:<​b><​font color=#​0000FF>​ Improving Interpretable Genre Recognition with Audio Feature Statistics Based on Zygonic Theory</​font></​b>​. Proceedings of the 2nd Nordic Sound and Computing Conference (NordicSMC),​ <font color=#​996600>​2021</​font></​html>​
  
-<​html><​b><​font color=#​006633>​[c36]</​font></​b>​ <i> I. Vatolkin, P. Ginsel, and G. Rudolph</​i>:<​b><​font color=#​0000FF>​ Advancements in the Music Information Retrieval Framework AMUSE over the Last Decade</​font></​b>​. ​Accepted for Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pp. 2383-2389, <font color=#​996600>​2021</​font></​html>​+<​html><​b><​font color=#​006633>​[c36]</​font></​b>​ <i> I. Vatolkin, P. Ginsel, and G. Rudolph</​i>:<​b><​font color=#​0000FF>​ Advancements in the Music Information Retrieval Framework AMUSE over the Last Decade</​font></​b>​. Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pp. 2383-2389, <font color=#​996600>​2021</​font></​html>​
  
-<​html><​b><​font color=#​006633>​[c35]</​font></​b>​ <i> I. Vatolkin, F. Ostermann, and M. Müller</​i>:<​b><​font color=#​0000FF>​ An Evolutionary Multi-Objective Feature Selection Approach for Detecting Music Segment Boundaries of Specific Types</​font></​b>​. ​Accepted for Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), pp. 1061-1069, <font color=#​996600>​2021</​font></​html>​+<​html><​b><​font color=#​006633>​[c35]</​font></​b>​ <i> I. Vatolkin, F. Ostermann, and M. Müller</​i>:<​b><​font color=#​0000FF>​ An Evolutionary Multi-Objective Feature Selection Approach for Detecting Music Segment Boundaries of Specific Types</​font></​b>​. Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), pp. 1061-1069, <font color=#​996600>​2021</​font></​html>​
  
-<​html><​b><​font color=#​006633>​[c34]</​font></​b>​ <i> I. Vatolkin, B. Adrian, and J. Kuzmic</​i>:<​b><​font color=#​0000FF>​ A Fusion of Deep and Shallow Learning to Predict Genres Based on Instrument and Timbre Features</​font></​b>​. ​Accepted for Proceedings of the 10th International Conference on Artificial Intelligence in Music, Sound, Art and Design (EvoMUSART),​ pp. 313-326, <font color=#​996600>​2021</​font></​html>​+<​html><​b><​font color=#​006633>​[c34]</​font></​b>​ <i> I. Vatolkin, B. Adrian, and J. Kuzmic</​i>:<​b><​font color=#​0000FF>​ A Fusion of Deep and Shallow Learning to Predict Genres Based on Instrument and Timbre Features</​font></​b>​. Proceedings of the 10th International Conference on Artificial Intelligence in Music, Sound, Art and Design (EvoMUSART),​ pp. 313-326, <font color=#​996600>​2021</​font></​html>​
  
-<​html><​b><​font color=#​006633>​[c33]</​font></​b>​ <i> I. Vatolkin, M. Koch, and M. Müller</​i>:<​b><​font color=#​0000FF>​ A Multi-Objective Evolutionary Approach to Identify Relevant Audio Features for Music Segmentation</​font></​b>​. ​Accepted for Proceedings of the 10th International Conference on Artificial Intelligence in Music, Sound, Art and Design (EvoMUSART),​ pp. 327-343, <font color=#​996600>​2021</​font></​html>​+<​html><​b><​font color=#​006633>​[c33]</​font></​b>​ <i> I. Vatolkin, M. Koch, and M. Müller</​i>:<​b><​font color=#​0000FF>​ A Multi-Objective Evolutionary Approach to Identify Relevant Audio Features for Music Segmentation</​font></​b>​. Proceedings of the 10th International Conference on Artificial Intelligence in Music, Sound, Art and Design (EvoMUSART),​ pp. 327-343, <font color=#​996600>​2021</​font></​html>​
  
 <​html><​b><​font color=#​006633>​[c32]</​font></​b>​ <i> I. Vatolkin</​i>:<​b><​font color=#​0000FF>​ Evolutionary Approximation of Instrumental Texture in Polyphonic Audio Recordings</​font></​b>​. Proceedings of the IEEE Congress on Evolutionary Computation (CEC), pp. 1-8, <font color=#​996600>​2020</​font></​html>​ <​html><​b><​font color=#​006633>​[c32]</​font></​b>​ <i> I. Vatolkin</​i>:<​b><​font color=#​0000FF>​ Evolutionary Approximation of Instrumental Texture in Polyphonic Audio Recordings</​font></​b>​. Proceedings of the IEEE Congress on Evolutionary Computation (CEC), pp. 1-8, <font color=#​996600>​2020</​font></​html>​
 
Last modified: 2023-03-21 18:36 by igor.vatolkin
DokuWikiRSS-Feed