Mood Tracking of Radio Station Broadcasts
|
|
- Alvin Gordon
- 6 years ago
- Views:
Transcription
1 Mood Tracking of Radio Station Broadcasts Jacek Grekow Faculty of Computer Science, Bialystok University of Technology, Wiejska 45A, Bialystok , Poland Abstract. This paper presents an example of a system for the analysis of emotions contained within radio broadcasts. We prepared training data, did feature extraction, built classifiers for music/speech discrimination and for emotion detection in music. To study changes in emotions, we used recorded broadcasts from 4 selected European radio stations. The collected data allowed us to determine the dominant emotion in the radio broadcasts and construct maps visualizing the distribution of emotions in time. The obtained results provide a new interesting view of the emotional content of radio station broadcasts. Keywords: Emotion detection, Mood tracking, Audio feature extraction, Music information retrieval, Radio broadcasts. 1 Introduction The overwhelming number of media outlets is constantly growing. This also applies to radio stations available on the Internet, over satellite, and over the air. On the one hand, the number of opportunities to listen to various radio shows has grown, but on the other, choosing the right station has become more difficult. Music information retrieval helps those people who listen to the radio mainly for the music. This technology is able to make a general detection of the genre, artist, and even emotion. Listening to music is particularly emotional. People need a variety of emotions, and music is perfectly suited to provide them. Listening to the radio station throughout the day, whether we want to it not, we are affected by the transmitted emotional content. In this paper, we focus on emotional analysis of the music presented by radio stations. During the course of a radio broadcast, these emotions can take on a variety of shades, change several times with varying intensity. This paper presents a method of tracking changing emotions during the course of a radio broadcast. The collected data allowed to determine the dominant emotion in the radio broadcast and construct maps visualizing the distribution of emotions in time. Music emotion detection studies are mainly based on two popular approaches: categorical or dimensional. The categorical approach [1][2][3][4] describes emotions with a discrete number of classes - affective adjectives. In the dimensional approach [5][6], emotions are described as numerical values of valence
2 2 Jacek Grekow and arousal. In this way an emotion of a song is represented as a point on an emotion space. In this work, we use the categorical approach. There are several other studies on the issue of mood tracking [7][8][9]. Lu et al. [3], apart from detecting emotions, tracked them, divided the music into several independent segments, each of which contained a homogeneous emotional expression. The use of mood tracking for indexing and searching multimedia databases has been used in the work of Grekow and Ras [10]. One wonders how long it takes a person to recognize emotion in musical compositions to which he/she is listening. Bachorik et al. [11] concluded that the majority of music listeners need 8 seconds to identify the emotion of a piece. This time is closely related to the length of a segment of music during emotion detection. Xiao et al. [12] found that the segment length should be no shorter than 4 sec and no longer than 16 sec. In various studies on the detection of emotion in music the segments are varying lengths. In [4][1], segment length is 30 sec. A 25 second segment was used by Yang et al. [5]. Fifteen second clips are used as ground truth data by Schmidt et al. [13]. The length of 1 sec segment was used by Schmidt and Kim in [6]. In this work, we use 6-second segments. A comprehensive review of the methods that have been proposed for music emotion recognition was prepared by Yang et al. [14]. Another paper surveying state-of-the-art automatic emotion recognition was presented by Kim et al. in [15]. The issue of mood tracking is not only limited to music. The paper by Mohammad [16] is an interesting extension of the topic; the author investigated the development of emotions in literary texts. Yeh et al. [17] tracked the continuous changes of emotional expressions in Mandarin speech. A method of profiling radio stations was described by Lidy and Rauber [18]. They used a technique of Self-Organizing Maps to organize the program coverage of radio stations on a two-dimensional map. This approach allows to profile the complete program of a radio station. 2 Music Data To conduct the study of emotion detection of radio stations, we prepared two sets of data. One of them was used for music/speech discrimination, and the other for the detection of emotion in music. A set of training data for music/speech discrimination consisted of 128 wav files, including 64 designated as speech and 64 marked as music. The training data were taken from the generally accessible data collection project MARSYAS ( The training data set for emotion detection consisted of 374 six-second fragments of different genres of music: classical, jazz, blues, country, disco, hip-hop, metal, pop, reggae, and rock. The tracks were all 22050Hz Mono 16-bit audio files in wav format. In this research we use 4 emotion classes: energetic-positive, energetic-negative, calm-negative, calm-positive, presented with their abbreviation in Table 1. They
3 Mood Tracking of Radio Station Broadcasts 3 cover the four quadrants of the 2 dimensional Thayer model of emotion [19]. They correspond to four basic emotion classes: happy, angry, sad and relaxed. Abbreviation Description e1 energetic-positive e2 energetic-negative e3 calm-negative e4 calm-positive Table 1. Description of mood labels Music samples were labeled by the author of this paper, a music expert with a university musical education. Six-second music samples were listened to and then labeled with one of the emotions (e1, e2, e3, e4). In the case when the music expert was not certain which emotion to assign, such a sample was rejected. In this way, the created labels were associated with only one emotion in the file. As a result, we obtained 4 sets of files: 101 files labeled e1, 107 files labeled e2, 78 files labeled e3, and 88 files labeled e4. To study changes in emotions, we used recorded broadcasts from 4 selected European radio stations: Polish Radio Dwojka (Classical/Culture), recorded on ; Polish Radio Trojka (Pop/Rock), recorded on ; BBC Radio 3 (Classical), recorded on ; ORF OE1 (Information/Culture), recorded on For each station we recorded 10 hours beginning at 10 A.M. The recorded broadcasts were segmented into 6-second fragments using sfplay.exe from MARSYAS software. For example, we obtained 6000 segments from one 10 h broadcast. 3 Features Extraction For features extraction, we used the framework for audio analysis of MARSYAS software, written by George Tzanetakis [20]. MARSYAS is implemented in C++ and retains the ability to output feature extraction data to ARFF format [21]. With the tool bextract.exe, the following features can be extracted: Zero Crossings, Spectral Centroid, Spectral Flux, Spectral Rolloff, Mel-Frequency Cepstral Coefficients (MFCC), and chroma features - 31 features in total. For each of these basic features, four statistic features were calculated: 1. The mean of the mean (calculate mean over the 20 frames, and then calculate the mean of this statistic over the entire segment); 2. The mean of the standard deviation (calculate the standard deviation of the feature over 20 frames, and then calculate the mean these standard deviations over the entire segment); 3. The standard deviation of the mean (calculate the mean of the feature over
4 4 Jacek Grekow 20 frames, and then calculate the standard deviation of these values over the entire segment); 4. The standard deviation of the standard deviation (calculate the standard deviation of the feature over 20 frames, and then calculate the standard deviation of these values over the entire segment). In this way, we obtained 124 features. The input data during features extraction were 6-second segments in wav format, sample rate 22050, channels: 1, Bits: 16. An example of using bextract.exe from the MARSYAS v0.2 package to extract features: bextract.exe -fe -sv colllection -w outputfile where collection is file with list of input files and outputfile is name of output file in ARFF format. For each 6-second file we obtained a representative single feature vector. The obtained vectors were used for building classifiers and for predicting new instances. 4 Classification 4.1 The Construction of Classifiers We built two classifiers, one for music/speech discrimination and the second for emotion detection, using the WEKA package [21]. During the construction of the classifier for music/speech discrimination, we tested the following algorithms: J48, RandomForest, BayesNet, SMO [22]. The classification results were calculated using a cross validation evaluation CV-10. The best accuracy (98%) was achieved using SMO algorithm, which is an implementation of support vector machines (SVM) algorithm. The second best algorithm was Random-Forest (94% accuracy). During the construction of the classifier for emotion detection, we tested the following algorithms: J48, RandomForest, BayesNet, IBk (K-nn), SMO (SVM). The highest accuracy (55.61%) was obtained for SMO algorithm. SMO was trained using polynominal kernel. The classification results were calculated using a cross validation evaluation CV-10. After applying attribute selection (attribute evaluator: WrapperSubsetEval, search method BestFirst), classifier accuracy improved to 60.69%. classified as a b c d a = e b = e c = e d = e Table 2. Confusion matrix The confusion matrix (Table 2) obtained during classifier evaluation shows that the most recognized emotion was e2 (Precision 0.706, Recall 0.72, F-measure
5 Mood Tracking of Radio Station Broadcasts ), and the next emotion was e1 (Precision 0.542, Recall 0.644, F-measure 0.588). We may notice a considerable amount of mistakes between the emotions of the left and right quadrants of the Thayer model, that is between e1 and e2, and analogously between e3 and e4. This is confirmed by the fact that detection on the arousal axis of Thayers model is easier. There are less mistakes made between the top and bottom quadrants. At the same time, recognition of emotions on the valence axis (positive-negative) is more difficult. 4.2 Analysis of Recordings During the analysis of the recorded radio broadcasts, we conducted a two-phase classification. The recorded radio program was divided into 6-second segments. For each segment, we extracted a feature vector. This feature vector was first used to detect if the given segment is speech or music. If the current segment was music, then we used a second classifier to predict what type of emotion it contained. For features extraction, file segmentation, use of classifiers to predict new instances, and visualization of results, we wrote a Java application that connected different software products: MARSYAS, MATLAB and WEKA package. 5 Results of Mood Tracking in Radio Stations The percentages of speech, music, and emotion in music obtained during the segment classification of 10-hour broadcasts of four radio stations are presented in Table 3. On the basis of these results, radio stations can be compared in two ways. The first way is to compare the amount of music and speech in the radio broadcasts, and the second is to compare the occurrence of individual emotions. PR Dwojka PR Trojka BBC Radio 3 ORF OE1 speech 59.37% 73.35% 32.25% 69.10% music 40.63% 26.65% 67.75% 30.90% e1 4.78% 4.35% 2.43% 2.48% e2 5.35% 14.43% 1.00% 0.92% e % 6.02% 56.19% 22.53% e % 1.85% 8.13% 4.97% e1 in music 11.76% 16.32% 3.58% 8.02% e2 in music 13.16% 54.14% 1.47% 2.98% e3 in music 49.89% 22.59% 82.93% 72.91% e4 in music 25.17% 6.94% 12.00% 16.08% Table 3. Percentage of speech, music, and emotion in music in 10-hour broadcasts of four radio stations
6 6 Jacek Grekow 5.1 Comparison of Radio Stations The dominant station in the amount of music presented was BBC Radio 3 (67.75%). We noted a similar ratio of speech to music in the broadcasts of PR Trojka and ORF OE1, in both of which speech dominated (73.35% and 69.10%, respectively). A more balanced amount of speech and music was noted on PR Dwojka (59.37% and 40.63%, respectively). Comparing the content of emotions, we can see that PR Trojka clearly differs from the other radio stations, because the dominant emotion is e2 energeticnegative (54.14%) and e4 calm-positive occurs the least often (6.94%). We noted a clear similarity between BBC Radio 3 and ORF OE1, where the dominant emotion was e3 calm-negative (82.93% and 72.91%, respectively). Also, the proportions of the other emotions (e1, e2, e4) were similar for these stations. We could say that emotionally these stations are similar, except that considering the speech to music ratio, BBC Radio 3 had much more music. The dominant emotion for PR Dwojka was e3, which is somewhat similar to BBC Radio 3 and ORF OE1. Compared to the other stations, PR Dwojka had the most (25.17%) e4 calm-positive music. 5.2 Emotion Maps The figures (Fig. 1, Fig. 2, Fig. 3, Fig. 4) present speech and emotion maps for each radio broadcast. Each point on the map is the value obtained from the classification of a 6-second segment. These show which emotions occurred at given hours of the broadcasts. For PR Dwojka (Fig. 1), there are clear musical segments ( , ) during which e3 dominated. At the end of the day ( ), emotion e2 occurs sporadically. It is interesting that e1 and e4 (from right half of the Thayer model) did not occur in the morning. For PR Trojka (Fig. 2), emotion e4 did not occur in the morning, and e2 and e3 dominated (segments and ). For BBC Radio 3 (Fig. 3), we observed almost a complete lack of energetic emotions (e1 and e2) in the afternoon (segments after 3200). For ORF OE1 (Fig. 4), e3 dominated up to segment 3600, and then broadcasts without music dominated. The presented analyses of maps of emotions could be developed by examining the quantity of changes of emotions or the distribution of daily emotions. 6 Conclusions This paper presents an example of a system for the analysis of emotions contained within radio broadcasts. The collected data allowed to determine the dominant emotion in the radio broadcast and present the amount of speech and music. The obtained results provide a new interesting view of the emotional content of radio stations. The precision of the constructed maps visualizing the distribution of emotions in time obviously depends on the precision of the classifiers of emotion detection.
7 Mood Tracking of Radio Station Broadcasts 7 Fig. 1. Map of speech and music emotion in PR Dwojka 10h broadcast Fig. 2. Map of speech and music emotion in PR Trojka 10h broadcast
8 8 Jacek Grekow Fig. 3. Map of speech and music emotion in BBC Radio 3 10h broadcast Fig. 4. Map of speech and music emotion in ORF OE1 10h broadcast
9 Mood Tracking of Radio Station Broadcasts 9 Their accuracy could be better. This is still associated with the imperfection of features for audio analysis. In this matter, there is still much to be done. We could also test audio features extracted by other software for feature extraction, such as jaudio or MIR toolbox. Also, musical file labeling, which are input data for learning classifiers, could be made by a bigger number of music experts; this would enhance the reliability of the classifiers. Development of the presented system to include emotion detection in speech also seems to be a logical prospect in the future. A system for the analysis of emotions contained within radio broadcasts could be a helpful tool for people planning radio programs enabling them to consciously plan the emotional distribution in the broadcast music. Another example of applying this system could be an additional tool for radio station searching. Because the perception of emotions can be subjective and different people perceive emotions slightly differently, the emotional analysis of radio stations could be dependent on the user s preferences. Search profiling of radio stations taking into consideration the user would be an interesting solution. Acknowledgments. This paper is supported by the S/WI/3/2013. References 1. Li, T. and Ogihara, M.: Detecting emotion in music. Proceedings of the Fifth International Symposium on Music Information Retrieval, pp (2003) 2. Grekow, J. and Ras, Z.: Detecting emotions in classical music from MIDI files. Foundations of Intelligent Systems: ISMIS 2009, LNAI, Vol. 5722, pp (2009) 3. Lu, L., Liu, D. and Zhang, H.J.: Automatic mood detection and tracking of music audio signals. IEEE Transactions on Audio, Speech and Language Processing, vol. 14, no. 1, pp (2006) 4. Song, Y., Dixon, S. and Pearce, M.: Evaluation of Musical Features for Emotion Classifica-tion. In Proceedings of the 13th International Society for Music Information Retrieval Con-ference (2012) 5. Yang,Y.-H., Lin,Y.C., Su,Y.F. and Chen, H.H.: A regression approach to music emotion recognition. IEEE Transactions on Audio, Speech, and Language Processing, Volume 16, Issue 2, pp (2008) 6. Schmidt, E. and Kim, Y.: Modeling Musical Emotion Dynamics with Conditional Random Fields. In Proceedings of the 12th International Society for Music Information Retrieval Conference, pp (2011) 7. Schmidt, E.M., Turnbull, D. and Kim Y.E.: Feature Selection for Content-Based, Time-Varying Musical Emotion Regression. Proc. ACM SIGMM International Conference on Multimedia Information Retrieval, Philadelphia, PA (2010) 8. Schmidt, E.M. and Kim, Y.E.: Prediction of time-varying musical mood distributions from audio. Proceedings of the 2010 International Society for Music Information Retrieval Conference, Utrecht, Netherlands (2010) 9. Grekow, J.: Mood tracking of musical compositions. Foundations of Intelligent Systems: ISMIS 2012, Lecture Notes in Computer Science, pp , eds. Li Chen, Alexander Felfernig, Jiming Liu, Zbigniew Ras; 20th International Symposium, Macau, China (2012)
10 10 Jacek Grekow 10. Grekow, J. and Ras, Z.: Emotion Based MIDI Files Retrieval System. Advances in Music Information Retrieval, Studies in Computational Intelligence, Springer (2010) 11. Bachorik, J.P., Bangert, M., Loui, P., Larke, K., Berger, J., Rowe, R. and Schlaug, G.: Emotion in motion: Investigating the time-course of emotional judgments of musical stimuli. Music Perception, vol. 26, no. 4, pp (2009) 12. Xiao, Z., Dellandrea, E., Dou, W. and Chen, L.: What is the best segment duration for music mood analysis? International Workshop on Content-Based Multimedia Indexing (CBMI 2008), pp (2008) 13. Schmidt, E. M., Scott, J.J. and Kim, Y.E.: Feature Learning in Dynamic Environments: Modeling the Acoustic Structure of Musical Emotion. In Proceedings of the 12th International Society for Music Information Retrieval Conference, pp (2012) 14. Yang Y.H. and Homer H. Chen, H.H.: Machine Recognition of Music Emotion: A Re-view. ACM Transactions on Intelligent Systems and Technology, Volume 3, Issue 3, Article No. 40 (2012) 15. Kim, Y., Schmidt, E., Migneco, R., Morton, B., Richardson, P., Scott, J., Speck, J. and Turnbull, D.: State of the Art Report: Music Emotion Recognition: A State of the Art Review. In Proceedings of the 11th International Society for Music Information Retrieval Conference, pp (2010) 16. Mohammad, S.: From Once Upon a Time to Happily Ever After: Tracking Emotions in Novels and Fairy Tales. Proceedings of the ACL 2011 Workshop on Language Technology for Cultural Heritage, Social Sciences, and Humanities, pp , Portland, OR, USA (2011) 17. Yeh, J., Pao, T., Pai, Ch., Cheng, Y.: Tracking and Visualizing the Changes of Mandarin Emotional Expression. ICIC 2008, LNCS 5226, pp (2008) 18. Lidy, T. and Rauber, A.: Visually Profiling Radio Stations. In Proceedings of the 7th Inter-national Conference on Music Information Retrieval (2006) 19. Thayer, R.E.: The biopsychology arousal. Oxford University Press (1989) 20. Tzanetakis, G. and Cook, P.: Marsyas: A framework for audio analysis. Organized Sound, 10: (2000) 21. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P. and Witten, I.H.: The WEKA Data Mining Software: An Update; SIGKDD Explorations, Volume 11, Issue 1 (2009) 22. Witten, I.H. and Frank, E.: Data Mining: Practical machine learning tools and techniques. Morgan Kaufmann, San Francisco, CA, USA (2005)
INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION
INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION ULAŞ BAĞCI AND ENGIN ERZIN arxiv:0907.3220v1 [cs.sd] 18 Jul 2009 ABSTRACT. Music genre classification is an essential tool for
More informationMusic Emotion Recognition. Jaesung Lee. Chung-Ang University
Music Emotion Recognition Jaesung Lee Chung-Ang University Introduction Searching Music in Music Information Retrieval Some information about target music is available Query by Text: Title, Artist, or
More informationMusic Genre Classification and Variance Comparison on Number of Genres
Music Genre Classification and Variance Comparison on Number of Genres Miguel Francisco, miguelf@stanford.edu Dong Myung Kim, dmk8265@stanford.edu 1 Abstract In this project we apply machine learning techniques
More informationPredicting Time-Varying Musical Emotion Distributions from Multi-Track Audio
Predicting Time-Varying Musical Emotion Distributions from Multi-Track Audio Jeffrey Scott, Erik M. Schmidt, Matthew Prockup, Brandon Morton, and Youngmoo E. Kim Music and Entertainment Technology Laboratory
More informationExploring Relationships between Audio Features and Emotion in Music
Exploring Relationships between Audio Features and Emotion in Music Cyril Laurier, *1 Olivier Lartillot, #2 Tuomas Eerola #3, Petri Toiviainen #4 * Music Technology Group, Universitat Pompeu Fabra, Barcelona,
More informationA Categorical Approach for Recognizing Emotional Effects of Music
A Categorical Approach for Recognizing Emotional Effects of Music Mohsen Sahraei Ardakani 1 and Ehsan Arbabi School of Electrical and Computer Engineering, College of Engineering, University of Tehran,
More informationMusic Information Retrieval with Temporal Features and Timbre
Music Information Retrieval with Temporal Features and Timbre Angelina A. Tzacheva and Keith J. Bell University of South Carolina Upstate, Department of Informatics 800 University Way, Spartanburg, SC
More informationABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC
ABSOLUTE OR RELATIVE? A NEW APPROACH TO BUILDING FEATURE VECTORS FOR EMOTION TRACKING IN MUSIC Vaiva Imbrasaitė, Peter Robinson Computer Laboratory, University of Cambridge, UK Vaiva.Imbrasaite@cl.cam.ac.uk
More informationDimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features
Dimensional Music Emotion Recognition: Combining Standard and Melodic Audio Features R. Panda 1, B. Rocha 1 and R. P. Paiva 1, 1 CISUC Centre for Informatics and Systems of the University of Coimbra, Portugal
More informationSubjective Similarity of Music: Data Collection for Individuality Analysis
Subjective Similarity of Music: Data Collection for Individuality Analysis Shota Kawabuchi and Chiyomi Miyajima and Norihide Kitaoka and Kazuya Takeda Nagoya University, Nagoya, Japan E-mail: shota.kawabuchi@g.sp.m.is.nagoya-u.ac.jp
More informationMUSI-6201 Computational Music Analysis
MUSI-6201 Computational Music Analysis Part 9.1: Genre Classification alexander lerch November 4, 2015 temporal analysis overview text book Chapter 8: Musical Genre, Similarity, and Mood (pp. 151 155)
More informationBi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset
Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset Ricardo Malheiro, Renato Panda, Paulo Gomes, Rui Paiva CISUC Centre for Informatics and Systems of the University of Coimbra {rsmal,
More informationA QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM
A QUER B EAMPLE MUSIC RETRIEVAL ALGORITHM H. HARB AND L. CHEN Maths-Info department, Ecole Centrale de Lyon. 36, av. Guy de Collongue, 69134, Ecully, France, EUROPE E-mail: {hadi.harb, liming.chen}@ec-lyon.fr
More informationAutomatic Extraction of Popular Music Ringtones Based on Music Structure Analysis
Automatic Extraction of Popular Music Ringtones Based on Music Structure Analysis Fengyan Wu fengyanyy@163.com Shutao Sun stsun@cuc.edu.cn Weiyao Xue Wyxue_std@163.com Abstract Automatic extraction of
More informationVECTOR REPRESENTATION OF EMOTION FLOW FOR POPULAR MUSIC. Chia-Hao Chung and Homer Chen
VECTOR REPRESENTATION OF EMOTION FLOW FOR POPULAR MUSIC Chia-Hao Chung and Homer Chen National Taiwan University Emails: {b99505003, homer}@ntu.edu.tw ABSTRACT The flow of emotion expressed by music through
More informationOutline. Why do we classify? Audio Classification
Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification Implementation Future Work Why do we classify
More informationComposer Identification of Digital Audio Modeling Content Specific Features Through Markov Models
Composer Identification of Digital Audio Modeling Content Specific Features Through Markov Models Aric Bartle (abartle@stanford.edu) December 14, 2012 1 Background The field of composer recognition has
More informationTOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC
TOWARD AN INTELLIGENT EDITOR FOR JAZZ MUSIC G.TZANETAKIS, N.HU, AND R.B. DANNENBERG Computer Science Department, Carnegie Mellon University 5000 Forbes Avenue, Pittsburgh, PA 15213, USA E-mail: gtzan@cs.cmu.edu
More informationMultimodal Music Mood Classification Framework for Christian Kokborok Music
Journal of Engineering Technology (ISSN. 0747-9964) Volume 8, Issue 1, Jan. 2019, PP.506-515 Multimodal Music Mood Classification Framework for Christian Kokborok Music Sanchali Das 1*, Sambit Satpathy
More informationWHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?
WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG? NICHOLAS BORG AND GEORGE HOKKANEN Abstract. The possibility of a hit song prediction algorithm is both academically interesting and industry motivated.
More informationSinger Identification
Singer Identification Bertrand SCHERRER McGill University March 15, 2007 Bertrand SCHERRER (McGill University) Singer Identification March 15, 2007 1 / 27 Outline 1 Introduction Applications Challenges
More informationAnalytic Comparison of Audio Feature Sets using Self-Organising Maps
Analytic Comparison of Audio Feature Sets using Self-Organising Maps Rudolf Mayer, Jakob Frank, Andreas Rauber Institute of Software Technology and Interactive Systems Vienna University of Technology,
More informationChord Classification of an Audio Signal using Artificial Neural Network
Chord Classification of an Audio Signal using Artificial Neural Network Ronesh Shrestha Student, Department of Electrical and Electronic Engineering, Kathmandu University, Dhulikhel, Nepal ---------------------------------------------------------------------***---------------------------------------------------------------------
More informationMusic Mood Classification - an SVM based approach. Sebastian Napiorkowski
Music Mood Classification - an SVM based approach Sebastian Napiorkowski Topics on Computer Music (Seminar Report) HPAC - RWTH - SS2015 Contents 1. Motivation 2. Quantification and Definition of Mood 3.
More informationA Study on Cross-cultural and Cross-dataset Generalizability of Music Mood Regression Models
A Study on Cross-cultural and Cross-dataset Generalizability of Music Mood Regression Models Xiao Hu University of Hong Kong xiaoxhu@hku.hk Yi-Hsuan Yang Academia Sinica yang@citi.sinica.edu.tw ABSTRACT
More informationResearch & Development. White Paper WHP 232. A Large Scale Experiment for Mood-based Classification of TV Programmes BRITISH BROADCASTING CORPORATION
Research & Development White Paper WHP 232 September 2012 A Large Scale Experiment for Mood-based Classification of TV Programmes Jana Eggink, Denise Bland BRITISH BROADCASTING CORPORATION White Paper
More informationInternational Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC
Scientific Journal of Impact Factor (SJIF): 5.71 International Journal of Advance Engineering and Research Development Volume 5, Issue 04, April -2018 e-issn (O): 2348-4470 p-issn (P): 2348-6406 MUSICAL
More informationToward Multi-Modal Music Emotion Classification
Toward Multi-Modal Music Emotion Classification Yi-Hsuan Yang 1, Yu-Ching Lin 1, Heng-Tze Cheng 1, I-Bin Liao 2, Yeh-Chin Ho 2, and Homer H. Chen 1 1 National Taiwan University 2 Telecommunication Laboratories,
More informationSinger Traits Identification using Deep Neural Network
Singer Traits Identification using Deep Neural Network Zhengshan Shi Center for Computer Research in Music and Acoustics Stanford University kittyshi@stanford.edu Abstract The author investigates automatic
More informationA Music Retrieval System Using Melody and Lyric
202 IEEE International Conference on Multimedia and Expo Workshops A Music Retrieval System Using Melody and Lyric Zhiyuan Guo, Qiang Wang, Gang Liu, Jun Guo, Yueming Lu 2 Pattern Recognition and Intelligent
More informationA Large Scale Experiment for Mood-Based Classification of TV Programmes
2012 IEEE International Conference on Multimedia and Expo A Large Scale Experiment for Mood-Based Classification of TV Programmes Jana Eggink BBC R&D 56 Wood Lane London, W12 7SB, UK jana.eggink@bbc.co.uk
More informationCombination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections
1/23 Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections Rudolf Mayer, Andreas Rauber Vienna University of Technology {mayer,rauber}@ifs.tuwien.ac.at Robert Neumayer
More informationMODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET
MODELING MUSICAL MOOD FROM AUDIO FEATURES AND LISTENING CONTEXT ON AN IN-SITU DATA SET Diane Watson University of Saskatchewan diane.watson@usask.ca Regan L. Mandryk University of Saskatchewan regan.mandryk@usask.ca
More informationClassification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors
Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors Priyanka S. Jadhav M.E. (Computer Engineering) G. H. Raisoni College of Engg. & Mgmt. Wagholi, Pune, India E-mail:
More informationSupervised Learning in Genre Classification
Supervised Learning in Genre Classification Introduction & Motivation Mohit Rajani and Luke Ekkizogloy {i.mohit,luke.ekkizogloy}@gmail.com Stanford University, CS229: Machine Learning, 2009 Now that music
More informationAutomatic Laughter Detection
Automatic Laughter Detection Mary Knox Final Project (EECS 94) knoxm@eecs.berkeley.edu December 1, 006 1 Introduction Laughter is a powerful cue in communication. It communicates to listeners the emotional
More informationAutomatic Rhythmic Notation from Single Voice Audio Sources
Automatic Rhythmic Notation from Single Voice Audio Sources Jack O Reilly, Shashwat Udit Introduction In this project we used machine learning technique to make estimations of rhythmic notation of a sung
More informationLyric-Based Music Mood Recognition
Lyric-Based Music Mood Recognition Emil Ian V. Ascalon, Rafael Cabredo De La Salle University Manila, Philippines emil.ascalon@yahoo.com, rafael.cabredo@dlsu.edu.ph Abstract: In psychology, emotion is
More informationFeatures for Audio and Music Classification
Features for Audio and Music Classification Martin F. McKinney and Jeroen Breebaart Auditory and Multisensory Perception, Digital Signal Processing Group Philips Research Laboratories Eindhoven, The Netherlands
More informationAN EMOTION MODEL FOR MUSIC USING BRAIN WAVES
AN EMOTION MODEL FOR MUSIC USING BRAIN WAVES Rafael Cabredo 1,2, Roberto Legaspi 1, Paul Salvador Inventado 1,2, and Masayuki Numao 1 1 Institute of Scientific and Industrial Research, Osaka University,
More informationQuality of Music Classification Systems: How to build the Reference?
Quality of Music Classification Systems: How to build the Reference? Janto Skowronek, Martin F. McKinney Digital Signal Processing Philips Research Laboratories Eindhoven {janto.skowronek,martin.mckinney}@philips.com
More informationMusic Mood. Sheng Xu, Albert Peyton, Ryan Bhular
Music Mood Sheng Xu, Albert Peyton, Ryan Bhular What is Music Mood A psychological & musical topic Human emotions conveyed in music can be comprehended from two aspects: Lyrics Music Factors that affect
More informationCapturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness
Capturing the Temporal Domain in Echonest Features for Improved Classification Effectiveness Alexander Schindler 1,2 and Andreas Rauber 1 1 Department of Software Technology and Interactive Systems Vienna
More informationMelody classification using patterns
Melody classification using patterns Darrell Conklin Department of Computing City University London United Kingdom conklin@city.ac.uk Abstract. A new method for symbolic music classification is proposed,
More informationSome Experiments in Humour Recognition Using the Italian Wikiquote Collection
Some Experiments in Humour Recognition Using the Italian Wikiquote Collection Davide Buscaldi and Paolo Rosso Dpto. de Sistemas Informáticos y Computación (DSIC), Universidad Politécnica de Valencia, Spain
More informationCreating a Feature Vector to Identify Similarity between MIDI Files
Creating a Feature Vector to Identify Similarity between MIDI Files Joseph Stroud 2017 Honors Thesis Advised by Sergio Alvarez Computer Science Department, Boston College 1 Abstract Today there are many
More informationAutomatic Music Genre Classification
Automatic Music Genre Classification Nathan YongHoon Kwon, SUNY Binghamton Ingrid Tchakoua, Jackson State University Matthew Pietrosanu, University of Alberta Freya Fu, Colorado State University Yue Wang,
More informationAutomatic Mood Detection of Music Audio Signals: An Overview
Automatic Mood Detection of Music Audio Signals: An Overview Sonal P.Sumare 1 Mr. D.G.Bhalke 2 1.(PG Student Department of Electronics and Telecommunication Rajarshi Shahu College of Engineering Pune)
More informationDiscovering Similar Music for Alpha Wave Music
Discovering Similar Music for Alpha Wave Music Yu-Lung Lo ( ), Chien-Yu Chiu, and Ta-Wei Chang Department of Information Management, Chaoyang University of Technology, 168, Jifeng E. Road, Wufeng District,
More informationMusic Information Retrieval
Music Information Retrieval Automatic genre classification from acoustic features DANIEL RÖNNOW and THEODOR TWETMAN Bachelor of Science Thesis Stockholm, Sweden 2012 Music Information Retrieval Automatic
More informationMusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface
MusCat: A Music Browser Featuring Abstract Pictures and Zooming User Interface 1st Author 1st author's affiliation 1st line of address 2nd line of address Telephone number, incl. country code 1st author's
More informationAutomatic Detection of Emotion in Music: Interaction with Emotionally Sensitive Machines
Automatic Detection of Emotion in Music: Interaction with Emotionally Sensitive Machines Cyril Laurier, Perfecto Herrera Music Technology Group Universitat Pompeu Fabra Barcelona, Spain {cyril.laurier,perfecto.herrera}@upf.edu
More informationThe Role of Time in Music Emotion Recognition
The Role of Time in Music Emotion Recognition Marcelo Caetano 1 and Frans Wiering 2 1 Institute of Computer Science, Foundation for Research and Technology - Hellas FORTH-ICS, Heraklion, Crete, Greece
More informationAutomatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting
Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting Dalwon Jang 1, Seungjae Lee 2, Jun Seok Lee 2, Minho Jin 1, Jin S. Seo 2, Sunil Lee 1 and Chang D. Yoo 1 1 Korea Advanced
More informationImproving Music Mood Annotation Using Polygonal Circular Regression. Isabelle Dufour B.Sc., University of Victoria, 2013
Improving Music Mood Annotation Using Polygonal Circular Regression by Isabelle Dufour B.Sc., University of Victoria, 2013 A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of
More informationTOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS
TOWARD UNDERSTANDING EXPRESSIVE PERCUSSION THROUGH CONTENT BASED ANALYSIS Matthew Prockup, Erik M. Schmidt, Jeffrey Scott, and Youngmoo E. Kim Music and Entertainment Technology Laboratory (MET-lab) Electrical
More informationIMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM
IMPROVING GENRE CLASSIFICATION BY COMBINATION OF AUDIO AND SYMBOLIC DESCRIPTORS USING A TRANSCRIPTION SYSTEM Thomas Lidy, Andreas Rauber Vienna University of Technology, Austria Department of Software
More informationPsychophysiological measures of emotional response to Romantic orchestral music and their musical and acoustic correlates
Psychophysiological measures of emotional response to Romantic orchestral music and their musical and acoustic correlates Konstantinos Trochidis, David Sears, Dieu-Ly Tran, Stephen McAdams CIRMMT, Department
More informationarxiv: v1 [cs.ai] 30 Nov 2016
Fusion of EEG and Musical Features in Continuous Music-emotion Recognition Nattapong Thammasan 1,*, Ken-ichi Fukui 2, and Masayuki Numao 2 1 Graduate school of Information Science and Technology, Osaka
More informationMELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS
MELODY ANALYSIS FOR PREDICTION OF THE EMOTIONS CONVEYED BY SINHALA SONGS M.G.W. Lakshitha, K.L. Jayaratne University of Colombo School of Computing, Sri Lanka. ABSTRACT: This paper describes our attempt
More informationMulti-label classification of emotions in music
Multi-label classification of emotions in music Alicja Wieczorkowska 1, Piotr Synak 1, and Zbigniew W. Raś 2,1 1 Polish-Japanese Institute of Information Technology, Koszykowa 86, 02-008 Warsaw, Poland
More informationMood Classification Using Lyrics and Audio: A Case-Study in Greek Music
Mood Classification Using Lyrics and Audio: A Case-Study in Greek Music Spyros Brilis, Evagelia Gkatzou, Antonis Koursoumis, Karolos Talvis, Katia Kermanidis, Ioannis Karydis To cite this version: Spyros
More informationDetecting Musical Key with Supervised Learning
Detecting Musical Key with Supervised Learning Robert Mahieu Department of Electrical Engineering Stanford University rmahieu@stanford.edu Abstract This paper proposes and tests performance of two different
More informationHomework 2 Key-finding algorithm
Homework 2 Key-finding algorithm Li Su Research Center for IT Innovation, Academia, Taiwan lisu@citi.sinica.edu.tw (You don t need any solid understanding about the musical key before doing this homework,
More informationAnalysing Musical Pieces Using harmony-analyser.org Tools
Analysing Musical Pieces Using harmony-analyser.org Tools Ladislav Maršík Dept. of Software Engineering, Faculty of Mathematics and Physics Charles University, Malostranské nám. 25, 118 00 Prague 1, Czech
More informationRecommending Music for Language Learning: The Problem of Singing Voice Intelligibility
Recommending Music for Language Learning: The Problem of Singing Voice Intelligibility Karim M. Ibrahim (M.Sc.,Nile University, Cairo, 2016) A THESIS SUBMITTED FOR THE DEGREE OF MASTER OF SCIENCE DEPARTMENT
More informationComputational Models of Music Similarity. Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST)
Computational Models of Music Similarity 1 Elias Pampalk National Institute for Advanced Industrial Science and Technology (AIST) Abstract The perceived similarity of two pieces of music is multi-dimensional,
More informationarxiv: v1 [cs.ir] 16 Jan 2019
It s Only Words And Words Are All I Have Manash Pratim Barman 1, Kavish Dahekar 2, Abhinav Anshuman 3, and Amit Awekar 4 1 Indian Institute of Information Technology, Guwahati 2 SAP Labs, Bengaluru 3 Dell
More informationLyrics Classification using Naive Bayes
Lyrics Classification using Naive Bayes Dalibor Bužić *, Jasminka Dobša ** * College for Information Technologies, Klaićeva 7, Zagreb, Croatia ** Faculty of Organization and Informatics, Pavlinska 2, Varaždin,
More informationMPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND
MPEG-7 AUDIO SPECTRUM BASIS AS A SIGNATURE OF VIOLIN SOUND Aleksander Kaminiarz, Ewa Łukasik Institute of Computing Science, Poznań University of Technology. Piotrowo 2, 60-965 Poznań, Poland e-mail: Ewa.Lukasik@cs.put.poznan.pl
More informationHOW COOL IS BEBOP JAZZ? SPONTANEOUS
HOW COOL IS BEBOP JAZZ? SPONTANEOUS CLUSTERING AND DECODING OF JAZZ MUSIC Antonio RODÀ *1, Edoardo DA LIO a, Maddalena MURARI b, Sergio CANAZZA a a Dept. of Information Engineering, University of Padova,
More informationWHEN LYRICS OUTPERFORM AUDIO FOR MUSIC MOOD CLASSIFICATION: A FEATURE ANALYSIS
WHEN LYRICS OUTPERFORM AUDIO FOR MUSIC MOOD CLASSIFICATION: A FEATURE ANALYSIS Xiao Hu J. Stephen Downie Graduate School of Library and Information Science University of Illinois at Urbana-Champaign xiaohu@illinois.edu
More informationWorld Academy of Science, Engineering and Technology International Journal of Computer and Information Engineering Vol:6, No:12, 2012
A method for Music Classification based on Perceived Mood Detection for Indian Bollywood Music Vallabha Hampiholi Abstract A lot of research has been done in the past decade in the field of audio content
More informationEVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION
EVALUATION OF FEATURE EXTRACTORS AND PSYCHO-ACOUSTIC TRANSFORMATIONS FOR MUSIC GENRE CLASSIFICATION Thomas Lidy Andreas Rauber Vienna University of Technology Department of Software Technology and Interactive
More informationMulti-modal Analysis for Person Type Classification in News Video
Multi-modal Analysis for Person Type Classification in News Video Jun Yang, Alexander G. Hauptmann School of Computer Science, Carnegie Mellon University, 5000 Forbes Ave, PA 15213, USA {juny, alex}@cs.cmu.edu,
More informationHidden Markov Model based dance recognition
Hidden Markov Model based dance recognition Dragutin Hrenek, Nenad Mikša, Robert Perica, Pavle Prentašić and Boris Trubić University of Zagreb, Faculty of Electrical Engineering and Computing Unska 3,
More informationImproving Frame Based Automatic Laughter Detection
Improving Frame Based Automatic Laughter Detection Mary Knox EE225D Class Project knoxm@eecs.berkeley.edu December 13, 2007 Abstract Laughter recognition is an underexplored area of research. My goal for
More informationHUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH
Proc. of the th Int. Conference on Digital Audio Effects (DAFx-), Hamburg, Germany, September -8, HUMAN PERCEPTION AND COMPUTER EXTRACTION OF MUSICAL BEAT STRENGTH George Tzanetakis, Georg Essl Computer
More informationA CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION
A CLASSIFICATION APPROACH TO MELODY TRANSCRIPTION Graham E. Poliner and Daniel P.W. Ellis LabROSA, Dept. of Electrical Engineering Columbia University, New York NY 127 USA {graham,dpwe}@ee.columbia.edu
More informationMUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC
12th International Society for Music Information Retrieval Conference (ISMIR 2011) MUSICAL MOODS: A MASS PARTICIPATION EXPERIMENT FOR AFFECTIVE CLASSIFICATION OF MUSIC Sam Davies, Penelope Allen, Mark
More informationSpeech To Song Classification
Speech To Song Classification Emily Graber Center for Computer Research in Music and Acoustics, Department of Music, Stanford University Abstract The speech to song illusion is a perceptual phenomenon
More informationTopics in Computer Music Instrument Identification. Ioanna Karydi
Topics in Computer Music Instrument Identification Ioanna Karydi Presentation overview What is instrument identification? Sound attributes & Timbre Human performance The ideal algorithm Selected approaches
More informationSpeech and Speaker Recognition for the Command of an Industrial Robot
Speech and Speaker Recognition for the Command of an Industrial Robot CLAUDIA MOISA*, HELGA SILAGHI*, ANDREI SILAGHI** *Dept. of Electric Drives and Automation University of Oradea University Street, nr.
More informationA PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES
12th International Society for Music Information Retrieval Conference (ISMIR 2011) A PERPLEXITY BASED COVER SONG MATCHING SYSTEM FOR SHORT LENGTH QUERIES Erdem Unal 1 Elaine Chew 2 Panayiotis Georgiou
More informationGRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM
19th European Signal Processing Conference (EUSIPCO 2011) Barcelona, Spain, August 29 - September 2, 2011 GRADIENT-BASED MUSICAL FEATURE EXTRACTION BASED ON SCALE-INVARIANT FEATURE TRANSFORM Tomoko Matsui
More informationPerceptual dimensions of short audio clips and corresponding timbre features
Perceptual dimensions of short audio clips and corresponding timbre features Jason Musil, Budr El-Nusairi, Daniel Müllensiefen Department of Psychology, Goldsmiths, University of London Question How do
More informationAutomatic Laughter Detection
Automatic Laughter Detection Mary Knox 1803707 knoxm@eecs.berkeley.edu December 1, 006 Abstract We built a system to automatically detect laughter from acoustic features of audio. To implement the system,
More informationAutomatic Musical Pattern Feature Extraction Using Convolutional Neural Network
Automatic Musical Pattern Feature Extraction Using Convolutional Neural Network Tom LH. Li, Antoni B. Chan and Andy HW. Chun Abstract Music genre classification has been a challenging yet promising task
More informationDAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval
DAY 1 Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval Jay LeBoeuf Imagine Research jay{at}imagine-research.com Rebecca
More informationCoimbra, Coimbra, Portugal Published online: 18 Apr To link to this article:
This article was downloaded by: [Professor Rui Pedro Paiva] On: 14 May 2015, At: 03:23 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office:
More informationAudio-Based Video Editing with Two-Channel Microphone
Audio-Based Video Editing with Two-Channel Microphone Tetsuya Takiguchi Organization of Advanced Science and Technology Kobe University, Japan takigu@kobe-u.ac.jp Yasuo Ariki Organization of Advanced Science
More informationAcoustic Scene Classification
Acoustic Scene Classification Marc-Christoph Gerasch Seminar Topics in Computer Music - Acoustic Scene Classification 6/24/2015 1 Outline Acoustic Scene Classification - definition History and state of
More informationMelody Retrieval On The Web
Melody Retrieval On The Web Thesis proposal for the degree of Master of Science at the Massachusetts Institute of Technology M.I.T Media Laboratory Fall 2000 Thesis supervisor: Barry Vercoe Professor,
More informationHIT SONG SCIENCE IS NOT YET A SCIENCE
HIT SONG SCIENCE IS NOT YET A SCIENCE François Pachet Sony CSL pachet@csl.sony.fr Pierre Roy Sony CSL roy@csl.sony.fr ABSTRACT We describe a large-scale experiment aiming at validating the hypothesis that
More informationPolyphonic Audio Matching for Score Following and Intelligent Audio Editors
Polyphonic Audio Matching for Score Following and Intelligent Audio Editors Roger B. Dannenberg and Ning Hu School of Computer Science, Carnegie Mellon University email: dannenberg@cs.cmu.edu, ninghu@cs.cmu.edu,
More informationDrum Stroke Computing: Multimodal Signal Processing for Drum Stroke Identification and Performance Metrics
Drum Stroke Computing: Multimodal Signal Processing for Drum Stroke Identification and Performance Metrics Jordan Hochenbaum 1, 2 New Zealand School of Music 1 PO Box 2332 Wellington 6140, New Zealand
More informationSpeech Recognition Combining MFCCs and Image Features
Speech Recognition Combining MFCCs and Image Featres S. Karlos from Department of Mathematics N. Fazakis from Department of Electrical and Compter Engineering K. Karanikola from Department of Mathematics
More informationAUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION
AUTOREGRESSIVE MFCC MODELS FOR GENRE CLASSIFICATION IMPROVED BY HARMONIC-PERCUSSION SEPARATION Halfdan Rump, Shigeki Miyabe, Emiru Tsunoo, Nobukata Ono, Shigeki Sagama The University of Tokyo, Graduate
More informationA Survey Of Mood-Based Music Classification
A Survey Of Mood-Based Music Classification Sachin Dhande 1, Bhavana Tiple 2 1 Department of Computer Engineering, MIT PUNE, Pune, India, 2 Department of Computer Engineering, MIT PUNE, Pune, India, Abstract
More informationMusic Recommendation from Song Sets
Music Recommendation from Song Sets Beth Logan Cambridge Research Laboratory HP Laboratories Cambridge HPL-2004-148 August 30, 2004* E-mail: Beth.Logan@hp.com music analysis, information retrieval, multimedia
More information