Raga Identification of Carnatic Music For Music Information Retrieval
Raga Identification of Carnatic Music For Music Information Retrieval
net/publication/228960849
CITATIONS READS
53 5,085
2 authors:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by T.V Geetha on 21 May 2014.
Abstract— In this work we propose a method to identify the ascending order of the arrangement of notes and
raga of a Carnatic music signal. Raga is analogous to melody in avarohanam is defined as the descending order
Western music but is much richer and stronger in its arrangement of notes. A raga is classified into melakarta
representation. The main motive behind Raga identification is raga (parent) and janya raga (child). The various notes in
that it can be used as a good basis for music information
retrieval of Carnatic music songs or Film songs based on
Carnatic music are S, R, G, M, P, D, N which can be
Carnatic music. The input polyphonic music signal is analyzed thought of as analogous to C, D, E, F, G, B, A. Every
and made to pass through a signal separation algorithm to note has a frequency associated with it as given in [2]. A
separate the instrument and the vocal signal. After extracting the melakarta Ragam will have all the seven swaras and a
vocal signal we segment the vocal signal using our proposed janya ragam will have atleast 5 of the seven swaras [1].
segmentation algorithm. Using our proposed singer Every song in Carnatic music has an associated Talam.
identification algorithm we determine the singer to know the There are essentially 175 talams as explained in [1]. In
fundamental frequency of the singer. The frequency Carnatic music any song will be of time duration which is
components of the signal are then determined and we map these an integer multiple of a Talam. A talam is analogous to
frequency components into the swara sequence and thereby
determine the Raga of the particular song, which could be used
the beat of European music and the seven basic talams is
to index the songs and further for retrieval based on the Raga. discussed in [2]. It indicates the pacing of the music and
the placement of syllables in the composition. In Carnatic
Index Terms—Carnatic music, Raga, Music information music, the singer indicates the talam using hand gestures
retrieval [2].
Usually songs, which are based on South Indian
I. INTRODUCTION Carnatic music, will belong to one of the several Ragas or
a combination of Ragas and one of the Talam. Hence it is
Music information retrieval is the need of the hour due a good idea to index songs, which are based on the
to the availability of large amount of music on the
Carnatic music characteristics using these components,
Internet. Music information retrieval can be thought of as Raga and Talam. Hence, in this work we have proposed
part of Multimedia information retrieval. A lot of work an algorithm to determine the ragam and talam of a
has been done in the other components of multimedia like
Carnatic music song and have proposed that these
text, video and the one that is yet to be fully developed is components could be used as features for indexing and
audio. Audio processing involves processing speech and later for retrieval.
music. In this paper we discuss music processing which
could be used as the basis of music information retrieval
3. EXISTING WORK
using music characteristics. This paper is organized as
follows: Section 2 discusses about the Characteristics of Literature work in Carnatic music retrieval is on a slow
Carnatic music, Section 3 discusses on Existing work in pace compared to Western music. Some work is being
Raga and melody identification, Section 4 discusses on done in Swara identification [2] and Singer identification
the proposed System Architecture, Section 5 discusses on [3] of Carnatic music. In Hindustani music work has been
Results and Evaluation and Section 6 concludes the done in identifying the Raga of Hindustani music [4]. In
paper. [4] the authors have created a HMM based on which they
have identified two ragas of Hindustani music. The
2. CHARACTERISTICS OF CARNATIC MUSIC fundamental difference between Hindustani Raga pattern
and Carnatic Raga pattern is that in Hindustani we have
Indian music is broadly classified into South Indian R1, R2 as against R1, R2, R3 in Carnatic. Similarly G, D,
Carnatic music and North Indian Hindustani music. Both N all has three distinct frequencies in Carnatic music as
the systems of music are rich in their own style and compared to two frequencies in Hindustani [5]. This
Carnatic music is much more complex in the way the reduces the confusion in identifying the distinct
notes are arranged and rendered [1]. Indian music in
frequencies in Hindustani music as compared to Carnatic
general is based on Raga and Talam. Raga and Talam can music. The authors have not used polyphonic music
be thought of as equivalent to melody and rhythm in signal and have assumed that the input music signal is a
Western music. Raga is much more complex than melody
voice only signal. The fundamental frequency of the
and scale in western music. Raga can be thought of as the signal was also assumed and based on these features the
sequential arrangement of notes that is capable of raga identification process was done for two Hindustani
invoking the emotion of a song. A note is called as a
ragas. On the western music aspect, melody retrieval is
Swara in Carnatic music. Raga is characterized by being performed by researchers. The one proposed by [6]
Arohanam and Avarohanam. Arohanam defines the
571
After obtaining the talam associated with a segment, The fundamental frequency normally corresponds to S
the segment is divided into individual talam segments by and all other swaras of R, G, M, P, D, and N will depend
observing the time duration in the Talam database. Any on this S. The ratio of the other swaras relative to S is
Talam has a fixed predetermined component pattern. listed in Table I. Using literature on Carnatic music and
Each component depending on whether it is a count – ‘|’, normal singers singing pattern we have formulated a table
thattu ‘0’, veech ‘0’, has varying duration. After of singers with their corresponding fundamental
identifying the Talam in the first level of segmentation frequency. This is populated in the singer’s database and
the pattern of the identified Talam is known. Using this the fundamental frequency of the singers is retrieved for
feature the segmented signal is further segmented at each Swara determination. Using Table I and the extracted
count, Thattu or Veechu of the given Talam. At this point frequency components raga identification is performed.
of segmentation we have individual segments which
correspond to the Thattu, Veechu or count. In Carnatic 4.4 Frequency mapping and Raga determination
music, for a count, Thattu, Veechu there can be one, two, The swara is determined by identifying the singer’s
or four swaras depending on the tempo in which the song fundamental frequency and the dominant energy based
is sung. That is, each component of a talam obtained after frequency in each segment. Using the
the second level of segmentation can be realized as one, the ratio available in Table I the swara corresponding to
two or four swaras. Hence each of our segments at this every frequency is determined using the identified
level of segmentation could correspond to either one, two dominant frequency in every segment. We then identify
or four swaras. This requires us to further segment the the individual swaras available and use these swaras and
signal. Hence we segment each talam component compare it with a database of Raga to identify the raga of
obtained as a result of second level segmentation into the given song. The Raga database is a table consisting of
four equal frames. This segmented signal is then used for the name of the raga, the arohanam, avarohanam of raga
the purpose of Raga identification. in the form of swara components. We use a simple string
matching to compare the identified swara pattern with the
4.3. Feature Extraction
Raga database and determine the Raga of the input
After identifying the smallest possible segment we Carnatic song.
assume that this could correspond to a swara. We then
use the HPS algorithm as explained in [12] to identify the 5. RESULTS & EVALUATION
frequency components in a single segment. In each
segment there is a possibility for the presence of more The input signal is sampled at 44.1 KHz. We consider
than one frequency component. In order to determine a the identification of three melakarta Ragams for the
frequency component corresponding to a swara in each purpose of evaluating this algorithm. For the purpose of
segment, we use energy as the basis for identifying the Raga identification we have considered songs sung by
distinct frequency component in each segment. After like Nithyasree, M.S.Subalakshmi, Balamuralikrishna
identifying the frequency component present in each and Ilayaraja. From music knowledge we have identified
segment, we determine the one with the highest energy as their fundamental frequencies as 400 Hz, 320 Hz, 400 Hz
the one that corresponds to the swara. We now have to and 240 Hz respectively. We considered songs that
map this frequency component to the corresponding belong to melakarta ragas, Sankarabharanam, Kanakangi,
swara. In order to perform this mapping, we need to Karaharapriya ragam and belonging to Adi talam or
identify the fundamental frequency of the signal. The Roopaka Talam. We made the signal to pass through the
fundamental frequency is identified by performing the signal separation algorithm, and segmentation algorithm.
process of singer identification. The result showing the segmentation points for one input
We extract features pertaining to identifying the is shown in Fig 2. This is the first level of segmentation
singer of the song as explained in [3]. The singer of the where the protruding lines indicate the points of
song is identified as explained in [3] and this is used as segmentation. After identifying the segmentation points
the basis for identifying the fundamental frequency of the we determine the frequency components using the HPS
song. The fundamental frequency is essential for Carnatic algorithm and tabulated the frequency values which have
music analysis, since a singer can use a frequency in the the dominant energy. The result is shown in Figure 3. The
range of 240 to 400 Hz as the fundamental frequency. distinct frequencies obtained for the signal separated song
in Figure 2 is 400 600 427 420 417 554 418 420 For the ragam Karharapriya, Nithyasree’s voice gave
418 625 623 494 416 417 418. Using the singer correct results for 6 songs out of the tested 10 songs. This
identification system, we determined that the singer is is because of the fact that, those songs had a better
Nithyasree for the considered sample and then identified frequency range compared to other songs and hence the
the ragam as Sankarabharanam. The details of the Raga result is wrong because of the assumption of the
identified based on singer and the number of songs for fundamental frequency. The same algorithm was run for
which the raga was identified is given in Figure 4 Hindustani music and we observed that the algorithm
gave correct results for two of the Hindustani Ragas.
A Since Hindustani music and Carnatic music’s note
m indication is different as explained earlier, some
p combination which is not present in Hinudstani music is
l also likely to occur which is to be omitted.
i
t 6. CONCLUSION & FUTURE
u
d This work is based on low level signal features
e for Raga identification. We have tested for three Ragas
belonging to two Talams sung by three singers. The
Time system has to be tested for different singers, all melakarta
Ragas and ten most frequently occurring talams. Our
system has to be improved with a Hidden markov model
Figure 2: Input signal separated and segmented
based approach where we combine the low level features
and HMM for better and accurate identification. This
system’s basic disadvantage is the assumption of
F fundamental frequency and hence determination of
r
e
fundamental frequency is our next task. The system has
q to be modified to use the identified raga to index songs
u and use raga as a query for music information retrieval
e
n REFERENCES
c
y [1] Prof. Sambamurthy, South Indian Music, vol. 4, The Indian
Music Publishing House, Madras, India, 1982
[2] Rajeswari Sridhar and T.V. Geetha, “Swara identification of
Carnatic music”, IEEE Computer Society press, proceeding of
Figure 3: Frequency components ICIT 2006
[3] Rajeswari Sridhar and T.V. Geetha, “Music Information
Retrieval of Carnatic Songs Based on Carnatic Music Singer
Raga Identification Identification”, IEEE conference on Computer Science, ICCEE
12
2008
10 Tested
[4] Gaurav Pandey, et al “Tansen: A system for automatic Raga
identification” IICAI, 2003
No. of Songs
Balamuralikrishna
Nithyasree
Nithyasree
Ilayaraja
Nithyasree
2005.
[9] Mark R Every & John E Szymanski, “A Spectral Filtering
Sankarabharanam Kanakangi Karaharapriya
approach to music signal separation”, Proceedings of the 7th
Raga and Singers International Conference on Digital Audio Effects (DAFx ’04),
Naples, Italy, October, 2004, pp 197 – 200
[10] Rajeswari Sridhar and T.V. Geetha, “Comparison of signal
Figure 4: Raga identified based on singers separation algorithms as applied to Carnatic music”, FRSM
2007
For the singer Nithyasree, the number of songs for [11] Bello, J.P., Daudet, L., Abdallah, S., Duxbury, C., Davies,
which raga identified correctly was high compared to M., Sandler, M.B., "A Tutorial on Onset Detection in Music
M.S. Subbalakshmi and Balamuralikrishna because of the Signals", IEEE Transactions on Speech and Audio Processing,
fact that Nithyasree’s fundamental frequency, which we 2005
[12] Kyogu Lee, “Automatic chord recognition from audio
have assumed is more correct compared to other singers.
using enhanced pitch profile”, ICMC 2006