Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This...

141
Estimation of Pitch in Speech and Music Signals Using Modified Group Delay Functions A THESIS submitted by RAJEEV RAJAN for the award of the degree of DOCTOR OF PHILOSOPHY DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING INDIAN INSTITUTE OF TECHNOLOGY MADRAS. MARCH 2017

Transcript of Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This...

Page 1: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Estimation of Pitch in Speech and Music Signals Using

Modified Group Delay Functions

A THESIS

submitted by

RAJEEV RAJAN

for the award of the degree

of

DOCTOR OF PHILOSOPHY

DEPARTMENT OF COMPUTER SCIENCE AND

ENGINEERING

INDIAN INSTITUTE OF TECHNOLOGY MADRAS.

MARCH 2017

Page 2: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

THESIS CERTIFICATE

This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music Sig-

nals Using Modified Group Delay Functions”, submitted by Rajeev Rajan, to the

Indian Institute of Technology, Madras, for the award of the degree of Doctor of Phi-

losophy, is a bona fide record of the research work done by him under my supervision.

The contents of this thesis, in full or in parts, have not been submitted to any other

Institute or University for the award of any degree or diploma.

Prof. Hema A. Murthy

Research Guide

Professor

Dept. of Computer Science and Engineering

IIT-Madras, 600 036

Place: Chennai

Date:

Page 3: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

ACKNOWLEDGEMENTS

I would like to express my sincere gratitude to my guide Prof. Hema A. Murthy for

her constant support and guidance during the entire period of my research. She spent

enough time for discussions to teach me speech and music processing. She was always

helpful and a source of constant encouragement throughout my studies.

I would like to thank Heads of the department during my research period Prof. C.

Siv Ram Murthy, Prof. P. Sreenivasa Kumar, Prof. Krishna Moorthy Sivalingam

for providing excellent lab facilities to carry out the research work in the department.

I am thankful to my Doctoral Committee members Dr. D. Janakiram, Dr. Deepak

Khemani, Dr. Rajagopalan A.N, Dr. Sutanu Chakroborthi and Dr.Venktesh Ra-

maiyan for their valuable suggestions and feedback during the progress review meet-

ings. I take this opportunity to thank Dr. C. Chandrashekhar for his timely advice and

suggestions during various presentations in speech lab meetings and seminars. I would

like to thank AICTE and Directorate of Technical Education, Government of Kerala for

giving me an opportunity to study in IIT Madras.

All the members (past and present) of the DON Lab and MS Lab, were helpful

whenever I need assistance during my studies in the campus. I thank all my colleagues

in the lab, with whom I have interacted at some point: Padmanabhan, Abhijith, Golda,

Sarala, Srikanth, Subhadeep, Saad, Vignesh, Ashwin, Shrey Duta, Akshay, Cyriac,

Raghav, Manaswi, Jilt, Jom, Mari, Krishnaraj, Karthik, Saranya, Anusha, Jeena, Arun,

Anju, Manish, Nauman, Hari, Kasthuri, Gayathri and Mohana. I would like to thank

my friends in the institute, Sangeetha Jose, Preethi madam, Subhashini, Shajimohan,

Ajeesh Ramanujam, Jayesh, Manoj, Akesh, Rahul, Lio, Joe and Sam Sreenivas for

their constant support and guidance.

I also thank all the staff members in the Department of computer science and en-

gineering, IIT Madras and QIP office for their assistance and cooperation. I express

my gratitude to my wife, Beena for her encouragement and understanding throughout

my stay at IIT Madras. With great pleasure, I would like to give special thanks to my

parents, sister, father-, mother-, sisters- and brothers-in-law, for their support, prayers,

i

Page 4: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

patience and encouragement throughout my studies. I express my deep sense of love

to my kids, Nanma and Nathu. I thank GOD almighty for all the blessings, He has

bestowed upon me in my life.

Rajeev Rajan

ii

Page 5: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

ABSTRACT

KEYWORDS: melody, modified group delay, cepstral smoothing, multipitch,

source-group delay, genre, power spectrum, tracking

The estimation of fundamental frequency, or the pitch of constituent components of au-

dio signals has a wide range of applications in computational auditory scene analysis

(CASA), prosody analysis, source separation, speaker identification and music signal

processing. This thesis addresses pitch estimation in mixed speech and predominant

pitch estimation in polyphonic music. In the first task, predominant melodic pitch is ex-

tracted from music using modified group delay functions. Predominant pitch extraction

is performed using two different flavours of the modified group delay function. In the

first approach, modified group delay analysis is performed directly on the music signal.

Choice of the window for cepstral smoothing, and pitch range play a crucial role in

the accurate estimation of pitch. To overcome the stringent pitch range constraint in

the scheme, system characteristics are annihilated from the spectrum in the second ap-

proach, and the flattened spectrum is processed using modified group delay functions to

obtain an estimate of pitch. Multi-resolution analysis is employed to capture dynamic

variation of melody and dynamic programming is incorporated to ensure consistency in

pitch tracking. Automatic music genre classification using features computed from the

melodic contour is also explored.

In the second task, individual pitch tracks are estimated from co-channel speech

using modified group delay functions. When utterances of different speakers are com-

bined additively, and transmitted through a single channel, pitch cues of individual

sources will be weakened by the presence of mutual interference. In the proposed

work, flattened spectrum is analysed for predominant pitch in the first pass. Later, the

estimated pitch and its harmonics are annihilated from the spectrum, and the residual

spectrum is once again processed to obtain the next prominent pitch in the signal. The

individual pitch trajectories are refined in pitch grouping and post processing stage.

In the third task, a variant of source-group delay representation is introduced. The

iii

Page 6: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

flattened power spectrum is analyzed using group delay algorithm, followed by discrete

cosine transform to convert the source based modified group delay function to mean-

ingful features. The novel feature is effectively used for two applications, namely audio

indexing and estimation of number of speakers in mixed speech.

iv

Page 7: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

TABLE OF CONTENTS

ACKNOWLEDGEMENTS i

ABSTRACT iii

LIST OF TABLES viii

LIST OF FIGURES xi

ABBREVIATIONS xii

NOTATION xiii

1 INTRODUCTION 1

1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.2 Objective of the thesis . . . . . . . . . . . . . . . . . . . . . . . . . 2

1.3 Organization of the thesis . . . . . . . . . . . . . . . . . . . . . . . 2

1.4 Major contributions of the thesis . . . . . . . . . . . . . . . . . . . 3

2 Theory of Group Delay and Modified Group Delay Functions in Speech

Processing 4

2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4

2.2 Group delay in speech signal processing . . . . . . . . . . . . . . . 4

2.2.1 Properties of group delay functions . . . . . . . . . . . . . 7

2.3 Modified group delay function (MODGD) . . . . . . . . . . . . . . 12

2.4 Modified group delay on flattened spectrum

MODGD (Source) . . . . . . . . . . . . . . . . . . . . . . . . . . 15

2.5 Group delay applications in speech and music . . . . . . . . . . . . 17

2.5.1 Pitch and formant estimation . . . . . . . . . . . . . . . . . 18

2.5.2 Speech and speaker recognition . . . . . . . . . . . . . . . 19

2.5.3 Speech synthesis . . . . . . . . . . . . . . . . . . . . . . . 20

2.5.4 Speech emotion recognition . . . . . . . . . . . . . . . . . 21

2.5.5 Voice activity detection . . . . . . . . . . . . . . . . . . . . 21

2.5.6 Chirp group delay processing and ASR . . . . . . . . . . . 22

2.5.7 Tonic identification . . . . . . . . . . . . . . . . . . . . . . 22

2.5.8 Onset detection . . . . . . . . . . . . . . . . . . . . . . . . 23

3 Melody Extraction in Polyphonic Music 24

v

Page 8: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

3.2 Music information retrieval (MIR) . . . . . . . . . . . . . . . . . . 24

3.3 Melodic pitch estimation . . . . . . . . . . . . . . . . . . . . . . . 26

3.4 Melody extraction approaches . . . . . . . . . . . . . . . . . . . . 29

3.4.1 Salience based approaches . . . . . . . . . . . . . . . . . . 30

3.4.2 Source separation based approaches . . . . . . . . . . . . . 30

3.4.3 Salience and Source separation based approaches . . . . . . 31

3.4.4 Data driven approaches . . . . . . . . . . . . . . . . . . . . 32

3.4.5 Acoustic/Musicological model based approaches . . . . . . 32

3.5 Evaluation methodology . . . . . . . . . . . . . . . . . . . . . . . 33

3.6 Melody extraction: Challenges and Applications . . . . . . . . . . 34

3.7 Audio melody extraction from music using modified group delay func-

tions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

3.8 Melodic-pitch estimation using MODGD (Direct) . . . . . . . . . . 38

3.8.1 Theory of predominant melodic pitch estimation . . . . . . 38

3.8.2 Performance evaluation . . . . . . . . . . . . . . . . . . . . 40

3.9 Melody pitch extraction using flattened spectrum . . . . . . . . . . 43

3.9.1 Monopitch estimation using MODGD (Source) . . . . . . . 43

3.9.2 Melody pitch extraction using MODGD (Source) . . . . . . 44

3.9.3 Transient analysis using a Multi-resolution Framework . . . 47

3.9.4 Pitch tracking using dynamic programming . . . . . . . . . 48

3.9.5 Voice Activity Detection (VAD) . . . . . . . . . . . . . . . 51

3.9.6 Performance Evaluation . . . . . . . . . . . . . . . . . . . 52

3.9.7 Results, analysis and discussion . . . . . . . . . . . . . . . 53

3.10 Automatic music genre classification . . . . . . . . . . . . . . . . . 58

3.10.1 Proposed method . . . . . . . . . . . . . . . . . . . . . . . 59

3.10.2 Modified group delay feature (MODGDF) . . . . . . . . . . 60

3.10.3 Melodic Features . . . . . . . . . . . . . . . . . . . . . . . 60

3.10.4 Evaluation Dataset . . . . . . . . . . . . . . . . . . . . . . 63

3.10.5 Results and analysis . . . . . . . . . . . . . . . . . . . . . 64

4 Two-pitch Tracking in Co-channel Speech 66

4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

4.2 Multipitch Estimation . . . . . . . . . . . . . . . . . . . . . . . . . 66

4.3 Multipitch estimation algorithms : State-of-the-art approaches . . . 68

4.4 Evaluation methodology . . . . . . . . . . . . . . . . . . . . . . . 71

4.5 Applications of multipitch estimation . . . . . . . . . . . . . . . . 72

4.6 Modified group delay based two-pitch tracking in co-channel speech 73

4.6.1 Theory of pitch detection using modified group delay functions 74

4.6.2 Proposed system description . . . . . . . . . . . . . . . . . 75

vi

Page 9: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

4.6.3 Estimation of constituent pitches . . . . . . . . . . . . . . . 76

4.6.4 Pitch trajectory estimation by grouping . . . . . . . . . . . 79

4.6.5 Postprocessing . . . . . . . . . . . . . . . . . . . . . . . . 81

4.6.6 Pitch extraction in noisy and reverberant environment . . . . 83

4.7 Performance evaluation . . . . . . . . . . . . . . . . . . . . . . . 85

4.8 Results analysis and discussions . . . . . . . . . . . . . . . . . . . 87

4.9 Pitch estimation : more than two speakers . . . . . . . . . . . . . . 89

5 The Source-MODGD Coefficient Features (SMCF) in Speech/Music Anal-

ysis 92

5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92

5.2 Source-MODGD coefficient features(SMCF) . . . . . . . . . . . . 92

5.3 Estimating number of speakers in multipitch environment . . . . . . 94

5.3.1 Related work . . . . . . . . . . . . . . . . . . . . . . . . . 94

5.3.2 Proposed system . . . . . . . . . . . . . . . . . . . . . . . 95

5.3.3 Performance evaluation . . . . . . . . . . . . . . . . . . . 96

5.4 Speech/music classification and detection . . . . . . . . . . . . . . 96

5.4.1 Related work . . . . . . . . . . . . . . . . . . . . . . . . . 97

5.4.2 Speech/music classification . . . . . . . . . . . . . . . . . 97

5.4.3 Speech/music detection . . . . . . . . . . . . . . . . . . . . 98

6 Summary and Conclusions 102

6.1 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102

6.2 Criticism . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103

A Mel-frequency cepstral coefficients 105

B Melody extraction vamp plug-in 107

Page 10: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

LIST OF TABLES

2.1 Algorithm for the computation of MODGDF . . . . . . . . . . . . . 13

3.1 Music information retrieval and level of specificity . . . . . . . . . 26

3.2 Comparison of σe and RPA . . . . . . . . . . . . . . . . . . . . . . 42

3.3 Comparison of RPA and σe for MIR-1K dataset. . . . . . . . . . . . 42

3.4 Description of the datasets used for the evaluation . . . . . . . . . . 52

3.5 Comparison of results for ADC-2004 submitted in 2011/2012/2013/2014

evaluation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

3.6 Comparison of results for MIREX-2008 submitted in 2011/2012/2013/2014

evaluation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

3.7 Comparison of σe, RPA, RCA for LabROSA training dataset . . 56

3.8 Comparison of σe, RPA, RCA for Carnatic music dataset . . . . 57

3.9 Overall accuracy . . . . . . . . . . . . . . . . . . . . . . . . . . . 63

3.10 Classification accuracies obtained by existing approaches. . . . . . 63

3.11 Confusion matrix of fusion (MODGDF + melodic) experiment. . . . 65

4.1 Multipitch estimation-Summarization of few algorithms . . . . . . 71

4.2 Category of mixtures for Dataset:1 and 2 . . . . . . . . . . . . . . 85

4.3 Performance evaluation on Dataset-1 . . . . . . . . . . . . . . . . . 86

4.4 Performance evaluation on Dataset-2 . . . . . . . . . . . . . . . . . 86

4.5 Comparison of Efs corresponding to Accuracy10 (in Hz) (Dataset-1) 88

4.6 Comparison of Efs corresponding to Accuracy10 (in Hz) (Dataset-2) 88

4.7 Comparison of accuracy for different TMR : Database:1 . . . . . . 89

4.8 Comparison of accuracy for different TMR : Database:2 . . . . . . 89

4.9 Pitch estimation accuracy for three speaker mixed speech. . . . . . . 91

5.1 Confusion matrix for number of speaker estimation task. SP-1, SP-2

and SP-3 denote one speaker case, two speaker case and three speaker

case, respectively. Class wise accuracy is given as the last column entry.

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96

5.2 Confusion matrix for audio indexing task. Class wise accuracy is given

as the last column entry. . . . . . . . . . . . . . . . . . . . . . . . 98

5.3 Results on Speech/Music detection on audio recordings. Correctly in-

dexed speech/music frames in each recording is entered in %. Global

accuracy is given as the entry in the last column. . . . . . . . . . . 99

5.4 Performance comparison of Speech/Music detection in database, DB2 101

viii

Page 11: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

LIST OF FIGURES

2.1 (a) Fourier transform phase, (b) Unwrapped phase , (c) Group delay. 6

2.2 Illustration of additive property of the group delay functions. Three

signals x[n], h[n] and y[n], where y[n] = x[n] ∗ h[n] (upper pane);

magnitude spectra of three signals (middle pane); group delay functions

of three signals (the lower pane). . . . . . . . . . . . . . . . . . . . 7

2.3 Illustrating the high-resolution property for a two conjugate pole sys-

tem. The faster decay of group delay in comparison to magnitude spec-

trum can be seen. . . . . . . . . . . . . . . . . . . . . . . . . . . . 9

2.4 (a) Noisy composite signal, (b) Magnitude Spectrum, (c) Group delay

spectrum of (a), (d) Modified group delay spectrum of (a). . . . . . 9

2.5 (a) z-plane with two conjugate pole pairs, (b) magnitude spectrum of

the system shown in (a), (c) group delay spectrum of the system shown

in (a). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

2.6 (a) The z-plane with two conjugate pole pairs, (b) The group delay

spectrum of the system shown in (a), (c) The z-plane with two pole pairs

and zeros added, (d) the group delay spectrum of the system shown in

(c), (e) The z-plane with zeros pushed radially inward, (f) the group

delay spectrum of the system shown in (e). . . . . . . . . . . . . . . 13

2.7 (a) Frame of synthetic signal, (b) Magnitude spectrum, (c) Smoothened

spectrum, (d) Flattened spectrum. . . . . . . . . . . . . . . . . . . 16

2.8 Cepstrum method of smoothing . . . . . . . . . . . . . . . . . . . 16

2.9 Root cepstrum method . . . . . . . . . . . . . . . . . . . . . . . . 17

2.10 (a) A speech utterance, (b) First formant frequency estimated using

MODGD, (c) Pitch estimated using MODGD (Wavesurfer reference is

also given). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

3.1 Music information retrieval system . . . . . . . . . . . . . . . . . . 25

3.2 Melody extracted for a song . . . . . . . . . . . . . . . . . . . . . 27

3.3 Prominent melodic pitch (grey contour) on spectrogram. . . . . . . 27

3.4 Musical notation and its metadata representation for the song ‘Happy

Birthday to You’ (Source: https://en.wikipedia.org/wiki/HappyBirthdaytoYou) 28

3.5 (a) Time domain signal, (b) melodic contour (c) spectrogram . . . . 29

3.6 Melodic contours with different degrees of polyphony. (a) low, (b)

medium, and (c) high. (Melody is extracted using pYIN algorithm)

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

3.7 Octave errors in melodic pitch estimation . . . . . . . . . . . . . . 36

3.8 Two instances of melodic motif of raga Kamboji in an alapana . . . 37

3.9 (a) Cepstral window for cepstral smoothing, (b) Peaks correspond to

pitch and formants in MODGD plot. . . . . . . . . . . . . . . . . . 39

ix

Page 12: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

3.10 MODGD plot for a speech frame. stems indicate formant locations and

pitch. Wavesurfer (W.S) and estimated (Est) values are shown in boxes. 40

3.11 (a) MODGD plot for a frame of a music sample, (b) Melody pitch

extracted using MODGD (Direct). . . . . . . . . . . . . . . . . . . 41

3.12 (a) Frame of a speech, (b) Flattened power spectrum, (c) Peaks in the

MODGD feature space, (d) Pitch estimated for the entire utterance with

reference. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43

3.13 Overview of MODGD (Source) based melody pitch extraction . . . 44

3.14 (a) A frame of music signal, (b) Magnitude spectrum of (a), (c) Spectral

envelope of (a), (d) Flattened spectrum of (a). . . . . . . . . . . . . 45

3.15 Computation of optimal path by dynamic programming . . . . . . . 48

3.16 (a) MODGD plot for a frame, (b) Melody pitch extraction for opera

fem4.wav using MODGD (Source). . . . . . . . . . . . . . . . . . . 50

3.17 (a) Melody pitch extracted for North Indian classical audio segment,

(b) Melody pitch extracted for a Carnatic music segment. . . . . . . 52

3.18 Raw pitch accuracy of ADC-2004 dataset in box plot . . . . . . . . 53

3.19 Results on adaptive windowing on three window sizes . . . . . . . . 54

3.20 (a) RPA and RCA for DB1 [less accompaniments], (b) RPA and RCA

for DB2 [5-6 accompaniments], (W.S : Wavesurfer, pYIN : Probabilis-

tic YIN, MEL : MELODIA, MODGD : MODGD (Source) ) . . . . 57

3.21 Block diagram of the proposed system . . . . . . . . . . . . . . . 59

3.22 Spectrogram (top) and Modgdgram (bottom) for a song from jazz genre

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60

3.23 Kernel density estimate of melodic pitch for two genres. (a) Pop, (b)

Jazz. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62

3.24 Genre-wise classification accuracy for all the systems . . . . . . . . 64

4.1 Monopitch estimation algorithms fails at mixed speech example. Wavesurfer

pitch is plotted for a two-speaker mixed speech. Individual reference

pitches are also shown. . . . . . . . . . . . . . . . . . . . . . . . . 67

4.2 Iterative multipitch estimation (Klapuri, 2008) . . . . . . . . . . . . 69

4.3 Block diagram of the proposed method. Dotted block represents the

MODGD (Source) based predominant pitch estimation algorithm. . 76

4.4 Notch filter magnitude response and pole-zero plot . . . . . . . . . 78

4.5 (a) A frame of synthetic mixture, (b) Flattened spectrum of frame in (a), (c)

MODGD plot in the first pass. (d) MODGD of (a) after the application of

comb filter (second pass), (e) Pitch extracted for the mixed synthetic speech

segment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79

4.6 (a) MODGD plot for a frame in first pass and residual spectrum (dotted

line) for a real speech mixture, (b) Presence of stray values in the pitch

contour are marked in circle. . . . . . . . . . . . . . . . . . . . . . 80

4.7 (a) Harmonic energy of a segment, (b) Initial pitch estimates of the

segment in (a), (c) Final pitch trajectory in the post processing stage. 81

x

Page 13: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

4.8 (a) Initial pitch estimates for a speech mixture, (b) Final pitch trajecto-

ries estimated using the proposed algorithm, (c) Pitch trajectories esti-

mated using WWB algorithm. . . . . . . . . . . . . . . . . . . . . 83

4.9 Room impulse response . . . . . . . . . . . . . . . . . . . . . . . 85

4.10 Speakers with two closed pitch trajectories . . . . . . . . . . . . . 89

4.11 (a) Modified group delay function computed for a speech frame, (b)

Pitch contours computed for the speech segment. . . . . . . . . . . 90

4.12 Pitch contours computed for a natural speech mixture with three speak-

ers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90

5.1 (a) MODGD on flattened power spectrum of music frame, (b) MODGD

on flattened power spectrum of a speech frame . . . . . . . . . . . . 94

5.2 Two dimensional visualization of speech-music data with SMCF feature

using Sammon mapping . . . . . . . . . . . . . . . . . . . . . . . 98

5.3 Performance evaluation of an audio sample, 1 stands for speech and 2

stands for music . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100

B.1 Selection of plug-in on Sonic Visualiser . . . . . . . . . . . . . . . 108

B.2 Melodic sequences of proposed algorithm along with MELODIA refer-

ence. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109

xi

Page 14: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

ABBREVIATIONS

ADC Audio Description Contest

AMDF Average Magnitude Difference Function

ASR Automatic Speech Recognition

CASA Computational Auditory Scene Analysis

DCT Discrete Cosine Transform

GMM Gaussian Mixture Models

HMM Hidden Markov Model

HPSS Harmonic and Percussive Sound Separation

LPC Linear Predictive Coefficients

MFCC Mel-Frequency Cepstral Coefficients

MFDP Mel-Frequency Delta-Phase

MIR Music Information Retrieval

MODGD Modified Group Delay

MODGDF Modified Group Delay Feature

NCCF Normalized Cross Correlation Functions

NMF Non-negative Matrix Factorization

pYIN Probabilistic YIN

QBSH Query By Humming/Singing

RIR Room Impulse Response

SACF Summary Auto-Correlation Function

TMR Target-to-Masker Ratio

VAD Voice Activity Detection

xii

Page 15: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

NOTATION

σe Standard deviation of pitch detection

α,γ Modified group delay parameters

ω Angular frequency

τp Group delay derived from Fourier transform phase

τ Group delay

τm Modified group delay

f0 Fundamental frequency

ρ Autocorrelation coefficient

z z-transform

γ root-cepstrum parameter

ps Reference pitch

pp Estimated pitch

e Mean of pitch error

slen Signal length

tlen, hlen Hamming window parameters

δ Cepstral smoothing -smoothing parameter

σ Cepstral smoothing-roll of parameter

Pmin Minimum pitch

Pmax Maximum pitch

F (c) Value of pitch salience peak at c.Fmax Maximum value of pitch salience

ct Transition cost

cl Local cost

Fr(l) Reference frequency

Fe(l) Estimated frequency

µp Mean pitch height

xiii

Page 16: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

CHAPTER 1

INTRODUCTION

Fundamental frequency (fo) estimation continues to be a topical area of research in

speech and music technology domains. The fundamental frequency is usually the low-

est frequency component or partial present in the signal. Research in this area goes

under the name of pitch detection. Choosing an fo estimator for a speech/song discrim-

ination is a difficult task because detectors that work well for music may give degraded

performance in speech and vice versa (Gerhard, 2003). In this thesis, mainly three tasks

are addressed using group delay analysis. In the first task, predominant melodic pitch

is estimated in polyphonic music using modified group delay functions. Two schemes,

based on modified group delay functions are proposed to extract predominant melodic

pitch sequence from polyphonic music. Secondly, an algorithm for two-pitch track-

ing in co-channel speech is proposed and performed on a number of speech mixtures.

Thirdly, a variant of group delay feature is derived and is effectively used for estimating

number of speakers in mixed speech and speech/music detection.

1.1 Motivation

From the multicultural and multidisciplinary aspects of music, it is obvious that the

challenges facing music information retrieval (MIR) research and development are far

from trivial (Downie, 2004). Many difficulties still remain to be overcome before the

content-based MIR systems become reality. If it is query for a song or karaoke sepa-

ration from an orchestration, the user expects an efficient system naturally. Numerous

techniques are available to retrieve useful information from music. Group delay func-

tion, which is defined as the negative derivative of unwrapped Fourier transform phase

is the basis of the proposed tasks discussed in the thesis. The importance of phase

in speech perception and recognition has been already established (Murthy and Yeg-

nanarayana, 2011; Paliwal, 2003). The process of phase unwrapping; recovering the

original phase values from the principal values is a challenging task. The advantage

of group delay is that it can be computed directly from the signal. The competitive

Page 17: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

performance of modified group delay function in speech was the key motivation to

apply the same in predominant melodic pitch extraction in polyphonic music. Com-

putational methods replaced ‘searching and retrieval’ of songs based on “Meta-data”

from digital libraries. Automatic melody extraction is an inevitable step for query by

humming/singing (QBSH), music score following and computational music analysis.

The proposed method opens the door to a wide range of music processing applications.

1.2 Objective of the thesis

The objective is to study the usefulness of modified group delay function for pitch

extraction in speech and predominant melody extraction in music.

1.3 Organization of the thesis

• Chapter 2 reviews the theory of group delay functions and modified group delay

functions. Properties of group delay functions are also listed with suitable exam-

ples. Few applications of modified group delay functions in speech and music

signal processing are also discussed.

• Chapter 3 discusses the approaches, evaluation strategies in melody extraction

task followed by the proposed schemes. Two algorithms are described for pre-

dominant melodic pitch estimation. In the first approach, group delay analysis

is performed directly on music signal. In the second approach, the speech/music

signal is preprocessed to remove the effect of the vocal tract. This residual signal

(in the frequency domain) is then subjected to modified group delay analysis to

estimate the predominant pitch. The performance is evaluated using the standard

MIREX framework. Automatic music genre classification using fusion of high

level melodic features and low level features is also discussed.

• Chapter 4 discusses the proposed algorithm for pitch estimation in a two speaker

environment. The modified group delay based pitch extraction algorithm is used

to estimate the predominant pitch. Next, the predominant pitch and its harmonics

are eliminated using a comb filter. The residual signal is then subjected to group

delay analysis to estimate the pitch of the second speaker.

• Chapter 5 discusses the feature derived from the source group delay and its ap-

plications. The modified group delay feature is used in two speech and music

related tasks, namely 1) to determine the number of speakers in a speech mixture,

and 2) to distinguish between speech and music in a monophonic recording.

2

Page 18: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

1.4 Major contributions of the thesis

The major contributions of the thesis are:

• Two variants of modified group delay function based pitch extraction are pro-

posed.

• Modified group delay function based music genre classification is proposed.

• Two-pitch tracking in mixed speech is addressed using modified group delay

functions.

• Speech/music detection in a mixed audio recording using modified group delay

functions is performed.

3

Page 19: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

CHAPTER 2

Theory of Group Delay and Modified Group Delay

Functions in Speech Processing

2.1 Introduction

Group delay is a measure of time delay of the amplitude envelopes of the various si-

nusoidal components of a signal through a device under test, and is a function of the

frequency. Recent research (Yegnanarayana et al., 1984, 1991; Murthy and Yegna-

narayana, 1991b,a; Prasad et al., 2004; Rasipuram et al., 2008b; Dutta and Murthy,

2014; Gulati et al., 2014; Sebastian et al., 2017) has shown that group delay functions

are effective in processing various types of signals ranging for speech, music and brain

signals.

In this chapter, a review of group delay functions is presented. Section 2.2 discusses

group delay and its properties. Modified group delay functions are discussed in Section

2.3, followed by modified group delay analysis on flattened power spectrum in Section

2.4. Section 2.5 discusses few applications of modified group delay functions in speech

and music signal processing.

2.2 Group delay in speech signal processing

A complete representation of a signal in the Fourier domain requires both the Fourier

transform magnitude and Fourier transform phase. Earlier studies have already estab-

lished the significance of the short-term phase spectrum in speech processing applica-

tions. The resonances of the vocal tract manifest as peaks in the short-term magnitude

spectrum of the speech signal. These resonances often called formants, manifest as

transitions in the short-time phase spectrum. These transitions are masked due to the

phase wrapping at multiple of 2π as illustrated in Figure 2.1 (a). Hence any meaning-

ful use of the short-time phase spectrum for speech processing involves the process of

phase unwrapping (Figure 2.1 (b)). Group delay is defined as the negative derivative of

Page 20: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

the unwrapped Fourier transform phase. Group delay computed for a signal is shown in

2.1 (c). Mathematically the group delay function, τp(ejω) is given by,

τp(ejω) = −

d{arg(X(ejω))}

dω. (2.1)

where ω is the angular frequency, X(ejω) is the Fourier transform of the signal x[n]

and arg(X(ejω)) is the phase function. Similarly, a group delay function τm(ejω) de-

rived from Fourier transform magnitude function |X(ω)| can also be found in (Yegna-

narayana et al., 1984).

Consider a discrete time signal x[n]. Then

X(ejω) = |X(ejω)|ej arg(X(ejω)) (2.2)

Taking logarithm on both sides,

log X(ejω) = log (|X(ejω)|) + j{arg(X(ejω))} (2.3)

arg(X(ejω)) = Im[log X(ejω)]. (2.4)

where, Im corresponds to the imaginary part. The group delay function can be com-

puted by (Oppenheim and Schafer, 1990; Yegnanarayana et al., 1984),

τ(ejω) = −Imd(log(X(ejω)))

dω(2.5)

τ(ejω) =XR(e

jω)YR(ejω) + YI(e

jω)XI(ejω)

|X(ejω)|2(2.6)

where the subscripts R and I denote the real and imaginary parts, respectively. X(ejω)

and Y (ejω) are the Fourier transforms of x[n] and nx[n], respectively. |X(ejω)|2 in the

denominator, makes the group delay function noisy (Murthy, 1991). Hence group delay

functions are modified to eliminate the effects of this spikes to generate modified group

delay functions (Murthy, 1991; Hegde, Rajesh M., 2005).

Using Equation 2.2 (and assuming that the signal is minimum phase), it can also

be shown that (Oppenheim and Schafer, 1990; Yegnanarayana et al., 1984; Murthy and

Yegnanarayana, 2011)

5

Page 21: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

0 1000 2000 3000 4000 5000−5

0

5(a)

Frequency (Hz)

An

gle

(ra

dia

ns)

0 1000 2000 3000 4000 5000−1000

−500

0(b)

Frequency (Hz)

An

gle

(ra

dia

ns)

0 1000 2000 3000 4000 5000−0.2

0

0.2(c)

Frequency (Hz)

Gro

up

de

lay(s

ec)

Figure 2.1: (a) Fourier transform phase, (b) Unwrapped phase , (c) Group delay.

ln | X(ejω) |= c[0]/2 +∞∑

n=1

nc[n] cos nω (2.7)

and the unwrapped phase function is given by

arg(X(ejω)) = −∞∑

n=1

nc[n] sin nω (2.8)

where c[n] are the cepstral coefficients. Taking the negative derivative of Equation-

2.8 with respect to ω, we get

τ(ejω) = −∞∑

n=1

nc[n] cos nω (2.9)

From Equations 2.8 and 2.9, it is worth noting that, for a minimum phase signal, the

spectral phase and magnitude are related through the cepstral coefficients. Further, the

group delay function can be obtained as the Fourier transform of the weighted cepstrum.

6

Page 22: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

2.2.1 Properties of group delay functions

The group delay functions and their properties have been discussed extensively in

(Murthy and Yegnanarayana, 2011). The basic properties are listed below.

• Poles (Zeros) of the transfer function show as peaks (valleys) in the group delay

function.

• For a minimum phase signal

τp(ω) = τm(ω) (2.10)

• For maximum phase signal

τp(ω) = −τm(ω) (2.11)

• For a mixed phase signal

τp(ω) 6= τm(ω) (2.12)

50 100 150

−0.5

0

0.5

n

x[n]

50 100 150−2

−1

0

1

n

h[n]

50 100 150−4−2

024

n

y[n

]

200 400 6000

5

10

15

x 10−3

Frequency (Hz)

Mag

nitu

de

200 400 600

2468

x 10−3

Frequency (Hz)

Mag

nitu

de

200 400 6000

0.01

0.02

0.03

Frequency (Hz)

Mag

nitu

de

200 400 6000

5

10

15

x 10−3

Frequency (Hz)

Gro

up d

elay

(sec

)

200 400 60002468

x 10−3

Frequency (Hz)

Gro

up d

elay

(sec

)

200 400 600

05

1015

x 10−3

Frequency (Hz)

Gro

up d

elay

(sec

)

Figure 2.2: Illustration of additive property of the group delay functions. Three signals

x[n], h[n] and y[n], where y[n] = x[n] ∗ h[n] (upper pane); magnitude

spectra of three signals (middle pane); group delay functions of three signals

(the lower pane).

7

Page 23: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Additive property

The group delay function exhibits additive property. Convolution of signals in the time

domain is reflected as a summation in the group delay domain.

Consider a linear time invariant (LTI) system with impulse response given by

H(ejω) = H1(ejω)H2(e

jω) (2.13)

whereH1(ejω) andH1(e

jω) are the responses of the two resonators whose product gives

the overall system response. Taking absolute value on both sides we have,

|H(ejω)| = |H1(ejω)|.|H2(e

jω)| (2.14)

Using the additive property of the Fourier Transform phase,

arg(H(ejω)) = arg(H1(ejω)) + arg(H2(e

jω)) (2.15)

Then the overallcgroup delay function τh(ejω) is given by,

τh(ejω) = −

∂(arg(H1(ejω)))

∂ω−∂(arg(H2(e

jω)))

∂ω(2.16)

= τh1(ejω) + τh2(e

jω) (2.17)

Additive property can be observed in the example shown in Figure 2.2. Consider

two time domain signals x[n] , h[n] and the convolution of x[n] and h[n] (denoted by

y[n] ), shown in the upper pane in Figure 2.2. The middle pane in Figure 2.2 shows

corresponding magnitude spectra. The lower pane illustrates the additive property of

group delay functions.

High resolution property

The group delay function has a higher resolving power as compared to the magnitude

spectrum. The high-resolution property of group delay functions is illustrated in Fig-

ure 2.3 (Sebastian et al., 2015a). The faster decay of group delay in comparison to

magnitude spectrum, shows the high resolving power of group delay functions.

8

Page 24: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

100 200 300 400 5000

20

40

60

80

100

Frequency(Hz)

Am

plit

ud

e

magnitude spectrumGroup delay

Figure 2.3: Illustrating the high-resolution property for a two conjugate pole system.

The faster decay of group delay in comparison to magnitude spectrum can

be seen.

The high resolution property enables group delay function to estimate the frequency

components present in a composite signal. Figure 2.4 (a) shows a composite signal

comprising two frequency components. Figure 2.4 (b), (c) and (d) represent magnitude

spectrum, group delay and modified group delay spectrum for a frame, respectively.

Two prominent peaks are observed corresponding to the two components of the signal

in group delay spectrum ( Figure 2.4 (d)).

50 100 150 200 250 300

−0.20

0.20.40.6

Time(ms)

Ampl

itude

(a)

200 400 600 800 100002468

Frequency(Hz)

Ampl

itude

(b)

200 400 600 800 1000

−50

0

50

Frequency(Hz)

Gro

up d

elay

(ms)

(c)

200 400 600 800 10000

200

400

Frequency(Hz)

MO

DG

D(m

s)

(d)

Figure 2.4: (a) Noisy composite signal, (b) Magnitude Spectrum, (c) Group delay spec-

trum of (a), (d) Modified group delay spectrum of (a).

9

Page 25: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

A z-plane plot of a system consisting of two complex conjugate pole pairs is shown

in Figure 2.5 (a). The corresponding magnitude and group delay spectrum are shown

in Figure 2.5 (b) and 2.5 (c), respectively. Two peaks are well resolved in group delay

domain. The ability of the group delay function to resolve closely spaced formants in

the speech spectrum is explained in (Murthy and Yegnanarayana, 1991a). In (Kumar,

2015), examples of single-resonator and multi-resonator systems are considered and

the resolution of group delay and magnitude spectrum are studied. Using kurtosis and

spectral flatness as measures, it is shown that the group delay function exhibits greater

peakedness compared to that of the magnitude spectrum.

Consider a causal, discrete time signal x[n], whose Z−transform X(z) is given by a 4th

order all pole model defined by:

X(z) =1

(z − z∗0)(z − z0)(z − z∗1)(z − z1)(2.18)

where * indicates complex conjugation. zi = e−σi+jωi , where e−σ0 , e−σ1 determines

proximity of the root to unit circle.

X(ejω) =

1∏

i=0

1

(ejω − e−(σi−jωi))

1∏

i=0

1

(ejω − e−(σi+jωi)). (2.19)

The phase spectrum of the system defined by Equation 2.18 is given by,

θ(ω) = −

1∑

i=0

tan−1 sinω − e−σi sinωi

cosω − e−σi cosωi+

1∑

i=0

tan−1 sinω + e−σi sinωi

cosω − e−σi cosωi. (2.20)

Differentiating Equation 2.20 with respect ω, it can be shown that:

θ′

(ω) = θ′

1(ω) + θ′

2(ω) + θ′

3(ω) + θ′

4(ω)

τ(ω) = τ1(ω) + τ2(ω) + τ3(ω) + τ4(ω) (2.21)

where θ′

i corresponds to derivative of each of the terms in Equation 2.20, and τi(ω)

corresponds to the group delay functions of each of the first order poles, respectively.

For simplicity, we consider group delay function of the first two terms alone. For sake

10

Page 26: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

−5 −4 −3 −2 −1 0 1 2 3 4 5−1

0

1

4

Real Part

Imag

inary

Par

t

(a)

100 200 300 400 5000

0.01

0.02

0.03

Frequency(Hz)

Ampli

tude

(b)

100 200 300 400 500

05

1015

x 10−3

Frequency(Hz)

Grou

p de

lay (s

ec)

(c)

Figure 2.5: (a) z-plane with two conjugate pole pairs, (b) magnitude spectrum of the

system shown in (a), (c) group delay spectrum of the system shown in (a).

of argument, let σ1 = 3σ0

τ+(ω) = −1− e−σ0 cos(ω − ω0)

1 + e−2σ0 − 2e−σ0 cos(ω − ω0)+

−1− e−3σ0 cos(ω − ω1)

1 + e−6σ0 − 2e−3σ0 cos(ω − ω1). (2.22)

Taking the derivative of τ+(ω) and setting it to zero, we have

τ′

+(ω) =(e−σ0 − e−3σ0) sin(ω − ω0)

(1 + e−2σ0 − 2e−σ0 cos(ω − ω0))2+

(e−3σ0 − e−9σ0) sin(ω − ω1)

(1 + e−9σ0 − 2e−3σ0 cos(ω − ω1))2= 0 (2.23)

When σ0 = 0, Equation 2.23 becomes identically zero. This corresponds to zeros

or poles that lie exactly on the unit circle, for example, the everlasting sinusoids. At

ω = ω0, first term becomes zero, while the second term takes values that are <<< 1

owing to the factor (e−3σ0 − e−9σ0). Owing to the cos and (e−3σ0 − e−9σ0), the group

delay function is symmetric and decays fast about the resonant frequency ω0. A similar

observation can be made when ω = ω1, as long as σ0 > 0 and ω − ω0, ω − ω1 are less

than π. The narrow width around the resonant frequency and the relation between the

height and bandwidth of a resonance is referred to as the high resolution property.

11

Page 27: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

2.3 Modified group delay function (MODGD)

It has been shown that group delay functions can be used to accurately represent signal

information as long as the roots of the z-transform of the signal are not too close to the

unit circle in the z-plane (Madhumurthy and Yegnanarayana, 1989). The group delay

function exhibits large spikes due to zeros that are very near to the unit circle. For

speech signals, the zeros are introduced by pitch, nasals, the glottal return phase and

the short-time analysis. Hence, the group delay function has to be modified to eliminate

the effects of these spikes. The group delay function is modified by replacing |X(ejω)|

in the denominator by its cepstral smoothed envelope |S(ejω)|. The significance of

cepstral smoothing is explained in detail in (Hegde, Rajesh M., 2005). The Equation

2.6 is modified to obtain τc(ejω) as,

τc(ejω) =

XR(ejω)YR(e

jω) + YI(ejω)XI(e

jω)

|S(ejω)|2(2.24)

Since the peaks at the formant locations are very spiky in nature, parameters are

introduced to reduce the amplitude of these peaks and to restore the dynamic range

of the speech spectrum. The algorithm for computation of the modified group delay

function is described in (Hegde et al., 2007b) and it can be obtained as,

τm(ejω) = (

τ(ejω)

|τ(ejω)|)|τ(ejω)|α, (2.25)

where,

τ(ejω) =XR(e

jω)YR(ejω) + YI(e

jω)XI(ejω)

|S(ejω)|2γ. (2.26)

Two new parameters, α and γ are introduced to control the dynamic range of MODGD

such that 0 < α ≤ 1 and 0 < γ ≤ 1 (Murthy, 1991; Hegde, Rajesh M., 2005). The

algorithm is summarized in the Table 2.1. Modified group delay analysis performed

directly on the speech signal is henceforth denoted by MODGD (Direct).

The pole-zero plot of a system characterized by two poles and their complex con-

jugates is shown in Figure 2.6 (a). The corresponding group delay spectrum is shown

in Figure 2.6 (b). In Figure 2.6 (c), the pole-zero plot of the same system is shown

with zeros added uniformly in very close proximity to the unit circle. It is evident from

Figure 2.6 (d), that the group delay spectrum for such a system becomes very spiky and

12

Page 28: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

ill defined. In Figure 2.6 (e), all the zeros are manually moved (radially) into the unit

circle and the group delay function of such a system is recomputed. The group delay

spectrum of the system in Figure 2.6 (e) is shown in Figure 2.6 (f).

Table 2.1: Algorithm for the computation of MODGDF

Steps Algorithm

1 Let x[n] be the given sequence.

2 Compute the DFT X [k] , Y [k], of x[n] and nx[n], respectively.

3 Group delay function is τx[k] =XR[k]YR[k]+XI[k]YI [k]

|X[k]|2

R and I represents real and imaginary parts, respectively.

4 Compute modified group delay τc[k] =XR[k]YR[k]+XI [k]YI [k]

|S[k]|2,

where S[k] is the smoothed version of X [k]5 Two new parameters α and γ are introduced in the

equation to convert to modified group delay function, τm[k] by,

τm[k] =τ [k]|τ [k]|

|τ [k]|α

τ [k] = XR[k]YR[k]+XI [k]YI [k]|S[k]|2γ

−2 −1 0 1 2−1

0

1

4

Real Part

Imag

inar

y P

art

(a)

100 200 300 400 5000

5

10

15

x 10−3

Frequency(Hz)

Gro

up d

elay

(sec

) (b)

−2 −1 0 1 2−1

0

1

8

Real Part

Imag

inar

y P

art

(c)

100 200 300 400 500

−1

0

1

x 1014

Frequency(Hz)

Gro

up d

elay

(sec

) (d)

−2 −1 0 1 2−1

0

1

8

Real Part

Imag

inar

y P

art

(e)

100 200 300 400 500

05

1015

x 10−3

Frequency(Hz)

Gro

up d

elay

(sec

) (f)

Figure 2.6: (a) The z-plane with two conjugate pole pairs, (b) The group delay spec-

trum of the system shown in (a), (c) The z-plane with two pole pairs and

zeros added, (d) the group delay spectrum of the system shown in (c), (e)

The z-plane with zeros pushed radially inward, (f) the group delay spectrum

of the system shown in (e).

13

Page 29: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Modified group delay function as a feature (MODGDF)

The modified group delay function has a dimension that is equal to that of the order of

the discrete Fourier transform (DFT) chosen. The dimension is reduced by transforming

the MODGD spectrum via the discrete cosine transform (DCT) to the cepstral domain.

The DCT is primarily used as a data independent decorrelator (Yip and Rao, 1997).

DCT is used to convert the modified group delay spectra into cepstral features, c[n]

given by,

c[n] =

k=Nf∑

k=0

τm[k] cos(n(2k + 1)π

Nf) (2.27)

where Nf is the DFT order and τm[k] is the modified group delay spectrum. This is

known as modified group delay feature MODGDF. The velocity and acceleration param-

eters for the new group delay function are defined in the cepstral domain, in a manner

similar to that of the velocity and acceleration parameters for mel-frequency cepstral

coefficient (MFCC).

Modified group delay features are used extensively in speech recognition tasks

(Hegde et al., 2007a), and robust voice activity detection (Hari Krishnan P et al., 2006).

The complementarity of the group delay features with respect to other conventional

acoustic features is discussed in (Rasipuram et al., 2008b). In (Rasipuram et al., 2008b),

the joint features formed by concatenating spectral magnitude based MFCC and spec-

tral phase based MODGDF are used for recognizing phonetic unit classes and it has

shown improvement in performance compared to other combinations. In (Kumar et al.,

2010), automatic speech recognition using feature switching based on KL divergence is

proposed. A significant improvement in performance is observed over the joint feature

stream consisting of MFCC and MODGD.

Numerator of group delay functions

The characteristics of the vocal tract system can also be computed using a variant of

group delay functions, namely numerator of group delay (Yegnanarayana et al., 2011).

The idea of numerator of group delay is introduced in (Anand Joseph et al., 2006).

Group delay function can be expressed by (Anand Joseph et al., 2006),

τ(ejω) =XI(e

jω)X′

R(ejω)−XR(e

jω)X′

I(ejω)

XR(ejω)2 +XI(ejω)2. (2.28)

14

Page 30: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

where X(ejω) = XR(ejω) + jXI(e

jω) is the Fourier transform of the discrete-time

signal x[n], and X′

(ejω) = X′

R(ejω) + jX

I(ejω) is its derivative. The group delay

function of a signal around the resonant frequency is proportional to the square of the

magnitude of the Fourier transform, That is,

τ(ejω) ∝ |X(ejω)|2 (2.29)

If we ignore the denominator term, and consider the numerator term alone in the speech

signal, numerator of group delay function can be expressed by,

g(ejω) = XI(ejω)X

R(ejω)−XR(e

jω)X′

I(ejω) (2.30)

From Equation 2.28, we get,

g(ejω) = τ(ejω)|X(ejω)|2 (2.31)

Using the Expression 2.29, we can conclude that

g(ejω) ∝ |X(ejω)|4 (2.32)

Thus, the numerator of group delay shows sharper peaks near the resonances than τ(ejω)

and it has been successfully employed in formant estimation (Anand Joseph et al., 2006)

and automatic speech recognition (Zhu and Paliwal, 2004).

2.4 Modified group delay on flattened spectrum

MODGD (Source)

In the previous section, we have seen the group delay analysis on the speech signal

directly, MODGD (Direct). In this section, group delay on the flattened spectrum -

MODGD (Source) is explained. The power spectrum of the speech signal is flattened

using root cepstral smoothing (Murthy, 1994). A flat power spectrum is first obtained

by dividing the spectrum by its envelope. What remains are the picket fence harmon-

ics corresponding to that of the pitch of the source. This signal resembles a sinusoid

in noise. The modified group delay function computed for this signal is defined as

15

Page 31: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

0 0.005 0.01 0.015 0.02 0.025−1

−0.5

0

0.5

1

Time(secs)

Am

plitu

de

(a)

0 2000 4000 6000 8000−4

−2

0

2

4

Frequency(Hz)

log

ma

gn

itu

de

(b)

0 2000 4000 6000 8000−8

−6

−4

−2

0

2

4

Frequency(Hz)

log

ma

gn

itu

de

(c)

5000 6000 7000 8000

500

1000

1500

2000

Frequency(Hz)lo

g m

ag

nitu

de

(d)

1/To

Figure 2.7: (a) Frame of synthetic signal, (b) Magnitude spectrum, (c) Smoothened

spectrum, (d) Flattened spectrum.

MODGD (Source). The group delay spectrum produces peaks at multiples of pitch

period, T0. Figure 2.7 illustrates the process of flattening the spectrum for a synthetic

signal frame. It is observed that the harmonics due to pitch are clearly visible in the

flattened spectrum (Figure 2.7 (d)).

Cepstral smoothing

Cepstral smoothing approach for spectrum flattening is illustrated in Figure 2.8. In

the cepstral domain, the signal is decomposed into the spectral envelope (lower order

cepstral coefficients) and the spectral fine structure (higher order cepstral coefficients).

Speech is mainly represented by the lower order cepstral coefficients and a cepstral peak

in the upper order cepstral coefficients represents the pitch information.

x[n] DFTabs log IDFT

w[n]

DFT

wm[n]

Xm

Ym

Figure 2.8: Cepstrum method of smoothing

The smooth spectral envelope can be obtained by cepstral smoothing using the steps

16

Page 32: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

shown in Figure 2.8. The spectral envelope obtained by cepstral smoothing is,

Ym = DFT[w · DFT−1 log(|Xm|)︸ ︷︷ ︸

real cepstrum

] (2.33)

where w and Xm represent low pass lifter in the cepstral domain and DFT of x[n],

respectively. w can be expressed by,

w[n] =

1, |n| < nc

0.5, |n| = nc

0, |n| > nc

(2.34)

where, nc determines the liftering window length. The log magnitude spectrum of

X(ejω) is thus low pass filtered to obtain a smooth spectral envelope.

Root cepstrum method

The root cepstrum based method is a non-model based approach which is similar to

the conventional cepstrum approach, except that the logarithmic operation is replaced

by the | . |γ operation, with 0 < γ ≪ 1 (Murthy, 1994). The root cepstrum method is

shown in Figure 2.9. Root cepstral based smoothing is preferred as it not only helps

in capturing the dynamic range of the spectrum but also de-emphasizes the convolution

effects of the window function in the frequency domain.

x[n] IDFTDFTγ

w[n]

DFTX Ym

Figure 2.9: Root cepstrum method

2.5 Group delay applications in speech and music

The potential of group delay functions has been exploited in speech and music sig-

nal processing applications. Group delay function has already established its role in

pitch/formant estimation, voice activity detection, segmentation of speech into syllable

17

Page 33: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

boundaries, speech emotion detection and text-to-speech synthesis. The effectiveness

of segmentation of speech, and the features derived from the modified group delay

functions are demonstrated in applications such as language identification (Nagarajan

and Murthy, 2006), speech recognition (Hegde et al., 2007b) and speaker identifica-

tion (R.M.Hegde et al., 2004). MIR applications also employ group delay functions in

tonic identification, motif recognition and discovery, onset detection and audio melody

extraction tasks. Few applications are briefly discussed in the following section.

2.5.1 Pitch and formant estimation

Additive and high resolution properties of group delay functions are exploited in es-

timating system/source parameters. Group delay analysis on the flattened spectrum

can be used for pitch estimation (Murthy, 1991). The performance of the proposed

algorithm is superior to many algorithms especially, in noisy environment. Pitch and

first formant frequency computed for a speech utterance, using modified group delay

functions are shown in Figure 2.10. Wavesurfer (W.S-URL, 2012) references are also

given. The work proposed in (Sebastian et al., 2015b) explores the use of peakedness

and high resolution properties of the group delay functions and the ability of grating

compression transform (GCT) to smear harmonically related components in the spec-

trum to track pitch across frame. Modified group delay functions are converted to use-

ful cepstral features (MODGDF) and it is used effectively in speaker recognition and

verification tasks (Hegde, Rajesh M., 2005). In (Rao et al., 2007), the instants of signif-

icant excitation in speech signals are computed using Hilbert envelope and group delay

function. In the first phase, the approximate epoch locations are determined using the

Hilbert envelope of the linear prediction residual. Later, the accurate locations of the

instants of significant excitation are determined by computing the group delay around

the approximate epoch locations. A group delay based method for epoch extraction

from speech is proposed in (Murty and Yegnanarayana, 2008). The average slope of

the unwrapped phase of the short-time Fourier transform of LP residual is computed.

Instants, where the phase-slope function makes a positive zero-crossing are identified

as epochs. In (Murthy and Yegnanarayana, 1991.), a spectral root group delay function

approach is proposed for formant estimation. The algorithm is similar to the cepstral

smoothing approach for formant extraction using homomorphic deconvolution (Murthy

and Yegnanarayana, 2011).

18

Page 34: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1−1

−0.5

0

0.5

Time (sec)Am

plitud

e

(a)

20 30 40 50 60 70 80 90 100 1100

200

400

600

800

Frame Index

Frequ

ency

(Hz)

(b)

MODGDWavesurfer

20 30 40 50 60 70 80 90 100 110

100

150

200

250

300

350

Frame Index

Frequ

ency

(Hz)

(c)

MODGDWavesurfer

Figure 2.10: (a) A speech utterance, (b) First formant frequency estimated using

MODGD, (c) Pitch estimated using MODGD (Wavesurfer reference is

also given).

2.5.2 Speech and speaker recognition

Group delay functions are utilized in speech recognition task using two approaches

(Kumar and Murthy, 2009; Rasipuram et al., 2008a; Padmanabhan and Murthy, 2010;

Murthy and Yegnanarayana, 2011). In the first approach, segmentation algorithm is

used to segment the speech signal at syllable-level. These boundaries are used in a hid-

den Markov model (HMM) framework for two different tasks, namely, speech recogni-

tion and synthesis. In speech recognition, the boundaries are used to reduce the search

space for the language model, while in speech synthesis, the boundaries are used to ob-

tain accurate phoneme boundaries by restricting embedded re-estimation in the HMM

framework to syllable boundaries. In another approach, the modified group delay func-

tion is converted into features and subsequently used for speech recognition using ma-

chine learning algorithms. In (Sarada et al., 2009), a method is proposed to automati-

cally segment and label continuous speech signal into syllable-like units for Indian lan-

guages. In this approach, the continuous speech signal is first automatically segmented

into syllable-like units using group delay based algorithm. Similar syllable segments are

then grouped together using an unsupervised and incremental training (UIT) technique.

19

Page 35: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

The complementary nature of the MODGDF derived from the modified group delay

function with respect to features derived from the Fourier transform magnitude spectra

is illustrated with the help of extensive experiments in (Hegde, Rajesh M., 2005). The

work proposed in (Dey et al., 2011) explores the use of feature-switching framework in

speaker recognition task. The work shows the ability of MODGDF in capturing infor-

mation complementary to the mel-frequency cepstral coefficients and linear predictive

cepstral coefficients (LPCC). Sub-band approach proposed in (Thiruvaran et al., 2007)

to restrict the masking effect, shows a competitive performance in recognition task as

compared to other approaches. The group delay functions derived from parametric all-

pole models are also used in speaker identification (Rajan et al., 2013). In (Bastys et al.,

2010), phase spectrum of all-pole linear prediction model is used to derive features for

speaker identification. The results are reported using RUSsian speech dataBASE (RUS-

BASE) with competitive performance as compared to the features derived from the

power spectrum.

2.5.3 Speech synthesis

Concatenative speech synthesis using unit selection approach relies on a large database

of basic units. The idea behind unit selection synthesis is to select the best sequence of

speech units from all possible contexts from a database of speech units. Group delay

based segmentation has already successfully been employed for segmenting the speech

signal at syllable-level (Nagarajan et al., 2003). Segmenting the speech at syllable-level

is well suited for Indian languages, which are syllable-timed. The algorithm is modi-

fied using vowel onset point detection to reduce insertion and deletion errors. A labeling

tool is developed using this idea and was successfully implemented in six Indian Lan-

guages, namely, Tamil, Hindi, Bengali, Malayalam Telugu and Marathi (Deivapalan

et al., 2008). Group delay based segmentation of syllables has been employed for

building syllable based text-to-speech systems in (Pradhan et al., 2013) and (Vinodh

et al., 2010). An automatic algorithm for phoneme segmentation on HTS is proposed

in (Shanmugam and Murthy, 2014). In (Shanmugam and Murthy, 2014), group delay

based processing of short-time energy (STE) is used in tandem with HMM based forced

Viterbi algoithm to get accurate syllable boundaries. The group delay based boundaries

obtained in the vicinity of the HMM syllable boundaries are used as correct bound-

aries to reestimate the monophone HMM models, where the monophone HMMs are

20

Page 36: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

restricted to the syllable boundaries rather than the whole utterance. The reestimated

boundaries are again compared with the group delay boundaries and are corrected again

(Shanmugam and Murthy, 2014).

2.5.4 Speech emotion recognition

Nowadays, increasing attention has been directed to identify the emotional content of

a spoken utterance. It is being employed in numerous applications such as humanoid

robots, car industry, call centers, mobile communication and computer tutorial appli-

cations. Proper mapping of acoustic cues in terms of features has an important role in

speech emotion recognition systems. A novel feature based on the phase response of

an all-pole model of the vocal tract obtained from linear predictive coefficients (LPC)

is used in (Sethu et al., 2007). The performance of group delay features for depression

detection was also addressed (Ming et al., 2013). The experimental results demonstrate

the potential of mel-frequency delta-phase (MFDP) over MFCC. The feature switching

paradigm has been also implemented in speech emotion recognition systems (Dey et al.,

2011). The underlying principle behind the method is that some features are better at

discriminating some classes and other features for other classes. The early fusion of

MFCC and MODGDF has shown superior performance than individual feature set in

recognizing emotions, anger, boredom and anxiety. An information theoretic procedure

is described in the paper which can be used to determine the feature most suitable for a

given class.

2.5.5 Voice activity detection

Voice activity detection (VAD) also known as speech activity detection detects the pres-

ence or absence of human speech in an audio signal. VAD is an important enabling

technology for a variety of speech-based applications such as speech coding, speech

recognition, echo cancellation, speech enhancement and simultaneous voice/data ap-

plications. Voice activity detection using group delay is proposed in (Hari Krishnan P

et al., 2006). The speech regions of the signal are characterized by well-defined peaks in

the group delay spectrum, while the non-speech regions are identified by well-defined

valleys. Two modified group delay function based VAD algorithms, one using Gaus-

sian mixture models and the other using multilayer perceptrons are discussed in (Pad-

21

Page 37: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

manabahan, 2012). The evaluation of the proposed algorithms shows a superior per-

formance over standard VAD algorithms, with average error reduction of about 8%.

The proposed work in (Kemp, T. and Waibel, A. , 1999), investigates the joint use of

source and filter-based features in voice activity detection. A mutual information-based

assessment shows superior discrimination power of chirp group delay (CGD) of the

zero-phase signal in voice activity detection.

2.5.6 Chirp group delay processing and ASR

The chirp group delay functions, in which group delay is computed on a circle other

than the unit circle, are proposed in (Bozkurt et al., 2007). The negative derivative of

the phase spectrum computed from chirp z-transform is termed as chirp group delay

(Bozkurt et al., 2007). The usefulness of chirp group delay functions in feature extrac-

tion is demonstrated for automatic speech recognition (ASR) in (Bozkurt et al., 2007).

In (Jayesh and Ramalingam, 2014), a variant of chirp group delay analysis is proposed,

which is useful in better vocal tract estimation. Zeros outside the unit circle are first

reflected inside the unit circle. The chirp group delay function of the transformed sig-

nal is computed. It does not require knowledge of glottis closure instants (GCI) and is

much less sensitive to starting point of the analysis window and its duration.

2.5.7 Tonic identification

Group delay functions have also found many applications in music processing. Auto-

matic tonic identification in Carnatic music has a major role in performing data driven

computational melodic analysis (Bellur and Murthy, 2013). The tonic is the base pitch

chosen by a performer, which serves as a reference throughout a performance. Group

delay function is employed in the work proposed by Ashwin et al. (Bellur, 2013) to pro-

cess pitch histograms. Three methods, namely concert based method, template match-

ing and segmented histogram are proposed in (Bellur, 2013). In concert based method,

group delay histograms computed over all individual items are multiplied bin-wise to

one single histogram. The peak with maximum value in the histogram is mapped to

tonic pitch. The template matching method attempts to fit a template for every peak of

the histogram, each of which is treated as a candidate Sa. The best template match helps

to estimate the tonic pitch. In the third method, picking the global peak in the bin-wise

22

Page 38: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

product of the segmented group delay histogram results in tonic pitch. Amongst the

template matching and segmented histograms methods, template matching method is

found to perform better.

2.5.8 Onset detection

Bello et al. define an onset as the instant chosen to mark the transient (Bello et al.,

2005). Specifically in (Kumar et al., 2015), a method is proposed to identify stroke

instants in the case of percussive instruments. A novel algorithm for onset detection

which utilizes high spectral resolution of group delay functions is presented in (Ku-

mar et al., 2015). It employs the resemblance of Carnatic percussion instruments to a

generic amplitude-frequency modulated waveform. Group delay response of this signal

is treated as a detection function in this work. The algorithm is observed to perform

better on the mridangam which contains a considerable number of silent strokes. Two

improvisations are proposed; one removes the parameter dependency, and the other

extends group delay processing to a more generic onset detection algorithm. Andre

Holzapfel et al. introduce a novel approach to estimate onsets in musical signals using

the average of the group delay function (Holzapfel and Stylianou, 2008a). In this work,

onsets are detected by locating the positive zero-crossings of the phase slope function.

Summary

Group delay functions and the need for modified group delay functions are discussed

in this chapter. Two important properties of modified group delay functions; additive

property and high resolution property are explained with illustrations. Two variants

of modified group delay functions, MODGD (Direct) and MODGD (Source) are ex-

plained. Finally, a review on various applications of modified group delay function in

speech and music signal processing is given. These applications show the potential of

modified group delay functions in speech-music analysis and content based information

retrieval.

23

Page 39: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

CHAPTER 3

Melody Extraction in Polyphonic Music

3.1 Introduction

In many cultures, music is an important part of people’s way of life, as it plays a key

role in religious rituals, rite of passage ceremonies, social activities and cultural activi-

ties ranging from amateur karaoke singing to community choir.1 Music is an important

performance art form conveyed through sound and silence2. The common elements

of music include melodic pitch and rhythm. A variety of styles/genres of music have

evolved owing to the differences in these elements. Melody extraction is important for

music information retrieval tasks. In this chapter, the task of audio melody extraction,

evaluation framework and challenges are discussed. Two schemes for melody extrac-

tion, both based on modified group delay functions are proposed and evaluated on vari-

ous datasets. Section 3.2 introduces the concept of music information retrieval system.

The task of audio melody extraction is discussed in Section 3.3, followed by different

approaches in melody extraction in Section 3.4. The methodology used in the perfor-

mance evaluation of the algorithms is briefly discussed in Section 3.5. The challenges

in melody extraction from polyphonic music are described in Section 3.6. The moti-

vation for the use of modified group delay functions in audio melody extraction from

music is listed in Section 3.7. The first scheme for melody extraction based on modi-

fied group delay functions is discussed in Section 3.8 followed by a revised algorithm

in Section 3.9. Finally, the complementary nature of melodic feature is experimentally

shown using standard dataset for music genre classification in Section 3.10.

3.2 Music information retrieval (MIR)

MIR mainly focuses on the understanding and usefulness of music data through re-

search, development and application of computational approaches or tools. MIR in-

1https://en.wikipedia.org/wiki/Music2https://cobussen.com/publications-and-courses/

Page 40: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

cludes numerous tasks such as query-by humming (QBH), automatic genre classifica-

tion, singing voice/ instrument melody separation, audio melody extraction, instrument

recognition and automatic music transcription. There are two main approaches to MIR;

Query

Extract

Description

Match

Retrieve

Results

Description

Schema

MusicDescriptions

MusicDocuments

Figure 3.1: Music information retrieval system

metadata-based and content-based (Wiering, 2006). In the former, the issue is mainly

to find useful categories for describing music, distinguishing different recordings of the

same composition or to organize the many genres that exist across the globe. Music

information retrieval, when the music is associated with text tags, is a little easier than

music that is not tagged. Retrieval using only the audio content is still a topical problem

in MIR. Content-based MIR enables to find music that is similar to a set of features or

an example (Wiering, 2006). The main components of a content based MIR system are

shown in Figure 3.1. These components are query formation, description extraction,

matching and music document retrieval. Musical information systems can be classified

as per the scale of specificity (Casey et al., 2008). Systems with high specificity demand

exact matching of the query and the content retrieved. In systems with low specificity,

a query track will return tracks having little content directly in common with the query,

but with some global characteristics that match. Mid-specificity systems match high-

level music features, but do not match audio content. Table 3.1 enumerates some of the

MIR use cases and their specificities. The proposed work for audio melody extraction

which computes predominant melody from polyphonic music comes under the category

of mid-specific systems.

25

Page 41: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Table 3.1: Music information retrieval and level of specificity

Use Case Specificity Description

Music identification H Identify a compact disk, provide metadata about

an unknown track, mobile music information

retrieval

Plagiarism detection H Identify mis-attribution of musical perfor-

mances, mis-appropriation of music intellectual

property

Copyright monitor-

ing

H Monitor music broadcast for Copyright in-

fringement or royalty collection

Versions H/M Remixes, live vs. studio recordings, Cover

songs. Used for database normalization and

near-duplicate results elimination

Melody H/M Find works containing a melodic fragment

Identical work/title M Retrieve performances of same OPUS or song

title

Performer M Find music by a specific artist

Sounds like M Find music that sounds like a given recording

Performance align-

ment

M Mapping one performance onto another inde-

pendent of tempo and repetition structure

Composer M Find works by one composer

Recommendation M/L Find music that matches the user’s personal pro-

file

Mood L Find music using emotional concepts:Joy, En-

ergetic, Melancholy, Relaxing

Style/Genre L Find music that belongs to a generic category:

Jazz, Funk, Female vocal

Instrument (s) L Find works with same instrumentation

Music-Speech L Radio broadcast segmentation

3.3 Melodic pitch estimation

Conventional pitch extraction algorithms used in the context of speech fail in music due

to the presence of interfering partials. In music, unlike in speech, the pitch can span a

number of octaves. The extraction of prominent pitch from polyphonic music signals

has received substantial attention due to its relevance in music analysis. Prominent pitch

or melodic pitch is the pitch contour of the lead or dominant musical instrument in the

song. In a music with orchestration, there are several musical instruments playing along

with singing voice which result in strong harmonic interferences. Predominant melody

extraction becomes difficult owing not only to fast changes in pitch but also due to the

presence of multiple voices. The multiple voices may be heterophonic 1 as in Carnatic

music or polyphonic 2 as in Western music.

As per the widely accepted standard, melodic pitch is defined as “the single (mono-

phonic) pitch sequence that a listener might reproduce when asked to hum or whistle a

polyphonic piece of music” (Salamon and Gomez, 2012; Poliner et al., 2007). Histori-

1en.wikipedia.org/wiki/Heterophony2en.wikipedia.org/wiki/Polyphony

26

Page 42: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

200 400 600 800 1000 1200

50

100

150

200

250

300

350

400

450

Frame Index

Pitch

(Hz)

Figure 3.2: Melody extracted for a song

cally, the term was derived frommelos, which meant "song". In most songs, the melody

follows a logical, mathematical pattern that creates a memorable line of notes. Melody

extracted for a piece of music is shown in Figure 3.2. The importance of melody in mu-

sic is explained well in (Field, 1998). It says “It is melody that enables us to distinguish

one work from another. It is melody that human beings are innately able to reproduce

by singing, humming, and whistling. It is melody that makes music memorable: we are

likely to recall a tune long after we have forgotten its text”.

Time(ms)

Fre

quen

cy (

Hz)

Figure 3.3: Prominent melodic pitch (grey contour) on spectrogram.

In polyphonic music, there are usually many instruments playing at the same time

or with a lag. For example, different melodic sequences due to the accompaniments

can be seen in Figure 3.3 with prominent melody emphasized in grey colour. Generally

MIR community points to three definitions for melody (Salamon et al., 2014a),

• The f0 curve of the predominant melodic line drawn from a single source.

• The f0 curve of the predominant melodic line drawn from multiple sources.

• The f0 curves of all melodic lines drawn from multiple sources.

Definition 1 requires the choice of a lead instrument and gives the f0 curve for this

instrument. This annotation is the accepted standard in MIREX (Music Information

27

Page 43: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Figure 3.4: Musical notation and its metadata representation for the song ‘Happy Birth-

day to You’ (Source: https://en.wikipedia.org/wiki/HappyBirthdaytoYou)

Retrieval Evaluation eXchange) evaluations. Definition 2 expands on definition 1 by

allowing multiple instruments to contribute to melody. While a single lead instrument

need not be chosen, an indication of which instrument is predominant at each point in

time is required to resolve the f0 curve to a single point at each time frame. Definition

3 is the most complex, but requires special attention in retrieval process. In a solo,

the singing voice is the leading melody line. As per defintion 1, melody extraction

algorithm computes the pitch of singer as the prominent melody source. In fugue 3,

two melody lines are present. A fugue begins with the exposition of its theme in one

of the voices in the tonic key. After the statement of the theme, a second voice enters

and states the theme transposed to another key, which is known as the answer. As per

definition 2, the two voices contribute to the melody line for the entire performance. In

Carnatic music, harmonium is used as a secondary melody instrument for vocal music.

As per definition 3, two melody lines, one for the singer and the other for harmonium

is computed simultaneously throughout the performance.

Melodic information can be conveyed through various representations (Zatorre and

Baum, 2012). For example, consider a piece of music “Happy Birthday to You.”4. Five

different ways of representing this piece of music are illustrated in Figures 3.4 and 3.5.

Melodic notation and metadata are shown in Figure 3.4. The metadata tag contains

3https://en.wikipedia.org/wiki/Fugue4https://en.wikipedia.org/wiki/HappyBirthdaytoYou

28

Page 44: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

1 2 3 4 5 6−1

−0.5

0

0.5

Time (sec)

Am

plit

ud

e

(a)

100 200 300 400 500 6000

100

200

300

Frame Index

Pitch

(H

z)

(b)

(c)

Time (sec)

Fre

qu

en

cy(H

z)

Figure 3.5: (a) Time domain signal, (b) melodic contour (c) spectrogram

information such as artist, album title, track number, genre, lyrics and more. Figure 3.5

(a), (b), and (c) show the time domain representation, the pitch contour and spectrogram

representation, respectively for the same piece of music.

3.4 Melody extraction approaches

In the last two-three decades, audio melody extraction task has emerged as an active

research topic in music retrieval, and numerous algorithms for the same have been pro-

posed. Audio melody extraction is also known by several names such as predomi-

nant melody extraction, lead melody line computation and predominant fundamental

frequency estimation (Salamon et al., 2014a). Existing melody extraction algorithms

can be broadly classified into salience based approaches, source separation based ap-

proaches, salience and source separation based approaches, data-driven approaches and

acoustic/musicological model based approaches.

29

Page 45: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

3.4.1 Salience based approaches

A very successful approach to the task of melody extraction is salience based (Goto

and Hayamizu, 1999; Cao et al., 2007; Hsu and Jang, 2010a). Salience based methods

estimate potential f0 candidates per frame based on pitch salience and apply tracking

or transition rules to select the final melody line. In (Goto and Hayamizu, 1999), Goto

et al. focus on the estimation of predominant musical voice rather than the transcrip-

tion of all sound sources. The main steps of the melody extraction algorithm include

selection of a candidate set of fundamental frequencies based on spectral peaks, and

use of the subharmonic summation method to identify the fundamental frequency from

this candidate set. This method estimates the relative dominance of every possible f0

by maximum a posteriori probability (MAP) estimation. Cao et al. (Cao et al., 2007)

propose a melody extraction method based on the subharmonic summation spectrum

and the harmonic structure tracking strategy. They analyze the prominent pitch of the

mixture to find stable harmonic structure seeds which are used to estimate pitch. By

using the characteristics of vibrato and tremolo, Hsu and Jang (Hsu and Jang, 2010a)

distinguish the vocal partial from the music accompaniment partials.

In another approach 5, the music signal is first preprocessed and autocorrelation

function is used to form melodic fragments from the music. In the next step, these

fragments are evaluated against each other and finally the melodic path is constructed

such that it minimizes the steep changes in the tonal sequence. Salamon et al. propose

a pitch extraction algorithm that extracts the pitch even when the accompaniment is

strong (Salamon and Gomez, 2012). Pitch candidates are first estimated, and the context

is used to determine pitch contours. Features such as length, height, pitch deviation and

presence of vibrato are then used for framing rules for filtering non-melody segments.

3.4.2 Source separation based approaches

Another approach to melody extraction is, to use source separation methods (G.Hu and

D.L.Wang, 2010; Wang and Brown, 2006; Richard et al., 2008; Durrieu et al., 2010;

Tachibana et al., 2010). In source separation methods, the melody line is identified by

separating it from the accompaniments. Source separation methods include (Richard

et al., 2008), where an effort is made to separate the melody from polyphonic music.

5http://www.nyu.edu/classes/bello/ACA-files/4-periodicity.pdf

30

Page 46: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

In this approach, source/filter model of the singing voice and a back ground model

derived using non-negative matrix factorization (NMF) technique are used to estimate

the fundamental frequency of the singing melody. (Durrieu et al., 2010) propose a

new approach for the estimation and extraction of the main melody from polyphonic

audio, where the leading vocal part is explicitly represented by a specific source/filter

model. Two different representations are proposed for the source/filter model: a smooth

instantaneous mixture model (SIMM) and a smooth Gaussian scaled mixture model

(SGSMM). While the former represents the lead instrument as the instantaneous mix-

ture of all possible notes, the latter allows one source/filter couple to be active at any mo-

ment. The final melodic pitch sequence is obtained by Viterbi algorithm. The sustained

and percussive sounds are separated using harmonic and percussive sound separation

(HPSS) algorithm (Ono et al., 2008) in two passes in (Tachibana et al., 2010). From the

retained enhanced components, the melodic pitch sequence is obtained directly from

the spectrogram by finding the path which maximizes MAP of the frequency sequence.

3.4.3 Salience and Source separation based approaches

In this framework, source separation is first used as a preprocessing step to attenuate

the accompaniment and then a salient function is computed from the processed signal to

compute the melodic pitch sequence. Some of the methods which follow these steps can

be found in (Hsu and Jang, 2010b; Tzu et al., 2012; Hsu et al., 2011). In (Hsu and Jang,

2010b), sinusoid partials are first extracted from the musical audio signal. The vibrato

and tremolo information are then estimated for each partial. The vocal and instrument

partials are discriminated according to a given threshold, and the instrument partials are

then deleted. Instead of simply extracting the singing pitches by tracking the remain-

ing vocal partials, the normalized harmonic summation spectrum (NHSS) is used as a

salient function to track the potential candidates. Dynamic programming is later applied

to find the melodic pitch contour. Yeh et al. propose a salience based method after ac-

companiments pruning (Tzu et al., 2012). In this method, a harmonic/percussive sound

separation (HPSS) method is applied to suppress the energy produced by harmonic in-

struments. Later, sinusoidal partial extraction is performed by the multi-resolution fast

Fourier transform (MR-FFT). A grouping method is adopted to combine partials of the

main vocal and accompanying instruments. Then, vibrato and tremolo characteristics

are used to prune the instrument partials. Finally, pitch trend is determined from the of

31

Page 47: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

T-F (time-frequency) blocks using dynamic programming.

3.4.4 Data driven approaches

In (Poliner and Ellis, 2005; Ellis and Poliner, 2006; Bittner et al., 2005), a data driven

approach is used wherein the entire short-time magnitude spectrum is used as training

data for a support vector machine (SVM) classifier. The classifier is trained on thou-

sands of spectral slices for which the suitable melody is found out through manual or

human-corrected transcription of the original audio. In (Ellis and Poliner, 2006), a ma-

chine learning approach is proposed in which the system infers the correct melody label

based only on training with labeled examples. In the proposed algorithm, each note is

identified via SVM classifier trained directly from audio feature data, and the overall

melody sequence is smoothed via a hidden Markov model (HMM). In (Bittner et al.,

2005), the system uses pitch contours and contour features generated by MELODIA

(Salamon and Gomez, 2012) for the task. This method consists of a contour labeling

stage, a training stage where a discriminative classifier is used that distinguishes be-

tween melodic and nonmelodic contours, and a decoding stage which generates a final

f0 sequence. A classifier is used to score the pitch contours and remove those below

a certain threshold. Viterbi decoding is adopted to get the final pitch trajectories. In

(Ishwar, 2014), a combination of the present state-of-the-art systems and timbral char-

acteristics of the various melodic sources is used to compute predominant pitch from

audio music recording.

3.4.5 Acoustic/Musicological model based approaches

Some of the other prominent methods for melody extraction are described in (Ryynanen

and Klapuri, 2008a; Rao and Shandilya, 2004). A method based on multiple f0 estima-

tion followed by acoustic and musicological modeling is proposed in (Ryynanen and

Klapuri, 2008a). The acoustic model consists of separate models for singing notes and

non-melody segments. The musicological model uses key estimation and note bigrams

to determine the transition probabilities between notes. In (Rao and Shandilya, 2004),

a pitch detection method based on a perceptual model is proposed to track voice pitch

in the presence of a strong percussive background.

32

Page 48: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

3.5 Evaluation methodology

Evaluation methodology is the same as that of MIREX evaluation standards. The per-

formance of the melody pitch extraction for voiced frames is evaluated based on the

following rule. The reference frequency of an unvoiced frame is considered to be 0

Hz. The estimated pitch of a voiced frame is considered correct, when it satisfies the

following condition:

| Fr(l)− Fe(l) |≤1

4tone ( 50 cents) (3.1)

where Fr(l) and Fe(l) denote the reference frequency and the estimated pitch frequency

on the lth frame, respectively. 6

Datasets

A music genre or sub-genre may be defined by the styles, the context, content and spirit

of the themes. The method of melody extraction for one genre may not be optimal for

a different genre. In MIREX evaluation, the performance of the algorithms is evaluated

on a diverse collection of music styles. The music corpus includes ADC-2004, MIREX-

05, MIREX-2008, LabROSA training set and MIR-1K dataset (Salamon and Gomez,

2012; Ramakrishnan et al., 2008).

Evaluation metrics

The performance of algorithms is conventionally evaluated using six metrics. Metrics

defined in (Poliner et al., 2007) are as follows:

• Voicing Recall Rate (VR): Proportion of frames labeled voiced in the ground

truth that are estimated as voiced by the algorithm.

• Voicing False Alarm Rate (VF): Proportion of frames labeled unvoiced in the

ground truth that are estimated as voiced by the algorithm.

• Raw Pitch Accuracy (RPA) : The ratio between the number of the correct pitch

frames in voiced segments and the number of all voiced frames.

6In the quarter tone scale, an octave is divided into 24 equal steps (equal temperament). In this scale,

the quarter tone is the smallest step.

33

Page 49: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

• Raw Chroma Accuracy (RCA): Same as raw pitch accuracy, except that both

the estimated and ground truth f0 sequences are mapped onto a single octave.

RCA is greater than or equal to RPA.

• Overall Accuracy (OA): This measure combines the performance of the pitch

estimation and voicing detection tasks to give an overall performance score for

the algorithm. It is defined as the proportion of frames (out of the entire piece)

correctly estimated by the algorithm, where for unvoiced frames, this means that

the algorithm labeled them as unvoiced, and for voiced frames, the algorithm not

only determined them correctly as voiced frames but also estimated f0 accurately.

• The Standard deviation of the pitch detection (σe) : It is defined as,

σe =

1

N

(ps − p′s)2 − e2 (3.2)

where ps is the standard pitch, p′s is the detected pitch,N is the number of correct

pitch frames and e is the mean of the fine pitch error. e is defined as:

e =1

N

N

(ps − p′s). (3.3)

The accuracy and standard deviation measure the quality of the pitch estimation

algorithms, while voicing recall and voicing false alarm refer to accuracy of the distinc-

tion between melody and non melody segments. Overall accuracy is the overall average

performance of the melody extraction algorithm.

3.6 Melody extraction: Challenges and Applications

There are a number of algorithms available in the literature for the melody extraction

task. Nevertheless, many challenges still remain in the design and evaluation strategy

(Salamon et al., 2014a). Most of the algorithms developed so far mainly rely on the

uniqueness of the human voice in audio melody extraction task. In an orchestration,

where leading melody line is an instrument with other accompaniments, main melody

line may be closer to the melody of accompanying instruments in many parts. In such

situations, the task of distinguishing melody from the accompaniment is a challenging

task. Another issue is the degree of polyphony in the evaluation material. Degree of

polyphony is defined as the amount of orchestration as music styles vary. The errors that

occur in melodic pitch estimation as the degree of polyphony increases are illustrated

34

Page 50: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

in Figure 3.6. In the figure, the melodic pitch estimated using pYIN algorithm (Mauch

and Dixon, 2014) along with references are shown for the same song with varying or-

chestration. As degree of polyphony increases, melody estimation error also increases.

Octave error minimization techniques adopted by algorithms also play an important

role in reducing spurious pitch estimates. Figure 3.7 shows the pitch contour of a music

segment. The reference pitch contour is shown in red. It is obvious that in the circled

part, the algorithm estimates pitch value as the double of the reference pitch values.

In (Ramakrishnan et al., 2008), raw chroma accuracy of the algorithm is high due to

occurrence of octave errors, which is attributed to biasing of the two way mismatch

(TWM) error function towards pitches that lie in the lower and middle range regions

of the f0 search space. In the work proposed by Jo et al. (Jo and Yoo, 2010), the

state transition is controlled by a factor called degree of melody line which makes the

algorithm robust to octave mismatch. In (Paiva et al., 2006), octave correction stage

is added to tackle errors that appear in the note elimination stage. The rule based post

processing implementation helps to minimize the difference between RPA and RCA in

the work proposed by Dressler (Dressler, 2011).

0 50 100 150 200 250 3000

200

400

Frame Index

Pitc

h (

Hz)

(a)

Ref

pYIN

0 50 100 150 200 250 3000

200

400

Frame Index

Pitc

h (

Hz)

(b)

Ref

pYIN

0 50 100 150 200 250 3000

200

400

Frame Index

Pitc

h (

Hz)

(c)

Ref

pYIN

Figure 3.6: Melodic contours with different degrees of polyphony. (a) low, (b) medium,

and (c) high. (Melody is extracted using pYIN algorithm)

Evaluation materials are required for assessing the performance of melody extrac-

tion algorithms. The duration of the audio clips, time offsets in the ground truth anno-

35

Page 51: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

tations and the size/musical content of the databases also pose issues in evaluating the

algorithmic performance of many systems. Audio clips currently used are too short to

predict performance on full songs. Similarly, time offset between the annotation and

the audio, can have a dramatic effect on the results (Salamon and Urbano, 2012).

720 730 740 750 760 770 7800

100

200

300

400

500

600

700

800

900

Frame Index

Pitc

h (H

z)

MODGDREF

Figure 3.7: Octave errors in melodic pitch estimation

Voicing detection stage always plays an important role in the overall performance.

Till date, most approaches focus primarily on raw pitch extraction segment of melody

extraction, and less so on the voicing decision. Currently, even the algorithms with

the most effective voicing detection methods obtain an average voicing false alarm rate

of more than 20%. The best result in melody-nonmelody segmentation in (Ellis and

Poliner, 2006), are mainly due to the machine learning approach in the algorithm, which

classifies dubious frames as non-melodic frames. The computational complexity of

the algorithms is also very crucial in real time applications. Analysis shows that, the

variation in computational time of various algorithms may be a hindrance in real time

user-interactive applications (E.Gomez et al., 2006).

Melody extraction has a wide variety of application ranging from industry to re-

search. It has already been shown that automatically extracted melodic pitch can be

used in music retrieval (Salamon et al., 2013; Song et al., 2002; Ryynanen and Klapuri,

2008b; Qin et al., 2011; Tralie and Bendich, 2015), music classification (Su, 2012; Sim-

sekli, 2010; Shan et al., 2002; Machado Rocha, 2011), music de-soloing (Jean-Marc and

Christophe Ris, 1999; Ryynanen et al., 2008; Li and Wang, 2007), music transcription

(Benetos et al., 2012; Bello et al., 2000; Gomez et al., 2012) and computational mu-

sicology (Pikrakis et al., 2012; Ishwar et al., 2013; Koduri et al., 2012). For example,

melodic motif spotting using melodic contour is illustrated in Figure 3.8. The melodic

36

Page 52: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

contour extracted for an alapana 7 is shown in Figure 3.8. In the figure, two instances

of the same melodic motif of raga Kamboji are circled in the melodic contour of the

song.

100 200 300 400 500 6000

100

200

300

400

Frame Index

Pitc

h (H

z)

100 200 300 400 500 6000

50

100

150

200

250

Frame Index

Pitc

h (H

z)Figure 3.8: Two instances of melodic motif of raga Kamboji in an alapana

3.7 Audio melody extraction from music using modified

group delay functions

The primary motivation for this work arises from the effectiveness of group delay pro-

cessing in speech and music (Yegnanarayana and Murthy, 1992; Rajan and Murthy,

2013; Vijayan et al., 2014). The audio melody extraction task consists of two tasks: (1)

pitch detection (to extract the main melody in the presence of strong accompaniment)

(2) voicing detection (to identify the presence or absence of the main melody). Most al-

gorithms in the MIR literature are based on processing the harmonic structure of pitch

in the Fourier transform magnitude spectrum for music (Salamon et al., 2014b). We

propose two different approaches, where group delay processing is employed.

In the first approach, the modified group delay function (used earlier for formant

extraction (Yegnanarayana et al., 1991)) is used for pitch extraction. The peak corre-

sponding to the predominant pitch is emphasized in the modified group delay spectrum

and is mapped to melodic pitch. This is explained in Section 3.8.

In the second approach, the starting point is the power spectrum of the signal. The

power spectrum is subjected to group delay processing exploiting the additive and high

7An alapana is a melodic improvisation within the permitted phraseology of a given melody.

37

Page 53: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

resolution properties of group delay functions. The power spectrum is first flattened.

The flattened spectrum appears to be periodic owing to the zeros due to pitch. This

signal resembles a sinusoid in noise. Modified group delay function based processing

of this signal is performed to estimate the frequency of the sinusoid which corresponds

to pitch. A multi-resolution frame work is incorporated to capture dynamic variation in

melody. Dynamic programming is employed to ensure pitch consistency across frames.

The removal of spurious pitch estimates in the post-processing step refines the pitch

sequences, thus resulting in a smooth pitch trajectory. This is explained in Section 3.9.

3.8 Melodic-pitch estimation using MODGD (Direct)

The first approach in melodic pitch computation employs MODGD analysis on the

music signal directly. In this approach, audio signals are frame-wise analyzed using

modified group delay function. The proper selection of cepstral window emphasizes

the peak corresponding to pitch in the group delay spectrum. The peaks in the range

[Pmin, Pmax] are subjected to consistency check using dynamic programming, where

Pmin and Pmax represent minimum and maximum pitch, respectively. The optimal path

obtained in the consistency check is mapped to predominant melodic pitch sequence.

Finally, the pitch sequence is smoothed using median filtering to get a refined pitch

contour in the ‘ post-processing ’ phase.

3.8.1 Theory of predominant melodic pitch estimation

A source-filter model for music is given by

X(z) = H(z)E(z) (3.4)

where E(z) corresponds to z-transform of the source and H(z) corresponds to that of

the system spectrum. Using additive property of group delay, the overall group delay

function, τx(ejω) becomes

τx(ejω) = τh(e

jω) + τe(ejω). (3.5)

38

Page 54: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

where τh(ejω), τe(e

jω) represent system group delay and source group delay, respec-

tively.

100 200 300 400 500

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

quefrency(samples)

Mag

nitu

de(a)

hlen

slen/2

500 1000 1500

2

4

6

8

10

12

14

Frequency(Hz)

MO

DG

D (

sec)

(b)

Peak corresponding topitch

Figure 3.9: (a) Cepstral window for cepstral smoothing, (b) Peaks correspond to pitch

and formants in MODGD plot.

In the MODGDF computation, smoothed spectral envelope is obtained by cepstral

smoothing. When the cepstral window is selected properly, a peak that corresponds to

the pitch can be seen in the modified group delay function. The choice of shape and

size of cepstral window is important. The purpose of the cepstral window is to ensure

that the poles corresponding to the source are emphasized in the modified group delay

spectrum. The source in speech or music is due to the glottal pulse. The glottal pulse

produces a double pole at the pitch frequency 8. For the customization of the cepstral

window, new parameters are introduced in the computation, namely δ (smoothing) and

σ (roll off), both varying from 0 to 1. For δ = 0, window size is small and for δ = 1,

window size is large (no smoothing). Hanning window coefficients with parameters (δ,

σ) for the window size slen are computed, using the following equations.

hlen = round[tlen(σ/2)] (3.6)

where

tlen = 2 round[slen(δ/2)]. (3.7)

For example, a cepstral window for smoothing the spectrum is shown in Figure 3.9 (a).

Figure 3.9 (b) shows the modified group delay spectrum for a frame of speech signal

using the customized cepstral smoothing. Another example along with a reference ob-

tained from Wavesurfer (W.S-URL, 2012) is shown in Figure 3.10. In Figure 3.10, pitch

peak obtained by the cepstral smoothing is also shown along with formant locations for

8http://www.phonetik.unimuenchen.de/studium/

39

Page 55: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

a frame. Pitch frequency (232 Hz) and formant frequencies (450 Hz, 2540 Hz, 3130

Hz) estimated using the proposed algorithm more or less coincide with the reference

computed using Wavesurfer algorithm.

500 1000 1500 2000 2500 30000

0.5

1

1.5

Frequency(Hz)

MO

DG

D(s

ec)

F3:W.S =3131 HzEst =3130 Hz

PitchW.S =228 HzEst =232 Hz

F1:W.S =451.5 HzEst =450.0 Hz

F2:W.S =2544 HzEst =2540 Hz

Figure 3.10: MODGD plot for a speech frame. stems indicate formant locations and

pitch. Wavesurfer (W.S) and estimated (Est) values are shown in boxes.

In the proposed algorithm, melodic pitch is computed by mapping the location of

the peak in the range [Pmin, Pmax] in MODGD spectrum to candidate frequency, that

ensures consistency. For example, a plot obtained by MODGD analysis for a frame of

music is shown in Figure 3.11 (a). The pitch contour obtained by the proposed approach

for a music segment is also shown in Figure 3.11 (b). In this approach, the search space

in the spectrum is limited by the first formant frequency. For high pitched voices, this

results in erroneous pitch estimate where a low formant can be confused for the pitch

frequency. Thus the range constraint sets a limit on the performance of the system.

3.8.2 Performance evaluation

The performance of the proposed algorithm is evaluated using ADC-2004 dataset, MIREX-

2008 dataset, Carnatic music dataset, LabROSA Training set and MIR-1K subset. The

evaluation is performed using the metrics RPA and standard deviation of the pitch de-

tection (σe).

The details of datasets are given below.

• ADC 2004 Dataset (Joo et al., 2010): 20 audio clips (PCM, 16 bit, 44.1 kHz)

consisting of daisy, jazz, opera, MIDI and pop genres. Four excerpts are avail-

able for every category, each roughly 20 seconds in duration. The corresponding

reference data is given at 5.8 ms steps.

40

Page 56: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

1000 2000 3000 4000 5000

1

2

3

4

5

6

7

8

9

Frequency(Hz)

MODG

D(se

c)

(a)

1640 1660 1680 1700 1720 1740 1760

450

500

550

600

650

700

750

800

Frame Index

Pitch

(Hz)

(b)

MODGD Pitch

MIREX Ref

Figure 3.11: (a) MODGD plot for a frame of a music sample, (b) Melody pitch ex-

tracted using MODGD (Direct).

• LabROSA training set (Tachibana et al., 2010): A referential dataset of MIREX

is provided by LabROSA of Columbia University. The data set contains 13 audio

files and ground truth f0 data for each audio file. The audio files are of CD-quality

(PCM, 16 bit, 44.1 kHz), monaural, 20-30 s length short clips.

• MIREX-2008 (Indian Database) (Ramakrishnan et al., 2008): 4 excerpts of 1 min

duration each from North Indian classical vocal performances. There are two

different mixtures of each of the 4 excerpts with different accompaniments corre-

sponding to a total of 8 audio clips. These performances consist of the voice, tabla

(pitched percussions), tanpura (Indian instrument, perpetual background drone)

and a secondary melodic instrument called the harmonium, very similar to the

accordion.

• Carnatic: Carnatic music is a sub-genre of Indian classical music. Carnatic music

usually, consists of a principal performer (mostly a vocalist), a melodic accompa-

niment (a violin), a rhythm accompaniment (a mridangam), and a tambura, that

acts as a drone throughout the performance. Fourteen Carnatic alapanas are used

for evaluation purpose. Usually a vocal alapana consists of a singer with violin

accompaniment all tuned to a specific tonic.

• MIR-1K (Hsu et al., 2012): One thousand song clips with the music accompani-

ment and the singing voice are recorded at left and right channels, respectively.

The duration of each clip ranges from 4 to 13 seconds, and the total length of the

41

Page 57: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

dataset is 133 minutes. These clips are extracted from 110 karaoke songs which

contain a mixture track and a music accompaniment track. These songs are freely

selected from 5000 Chinese pop songs and sung by 8 females and 11 males. Most

of the singers are amateurs and do not have professional music training. Manual

annotations of the dataset include pitch contours in semitone, indices and types

for unvoiced frames, lyrics, and vocal/non-vocal segments.

In MIR-1K dataset, the music accompaniment and the singing voice are available

in left and right channels separately. Singing voice track and accompaniment track are

mixed with two signal accompaniment ratio (0 dB and 5 dB). 0 dB means the accom-

paniment is as strong as the singing voice. The performance of the proposed system

is compared with that of YIN algorithm (Cheveigne and Kawahara, 2002.). The re-

sults are tabulated in Table 3.2 and Table 3.3. Raw pitch accuracy of 62.80%, 49.00%,

49.07% and 51.00% are reported for ADC, LabROSA, MIREX-2008, Carnatic dataset,

respectively. In terms of RPA, the performance of the proposed system is far better

than YIN algorithm for ADC, MIREX-2008 dataset but poorer in the case of Carnatic,

LabROSA and MIR-1K dataset. The standard deviation error is high as compared to

YIN algorithm. In the experiments performed, it is observed that the choice of range

[Pmin, Pmax] plays an important role in the performance of the system. Although pitch

estimation is accurate, the range is restrictive. It is observed that the proposed algorithm

fails to estimate pitch accurately when the pitch of the voice is high. To overcome such

stringent constraints, an alternative method based on the flattened spectrum is proposed

in the next section. e such stringent constraints in the pitch estimation.

Table 3.2: Comparison of σe and RPA

RPA σe

Method ADC LabROSA MIREX-2008 Carnatic ADC LabROSA MIREX-2008 Carnatic

YIN 50.30 54.81 43.49 64.21 3.70 3.10 3.65 2.94

MODGD (Direct) 62.80 49.00 49.07 51.00 3.89 3.37 3.80 3.43

Table 3.3: Comparison of RPA and σe for MIR-1K dataset.

Method RPA σe

0dB 5dB 0dB 5dB

YIN 68.2 81.5 2.8 2.7

MODGD (Direct) 43.1 57.7 3.2 3.3

42

Page 58: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

3.9 Melody pitch extraction using flattened spectrum

As discussed in Section 3.8, the proposed method results in erroneous pitch estimates

when the pitch is very close to the first formant. Since we are interested in source

information only, if the system information can be annihilated, and the residual signal is

subjected to MODGD analysis there is a possibility of obtaining better pitch estimates.

This is precisely what is attempted next. Modified group delay analysis on flattened

spectrum is denoted as MODGD (Source) and it is discussed below.

3.9.1 Monopitch estimation using MODGD (Source)

Pitch estimation in speech using modified group delay was first attempted in (Murthy,

1991). The power spectrum of the speech is flattened and the modified power spectrum

is analysed using MODGD algorithm. The process is illustrated in Figure 3.12. Figure

3.12 (a) shows a frame of speech. Figure 3.12 (b) and Figure 3.12 (c) represent flat-

tened spectrum and MODGD obtained on the flattened spectrum, respectively. Pitch

estimated for an entire speech utterance using MODGD (Source) is shown in Figure

3.12 (d) along with the references. The performance is at par with any of the state-of-

the-art algorithm.

0 0.01 0.02 0.03 0.04−1

−0.5

0

0.5

1

Time(sec)

Am

plit

ud

e

(a)

5 10 15

x 10−3

0.02

0.04

0.06

0.08

Time(sec)

MO

DG

D (

se

c)

(c)

1000 2000 3000 4000 50000

20

40

Frequency(Hz)

Am

plit

ud

e

(b)

0 50 100 150 2000

100

200

300

Frame Index

Pitch

(H

z)

(d)

Ref

MODGD

Figure 3.12: (a) Frame of a speech, (b) Flattened power spectrum, (c) Peaks in the

MODGD feature space, (d) Pitch estimated for the entire utterance with

reference.

43

Page 59: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

3.9.2 Melody pitch extraction using MODGD (Source)

In polyphonic music, strong overlap between partials of different instruments and vo-

cal is present. The underlying task is to estimate the predominant pitch sequence. A

schematic diagram of the proposed algorithm is presented in Figure 3.13. The power

spectrum is flattened by annihilating the spectral magnitude contribution of the system

(filter) characteristics using root cepstrum based smoothing. Root cepstral smoothing

is employed, as the objective is to capture the dynamic range of the signal accurately,

while at the same time not emphasising window artifacts. The process of spectral flat-

tening for a music frame is illustrated in Figure 3.14. A music frame is shown in Figure

3.14 (a). It’s magnitude spectrum and spectral envelope are shown in Figure 3.14 (b)

and (c), respectively. The picket fence harmonics in the flattened spectrum are clearly

shown in 3.14 (d).

Figure 3.13: Overview of MODGD (Source) based melody pitch extraction

In the proposed method, flattened spectrum is analysed using MODGD processing.

Adaptive windowing is used to capture the dynamic variation of melody. The autocorre-

lation function is used to determine the transient segments and a window size is chosen

accordingly. Salient peaks at multiples of pitch periods can be found in the MODGD

feature space. A block by block consistency check on the peak locations helps to se-

lect the track corresponding to the melodic pitch using dynamic programming. In the

post-processing stage, spurious pitch estimates are removed using median smoothing.

A voice activity detection based on harmonic energy is implemented to identify the

melodic segments. Assuming a source-system model for production of sounds in mu-

44

Page 60: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

0 0.01 0.02 0.03−0.5

0

0.5

Time(secs)A

mp

litu

de

(a)

0.5 1 1.5 2

x 104

−4

−2

0

2

4

Frequency(Hz)

log

Ma

gn

itu

de

(b)

0.5 1 1.5 2

x 104

−4

−2

0

2

4

Frequency(Hz)

log

Ma

gn

itu

de

(c)

0.5 1 1.5 2

x 104

0

5

10

Frequency(Hz)

Am

plit

ud

e

(d)

Figure 3.14: (a) A frame of music signal, (b) Magnitude spectrum of (a), (c) Spectral

envelope of (a), (d) Flattened spectrum of (a).

sic, melody in music corresponds to the periodicity of the predominant source. If the

timbral information can be suppressed, the picket fence harmonics are essentially pure

sinusoids (see Figure 3.14). Owing to the artifacts introduced by windowing, the sinu-

soids can be thought of as sinusoids in noise. Consider the z-transform of a monopitch

stream of impulses separated by To. Then,

E(z) = 1 + z−T0 + z−2T0 + ...+ ...z−nT0 + ... (3.8)

Assume that a frame contains at most three pitch period. The power spectrum of the

source is given by

E(z)E∗(z) = (1 + z−T0 + z−2T0 + z−3T0)(1 + zT0 + z2T0 + z3T0). (3.9)

Substituting z = ejω,

|E(ejω)|2 = (1 + cos(ωT0) + cos(ω2T0 + cos(ω3T0))2+

(sin(ωT0) + sin(ω2T0) + sin(ω3T0))2 (3.10)

|E(ejω)|2 = 4 + 6 cos(ωT0) + 4 cos(ω2T0) + 2 cos(ω3T0). (3.11)

45

Page 61: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

The root power spectrum is given by:

|E(ejω)|2γ

= |4 + 6 cos(ωT0) + 4 cos(ω2T0) + 2 cos(ω3T0)|γ (3.12)

where 0 < γ ≤ 1. The parameter γ controls the flatness of the spectrum. Restricting to

two impulses per frame and evaluating the power spectrum on the unit circle, we have

| E(ejω) |2γ

= 2γ(1 + cos(ωT0))γ. (3.13)

Clearly from the signal in Equation 3.12, it can be seen that the flattened power spectrum

becomes a sum of sinusoids that are integral multiples of 1T0

. When the dc component

is removed, the signal is similar to that of a noisy sinusoid. The task is to estimate the

frequency of the sinusoid in this signal. Replacing ω by n and T0 by ω0 in Equation

3.12 yields:

s[n] = a cos(nω0) + b cos(n2ω0) + c cos(n3ω0) + . . . , (3.14)

where a, b, c are constants. This signal is subjected to modified group delay processing

to obtain pitch salience function, which is used in subsequent steps to estimate melodic

pitch.

When the signal is polyphonic, E(z) becomes:

E(z) = 1 + z−T0 + z−2T0 + ...+ ...z−nT0 +

z−T1 + z−2T1 + ...+ ...z−nT1 +

z−T2 + z−2T2 + ...+ ...z−nT2 +

+z−Ti + z−2Ti + ... + z−nTi + ... (3.15)

where Ti corresponds to the pitch of each of the sources. Assuming at most two

pitch components per frame, in the group delay domain, additional peaks at nTi −mTj

are introduced, where Ti and Tj correspond to the pitches of two different sources and

n, m are integers. Nevertheless, the salience corresponding to the prominent pitch

is high. The additional pitch estimates due to polyphony are eliminated in dynamic

programming based post processing. The exponential decay of the group delay function

46

Page 62: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

at each of these composite pitches is still observed.

3.9.3 Transient analysis using a Multi-resolution Framework

An adaptive window to capture the dynamic variation of melody (Joo et al., 2010) in-

stead of a fixed window (Poliner et al., 2007) is used in the proposed method. Common

instances include abrupt changes in amplitude or frequency because of the variation in

energy input of the performer, attack regions followed by onsets of percussive sources,

fast transitions in frequencies due to expressive pitch variations, and noise due to en-

vironmental sounds (Thornburg, 2005). If the input signal changes pitch during an

analysis frame or if there is a significant transient, the resulting pitch measurement may

be wrong. The duration of the short analysis window does play an important role in

accurate pitch marking. A longer window provides high frequency resolution but leads

to poor temporal resolution. On the other hand, the use of large windows fails to pick

up the rapid pitch modulation that may occur during the transients. Thus, the proposed

algorithm uses a multi-resolution framework in which short windows are used for tran-

sient segments and long windows otherwise. Signal-driven window length adaptation

has been used extensively in audio coding algorithms for capturing the information in

stationary and transient audio segments (Painter and Spanias, 2000; Jones and Parks,

1990). Signal sparsity indicators such as Kurtosis measure, L2 norm, Gini index and

Spectral flatness are widely used to automatically adapt the window length (Rao et al.,

2012). The proposed algorithm uses correlation coefficient (Ref : Equation 3.16) com-

puted on consecutive frames at a lag (τ ), to switch between two windows. At zero lag,

the autocorrelation coefficient will be the largest. The choice of τ decides whether the

frame is a transient or not. A low autocorrelation coefficient value indicates a transient

segment and the window size is resized to a smaller length in such a case. The auto-

correlation coefficient ρ(X, τ, l) of a signal X at the lth frame is defined in (Joo et al.,

2010) as,

ρ(X, τ, l) =

k | X(k, l) || X(k, l + τ) |√∑

k | X(k, l) |2| X(k, l + τ) |2(3.16)

where, X(k, l) denotes the kth coefficient of the discrete Fourier transform of the lth

frame. τ corresponds to the autocorrelation lag. Discrete Fourier transform of two

frames at a lag (τ ) of 50 ms, has been empirically chosen for the experiments. Two

different window functions are chosen, one that is short and other that is long. In the

47

Page 63: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

proposed algorithm, a threshold of 0.98 on the autocorrelation coefficient is used as a

measure to switch between the two windows.

3.9.4 Pitch tracking using dynamic programming

The peak locations in the modified group delay functions are the possible pitch candi-

dates in each frame. Dynamic programming is then applied to these peaks to obtain

the optimal sequence of pitch candidates across frames. Dynamic programming, where

pitch salience is defined by the height of the group delay peak in tandem with transition

cost between successive frames is used for this purpose. Thus dynamic programming

combines two sources of information for pitch period marking. One source of infor-

mation is the ‘local’ information corresponding to amplitudes of peaks in the MODGD

feature space. The other source of information is the ‘transition’ information corre-

sponding to the relative closeness of the distance between peaks in two consecutive

frames in MODGD feature space. The optimal path is selected by minimizing the total

cost, which is the sum of local cost and transition cost.

Figure 3.15: Computation of optimal path by dynamic programming

The computation of local cost, transition cost and optimal path selection are dis-

cussed next (Veldhuis, 2000). Local cost function Cl(c) (pitch salience) of a pitch can-

didate c in the MODGD feature domain is computed by

Cl(c) = 1−F (c)

Fmax

(3.17)

48

Page 64: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

where F (c) is the value of the peak of the pitch candidate c and Fmax is the maxi-

mum value of the peak present in MODGD feature space for each frame. The transition

cost Ct(cj/cj−1) is the distance between the pitch candidates cj and cj−1 of consecutive

frames and is defined as

Ct(cj/cj−1) =| Lj − Lj−1 |

lmax(3.18)

where Lj , Lj−1 are peak locations in consecutive frames. lmax is the maximum tran-

sition distance between peaks, computed from the range [ Pmin, Pmax]. The transition

cost is normalized with the maximum possible transition distance (empirically chosen).

The dynamic programming algorithm finds an optimal pitch sequence (c1...cM ) with

candidates c1 in the first and cM in the M th frame in a block by minimizing the total

cost (TC) (Veldhuis, 2000).

Total cost (TC) = Local Cost+ Transition Cost. (3.19)

Total cost TC(c1...cM) of pitch candidates c1 to cM is computed by

TC(c1...cM) = Cl(c1) +

j=M∑

j=2

Ct(cj/cj−1) + Cl(cj). (3.20)

The optimal sequence of pitch markers is determined by back tracking from the can-

didate cM in the M th frame in a block to its starting frame while minimizing the total

cost.

Assume that c, d are the pitch candidates in two adjacent frames. Let {e1,e2 . . . ..ek}

be the successors of a candidate d for k frames. The minimum distance from the can-

didate d to its successors, Cmin(d) is computed by minimizing Equation 3.20 and is

denoted as,

Cmin(d) = mine1,...ek

C(e1, . . . .., ek, d). (3.21)

The optimal pitch sequence starting from candidate c followed by d in the consecutive

frame, with successors {e1,e2 . . . ..ek } is computed as

TCmin = C1(c) + min(Cmin(d) + Ct(c/d)). (3.22)

49

Page 65: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Figure 3.15 illustrates the process of finding the optimal path for a block of consec-

utive frames. Peak locations for four consecutive frames are shown in the left block

and corresponding pitch saliences are shown in the right block. Peak locations are the

positions of peaks in the MODGD feature space with a magnitude above a threshold.

Pitch salience is the corresponding local cost computed as defined in Equation 3.17.

Cost estimation starts from the last frame in the block i.e {73, 128, 178, 232}, with cor-

responding local cost {0.88, 0.85, 0.92, 0.89} and backtracks until the first frame of the

block. After estimating the cost for all possible paths, the path with minimum total cost

is chosen as the optimal path. Optimal path computed by minimizing the total cost is

highlighted in the block obtained after dynamic programming. The optimal path cho-

sen in the given example is {129, 124, 133, 128} highlighted in box. Block by block

dynamic programming across sequence of voiced frames results in the optimal path of

the pitch sequence by ensuring continuity.

The optimal path obtained from the dynamic programming stage is selected and a

search is performed over the MODGD feature space at multiples of the first optimal

path. Figure 3.16 (a) shows the peaks at multiples of pitch for a music frame. This

location information is also used to reinforce the value of pitch obtained.

1 1.5 2 2.5 3

0.005

0.01

0.015

0.02

0.025

Time (ms)

MODG

D (se

c)

(a)

Peaks at multiples of predominant pitch

2000 2050 2100

350

400

450

500

550

600

650

700

Frame Index

Pitch

(Hz)

(b)

REF PitchMODGD Pitch

Figure 3.16: (a) MODGD plot for a frame, (b) Melody pitch extraction for opera

fem4.wav using MODGD (Source).

50

Page 66: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Pitch halving/doubling is reduced considerably by context-framed rules. Instanta-

neous transitions in the intra-voiced regions are not permitted. If detected, it is replaced

by an interpolated sequence followed by a median smoothing to ensure pitch continuity

through a post processing stage. Median smoothing is performed to eliminate spurious

pitch estimates occurred in the process of pitch computation. The predominant pitch

extracted using the proposed method is compared with ground truth for audio of a fe-

male singer operafem4.wav in Figure 3.16 (b). Figures 3.17 (a) and 3.17 (b) show

plots of the melodic pitch contour for a North Indian classical excerpt and a Carnatic

music excerpt, respectively.

3.9.5 Voice Activity Detection (VAD)

Most melody extraction algorithms also include voicing detection using loudness thresh-

old (Brossier, 2005), adaptive threshold (Cancela, 2008) and harmonic energy (Ramakr-

ishnan et al., 2008) as criteria. HMM based detector with a 2-state HMM to decode the

mixture input into voiced and unvoiced segments is used in (Hsu et al., 2009). As most

approaches yield more or less the same performance, we have used the normalized har-

monic energy for voice activity detection. The frame-wise normalized harmonic energy

is sorted in ascending order and the value at 25% of size of the array is set as the thresh-

old. The harmonic energy is computed from the estimated fundamental frequency f0

and its harmonics up to 5 kHz (Ramakrishnan et al., 2008). Multiples of the fundamen-

tal frequency are identified by searching for the local maxima with 3% tolerance.

51

Page 67: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

800 1000 1200 1400 1600 1800 2000

50

100

150

200

250

300

Frame Index

Pitch

(a)

MODGD PitchREF

4200 4250 4300 4350 4400 4450 4500 4550 4600 4650 4700

100

150

200

250

300

Frame Index

Pitch

(b)

MODGD PitchREF

Figure 3.17: (a) Melody pitch extracted for North Indian classical audio segment, (b)

Melody pitch extracted for a Carnatic music segment.

3.9.6 Performance Evaluation

The proposed algorithm is evaluated using four datasets. The datasets used in this

evaluation are given in Table 3.4. The duration referred to in Table 3.4 corresponds to

the total duration of all the audio clips in the given database. All the recordings in the

LabROSA, MIREX-2008 and Carnatic music datasets are processed frame-wise with

a hop size of 10 ms while ADC-2004 dataset is processed with a hop size of 5.8 ms

(reference in the dataset is given for hop size of 5.8 ms) .

Table 3.4: Description of the datasets used for the evaluation

Dataset Description Duration

ADC-2004 MIREX-2004 dataset (5 styles) 368sec

LabROSA MIREX-05 Training set (13 Files) 391sec

MIREX-2008 North Indian Classical Music (8 Files) 500sec

CDB Carnatic Music (14 Files) 876sec

Total 2135sec

Evaluation methodology is same as that of MIREX standards. The performance of

52

Page 68: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

the proposed algorithm is evaluated using six metrics discussed in Section 3.5.

3.9.7 Results, analysis and discussion

Table 3.5 compares the performance of the proposed algorithms with various methods

submitted by the participants for MIREX evaluation in 2011, 2012, 2013 and 2014 on

the ADC-2004 dataset. Table 3.6 tabulates the metrics for the MIREX-2008 dataset.

Most of the techniques in Table 3.5 and Table 3.6 are magnitude spectrum based ap-

proaches. The proposed MODGD (Source) phase based method performs at par with the

performance of other magnitude spectrum based approaches. As far as the performance

is concerned, RPA of the proposed algorithm appears to be good for styles where the

singing melody is predominant. In the ADC-2004 dataset, singing melody is predom-

inant in daisy, opera and pop. From the box plot in Figure 3.18, it is obvious that raw

pitch accuracy is higher for those styles as compared to others.

40

50

60

70

80

90

100

Daisy Jazz Midi Pop OperaStyles

Raw

pitc

h ac

cura

cy (

%)

Figure 3.18: Raw pitch accuracy of ADC-2004 dataset in box plot

Adaptive windowing is incorporated in the proposed algorithm to capture dynamic

variation in the melody. The experiment is performed on pop music with various fixed

and adapted window sizes. Experimental results on pop music genre, selected from

ADC-2004 dataset using adaptive windowing are illustrated in Figure 3.19. Initially,

three fixed window sizes 25 ms, 30 ms, 35 ms are selected and predominant melodic

pitch is computed. Later, the experiments are repeated by switching between two win-

dows, one with 80% of previously fixed window based on autocorrelation coefficient

criteria. The conjecture is that the smaller window should be used by the transient seg-

ments and larger window by the steady state segments. It is obvious from the plot that

53

Page 69: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

the window adaptation has benefited the overall performance. The improved results

motivated us to apply this technique to all the datasets.

25 30 3570

71

72

73

74

75

76

77

78

Window Size (ms)

Pitc

h ac

cura

cy (%

)

RPA−fixedRPA−AdaptRCA−fixedRCA−Adapt

Figure 3.19: Results on adaptive windowing on three window sizes

In (Salamon et al., 2014b), it is stated that the best performing algorithms obtain a

raw pitch accuracy between 70 and 80%. One of the submissions of Bin Liao et al. out-

performs other two submissions due to the effective utilization of three pitch tracking

schemes: two trend-estimation-based methods and one HMM-based method (Tzu et al.,

2012). Dynamic programming by the pitch salience peak, in-addition to the transition

information and adaptive windowing to capture the transients, have effectively con-

tributed to improve the accuracy compared to the first approach (MODGD (Direct)).

As compared to the earlier work, imposing rules in the post processing phase also ef-

fectively minimized spurious pitch estimates in the melodic pitch computation. From

Tables 3.5 and 3.6, it can be seen that the Dressler’ algorithm outperforms most of the

algorithms. The algorithm implements an auditory streaming model which builds upon

tone objects and salient pitches. The formation of voices is based on the regular update

of the frequency and the magnitude of so called streaming agents, which aim at salient

tones or pitches close to their preferred frequency range (Dressler, 2011). Dressler’s

method gives the best performance for the audio material, which contain a mixture of

vocal and instrumental pieces, but performs poorly for the other collections where the

melody is always vocal (Salamon et al., 2014b). In the case of MIREX-2008 dataset,

MODGD (Source) algorithm achieves almost the same performance as that of PolyPDA

algorithm (Rao and Rao, 2010). The results reported for this dataset are 73.9% and

76.3% for RPA and RCA, respectively. The proposed MODGD (Source) algorithm out-

performs the algorithm proposed by Stacy Hsueh et al. in RPA and at par with it in

OA.

54

Page 70: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

The metrics RPA, RCA and standard deviation of pitch detection error are used for

the performance evaluation of LabROSA dataset. We see that the proposed methods

achieve better overall performance than ESPS method (wavesurfer) (W.S-URL, 2012),

and MODGD(Source) based approach outperforms the PolyPDA algorithm in terms of

RPA. From Table 3.7, we can observe that the performance of the proposed MODGD

(Source) is at par with that of MELODIA (Salamon and Gomez, 2012) and the algorithm

of V.Arora et al. (Arora and Behera, 2013). The results on the Carnatic music dataset

are summarized in Table 3.8. In Carnatic music dataset, accompaniments are less for

alapanas, which results in better performance for MODGD (Source), as compared to

other datasets. Since singing voice is the predominant melody line in Carnatic music

dataset, [Pmin, Pmax] range selected is also less for MODGD (Direct) algorithm. In the

case of Carnatic music dataset, since the ground truth is not available, pitch estimated

using the algorithm proposed by V.Arora et al. (Arora and Behera, 2013) is used as the

reference for objective comparison. It was the best performing system in MIREX-2012

evaluation. The standard deviation of pitch detection error is less as compared to other

schemes and in terms of RPA, the proposed algorithm is ranked second. Correctness of

the extracted pitch is validated by a professional musician after synthesizing the music

from the extracted pitch.

The performance metric, overall accuracy points towards the need for proper choice

of voicing step. For example, as seen in the Table 3.5, even though the proposed

MODGD (Source) approach is successful in estimating melodic pitches for the major-

ity of frames correctly as that of Salamon et al. (Salamon and Gomez, 2012), voicing

detection deteriorated the overall performance. It can be seen that both the proposed

approach and Salmon et al.’s algorithm have a similar performance in terms of RPA

and RCA for ADC-2004 dataset. Salamon uses a threshold based on the salience distri-

bution of pitch contours to remove non-salient contours before proceeding to filter out

other non-melody contours. But, the system performance is sensitive to a parameter

which determines the lenience of the filtering. Effective use of auditory approach to

separate melody segments resulted in better OA for the algorithm proposed by Dressler

(Ref:Table 3.5) (Dressler, 2011). Since, till date most approaches focus primarily on the

raw pitch accuracy, and less on the voicing decision, overall algorithmic performance

of the melody extraction task will be affected adversely (Salamon et al., 2014b). In

voicing detection phase, the trade-off between VF and VR is also considered in fixing

55

Page 71: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

the optimal threshold on harmonic energy. As far as the post-processing step is con-

cerned, while the proposed algorithm relies on the basic post processing techniques,

there are algorithms which use different approaches like hidden Markov model (Kum

et al., 2016) and group characteristics (Yoon et al., 2011). The performance of the

proposed approach actually demonstrates the potential of the phase based method in

melody extraction task and other music retrieval applications.

Table 3.5: Comparison of results for ADC-2004 submitted in 2011/2012/2013/2014

evaluation.

Method OA RPA RCA VR VF

V.Arora et al 69.06 81.41 85.92 76.51 23.56

Sam Meyer 60.34 64.23 71.21 77.36 32.96

Bin Liao et al (1) 46.24 55.87 66.71 99.98 97.76

Bin Liao et al (2) 41.54 48.32 59.90 99.96 95.37

Bin Liao et al (3) 41.54 48.32 59.90 99.96 95.37

Salamon et al (1) 73.55 76.34 78.71 80.55 15.09

Salamon et al (2) 73.97 77.28 79.41 80.64 15.25

Tachibana et al 59.42 73.03 81.43 74.98 29.37

Yeh 46.99 56.40 65.33 99.91 93.69

C.Cannam et al 66.69 74.79 78.92 63.63 10.35

K. Dressler 86.30 87.10 87.62 91.63 15.76

MODGD (Direct) 55.20 62.80 76.03 80.26 35.00

MODGD (Source) 65.38 72.01 76.95 82.26 26.00

Table 3.6: Comparison of results for MIREX-2008 submitted in 2011/2012/2013/2014

evaluation.

Method OA RPA RCA VR VF

V.Arora et al 67.95 85.85 86.79 70.76 15.58

Sam Meyer 50.06 49.31 59.48 63.52 30.23

Bin Liao et al (1) 70.25 81.94 82.17 100.00 100.00

Bin Liao et al (2) 51.21 59.59 67.95 100.00 100.00

Bin Liao et al (3) 51.51 59.59 67.95 100.00 100.00

Salamon et al 82.78 87.55 88.02 89.26 17.86

Stacy Hsueh et al 63.57 67.64 73.20 78.69 34.25

C.Cannam et al 57.67 73.35 74.43 68.21 44.64

K. Dressler 82.20 87.73 88.48 91.26 28.58

MODGD (Direct) 44.01 49.00 60.05 80.30 34.96

MODGD (Source) 63.01 71.06 73.06 83.55 28.27

Table 3.7: Comparison of σe, RPA, RCA for LabROSA training dataset

Metrics

Method σe RPA (%) RCA (%)

ESPS method (Wavesurfer) 2.97 25.52 62.94

PolyPDA 1.70 62.87 72.67

V.Arora et al 1.85 66.00 72.80

Salamon et al (MELODIA) 2.10 68.62 71.26

MODGD (Direct) 3.37 49.07 60.05

MODGD (Source) 2.67 64.01 69.54

56

Page 72: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Table 3.8: Comparison of σe, RPA, RCA for Carnatic music dataset

Metrics

Method σe RPA (%) RCA (%)

MELODIA 2.62 85.21 90.70

PolyPDA 2.41 89.20 92.03

MODGD (Direct) 3.43 51.00 70.48

MODGD (Source) 2.30 85.50 88.49

Performance analysis on varying degree of polyphony

As mentioned in (Salamon et al., 2014b), varying degree of polyphony is a challenging

issue in the task of audio melody extraction. Regardless of the lead melody line, the task

becomes harder as we increase the number of instruments in orchestral background. In

such instances, the pitch contour of the predominant melodic instrument may be closer,

to melody line of other accompanying instruments, which makes the task of melody

extraction more complex (Salamon et al., 2014b). Salamon et al. discuss the effect of

degree of polyphony in (Salamon et al., 2015).

W.S pYIN MEL MODGD0

20

40

60

80

100

Pitc

h Ac

cura

cy (%

)

(a)

RPA

RCA

W.S pYIN MEL MODGD0

20

40

60

80

100

Pitc

h Ac

cura

cy (%

)

(b)

RPA

RCA

Figure 3.20: (a) RPA and RCA for DB1 [less accompaniments], (b) RPA and RCA for

DB2 [5-6 accompaniments], (W.S : Wavesurfer, pYIN : Probabilistic YIN,

MEL : MELODIA, MODGD : MODGD (Source) )

The proposed MODGD (Source) algorithm is also evaluated on varying level of

polyphony using MedleyDB dataset (Bittner et al., 2014). This dataset consists of an-

notated, royalty free multitrack recordings. The dataset consists of 122 songs, 108 of

which include melody annotations. The dataset consists of a stereo mix and both dry

57

Page 73: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

and processed multitrack stems for each song. During recording process, a set of mi-

crophones is used, such that there may be more than one microphone recording a single

source. The resulting files are raw unprocessed mono audio tracks. The raw files are

then grouped into stems, each corresponding to a specific sound source. The experi-

ments are conducted using two customized datasets; 1) leading singing voice with 1-2

accompaniments (DB1), 2) leading singing voice with 5-6 accompaniments of wider

pitch ranges (DB2). DB1 and DB2 are created by mixing leading singing voices with

background music stems available in the dataset using audacity. The performance of the

melody extraction task on DB1, DB2 is evaluated using Wavesurfer (W.S-URL, 2012),

pYIN (Mauch and Dixon, 2014), MELODIA and MODGD (Source) algorithms. The trend

in the results can be seen in Figure 3.20 (a) and (b). As the number of accompaniments

increases, considerable decrease in pitch accuracy is reported for Wavesurfer and pYIN.

The proposed MODGD (Source) algorithm and MELODIA yield a competitive perfor-

mance, both with RCA greater than 80% on DB1. As the task gets tougher moving on

from DB1 to DB2, where the degree of polyphony is high, the performance of the algo-

rithms declined but less as compared to pYIN and Wavesurfer. Even in the high degree

of polyphony, chroma accuracy of 70% is promising in melody extraction context.

3.10 Automatic music genre classification

Automatic musical genre classification provides a framework for developing and eval-

uating features for content-based analysis of music signals. Due to incomplete and

inconsistent meta data associated with large collection of music files, an increasing de-

mand for automatic retrieval of music content exists nowadays. Automatic music genre

classification task extracts distinctive features from audio excerpts and then automati-

cally categorises them into musical genres. A genre or sub genre can be distinguished

from each other by the style, the cultural context or the content and spirit of the themes.

Although the division of music into genres is somewhat subjective and arbitrary, there

exists perceptual criteria based on the texture, instrumentation and rhythmic structure

of music that can be used to characterize a particular genre (Tzanetakis et al., 2001).

Automatic music genre classification can be utilized effectively to automatically select

radio stations playing a particular genre of music or to design a selective band equalizer

based on the label (Barbedo and Lopes, 2007).

58

Page 74: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Figure 3.21: Block diagram of the proposed system

A comprehensive survey of both features and classification techniques used in genre

classification can be found in (N.Scaringella et al., 2006). In (Salamon et al., 2012b),

melodic features are computed directly from the polyphonic music and features are

fused at feature level with MFCC features to develop a genre classification scheme. In

(Aryafar et al., 2012), a method is introduced with l-SVM classifier which combines

the ideas of the classical SVM with the sparse approximation techniques in genre clas-

sification. A similar approach is also proposed in (Xu et al., 2003). Riedmiller et al. use

unsupervised learning to create a dictionary of features (Wulfing and Riedmille, 2012)

as part of the task. Two novel classifiers, using inter-genre similarity (IGS) modeling

are proposed in (UlaBagci and Erzin, 2007). The method also investigates the use of

dynamic timbrel texture features in order to improve automatic musical genre classifi-

cation performance. In (Tzanetakis and Cook, 2002), three feature sets for representing

timbrel texture, rhythmic content and pitch content are proposed. Extraction of psy-

choacoustic features related to music surface and high level semantic descriptions are

explored in (Scheirer, 2000) for music genre classification. The proposed method in this

work uses fusion of melodic features and modified group delay features for automatic

music genre classification task.

3.10.1 Proposed method

The proposed system is shown in Figure 3.21. The algorithm uses fusion of high level

features derived from melodic contour with low level features to effectively map the

significant characteristics of genres. Initially, modified group delay features are used to

perform genre classification. Later, modified group delay features are fused with high

level features derived from the predominant pitch contour to create a new feature set.

Support vector classifier is used in the classification phase. High level melodic features

include mean pitch height, pitch deviation, jitter, skewness, kurtosis and bandwidth

59

Page 75: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Time (sec)

Fre

qu

en

cy (

Hz)

0 5 10 15 20 25 300

2000

4000

6000

8000

10000

Figure 3.22: Spectrogram (top) and Modgdgram (bottom) for a song from jazz genre

derived from the kernel density estimate of melodic pitch sequences. The following

section describes the feature extraction and classification in detail.

3.10.2 Modified group delay feature (MODGDF)

In the proposed scheme, modified group delay features are used for genre classifica-

tion. 13 dimensional MODGDF features computed framewise are used for fusion ex-

periments. Modified group delay functions effectively map system characteristics in

feature space. For example, the modgdgram 9 and spectrogram of two songs, one from

blues and the other from jazz are shown in Figure 3.22. It is obvious that modgdgram

emphasizes system-specific information when compared to spectrogram.

3.10.3 Melodic Features

Prominent pitch or melodic pitch is the pitch contour of the lead or dominant musical

instrument in a song. In music with orchestration, there are several musical instruments

playing along with the singing voice which result in strong interferences. MELODIA

(Salamon and Gomez, 2012) is used for computing the prominent melodic pitch of all

9 Visual representation of modified group delay functions with time in vertical axis and frame index

in horizontal axis. A third dimension, indicating the amplitude of group delay function at a particular

time is represented by the intensity or color of each point in the image

60

Page 76: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

the excerpts in the proposed system. In (Salamon and Gomez, 2012), pitch candidates

are first estimated, and the context is used to determine pitch contours. Peaks of the

function are grouped over time using auditory streaming cues into pitch contours: time

and frequency continuous sequences of salient pitches. Contour features such as length,

height, pitch deviation and presence of vibrato are then used for filtering non-melody

segments. A set of contour characteristics is computed for each contour and used to

filter out non-melodic contours. Pitch is extracted using a 10 ms hopsize.

The following features are computed from the prominent melodic pitch contour and

used in the experiments.

• Mean pitch height ( µp) : Mean pitch height is given by,

µp =1

N

N∑

i=1

Pi (3.23)

where Pi, N represent pitch for a frame, number of frames, respectively.

• Pitch deviation (Pdev): Pitch deviation is computed using the equation:

Pdev =

√√√√(

1

N

N∑

i=1

(Pi − µp)2) (3.24)

• Jitter: It is the cycle-to-cycle variation of fundamental frequency, i.e. the average

absolute difference between consecutive periods, expressed as:

Jitter(absolute) =1

N − 1

N∑

i=2

| Pi − P(i−1) | (3.25)

Three parameters; skewness, kurtosis and bandwidth are also computed from the

Gaussian kernel density estimate of melodic pitch sequences. Kernel estimators centre

a smooth kernel function at each data point to get smooth density estimate when com-

pared to pitch histograms. If we denote the kernel function as K and its bandwidth by

h, for n data points, the estimated density 10 at any point x is,

f(x) =1

n

N∑

i=1

K(x− x(i)

h) (3.26)

where∫K(t)dt = 1.

• Skewness 11: Skewness is a measure of the asymmetry of the probability distri-

bution. The skewness of a random variable X is the third standardized moment

10http://homepages.inf.ed.ac.uk/rbf../kde.html11https://en.wikipedia.org/wiki/Skewness

61

Page 77: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

100 200 300 400 5000

0.005

0.01

0.015

0.02

0.025

0.03

Pitch (Hz)De

nsity

(a)

Song1Song2

100 200 300 400 5000

0.005

0.01

0.015

0.02

0.025

0.03

0.035

Pitch (Hz)

Dens

ity(b)

Song1Song2

Figure 3.23: Kernel density estimate of melodic pitch for two genres. (a) Pop, (b) Jazz.

γ1, defined by,

γ1 = E

[(X − µ

σ

)3]

=µ3

σ3=

E [(X − µ)3]

(E [(X − µ)2])3/2(3.27)

where µ is the mean, σ is the standard deviation, E is the expectation operator,

and µ3 is the third central moment.

• Kurtosis12: Kurtosis is a measure of the ‘tailedness’of the probability distribution

of random variable. The kurtosis is the fourth standardized moment, defined by ,

Kurt[X ] =µ4

σ4=

E[(X − µ)4]

(E[(X − µ)2])2, (3.28)

where µ4 is the fourth moment about the mean.

• Bandwidth: This is the smoothing parameter, h in Equation (3.26) that approxi-

mates the data optimally, when Gaussian approximation is used.

Kernel density estimate (KDE) of two genres are given in Figure 3.23. KDE of

pitch sequences varies from genre to genre. While spiky nature of density estimate is

observed in Jazz genre, spread estimate becomes the key characteristic of pop music.

12https://en.wikipedia.org/wiki/Kurtosis

62

Page 78: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Table 3.9: Overall accuracy

Sl.No Feature NameAccuracy

(%)

1 MFCC 66.03

2 MFCC+ Melodic 69.00

3 MODGDF 71.73

4 MODGDF + Melodic 75.50

Table 3.10: Classification accuracies obtained by existing approaches.

Sl.No Reference DatasetAccuracy

(%)

1 Li et al. (Li et al., 2003) GTZAN 78.50

2 Tzanetakis et al.(Tzanetakis and Cook, 2002) GTZAN 61.00

3 Panagakis et al.(C. Panagiotakis, 2005) GTZAN 78.20

4 Holzapfel et al.(Holzapfel and Stylianou, 2008b) GTZAN 74.00

5 Salamon et al.(Salamon et al., 2012b) GTZAN 82.00

6 Holzapfel et al.(Holzapfel and Stylianou, 2008b) ISMIR2004 83.50

7 Panagakis et al. (C. Panagiotakis, 2005) ISMIR2004 80.95

3.10.4 Evaluation Dataset

The performance of the proposed system is evaluated on a subset of GTZAN dataset

(Tzanetakis and Cook, 2002). GTZAN dataset is created by Tzanetakis and Cook and

it includes 1000 music excerpts of 30 seconds duration with 100 examples in each of

10 different categories. In the proposed work, five genres: Pop, Blues, Classical, Jazz,

Metal are considered. The classical dataset has the following classes: choir, orchestra,

piano, string quartet whereas the jazz dataset includes big band, cool, fusion, piano,

quartet and swing. The tracks are all 22050 Hz Mono 16-bit audio files in .wav format.

Initially, a baseline system using 13 dimensional MFCC with SVM classifier is cre-

ated. In the second phase, the experiment is repeated with modified group delay fea-

tures. MFCC and MODGDF (13 dimensions) feature sets are converted to utterance level

by averaging across dimensions and a representative feature vector is created for each

utterance. The experiments are extended by combining 6 dimensional melodic features

to MFCC/MODGDF features to create a new feature set (19 dimensions). In all the

experiments, 60% of files are used for training and the rest for testing.

63

Page 79: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Figure 3.24: Genre-wise classification accuracy for all the systems

3.10.5 Results and analysis

The results are tabulated in Table 3.9. Classification accuracies obtained by existing

approaches in music genre classification are also summarized in Table 3.10. Since the

classification frame work, test-train pattern, size of the dataset and segmental length

of the test pattern are different for each of the proposed methods in Table 3.10, direct

comparison only helps to understand the trend in classification framework. The pro-

posed system shows the effectiveness of the fusion of group delay features and melodic

features in automatic music genre classification. Whilst the baseline MFCC system re-

ports overall accuracy of 66.33%, modified group delay based system reports 71.73%.

When we combine the high level melodic features with low-level features, significant

improvement in performance is noted in both systems. Fusion of melodic features with

MFCC and MODGDF achieves an increase of 3% and 4%, respectively. Genre wise

classification for all the experiments are shown in Figure 3.24. For all the genres except

metal, MODGDF fusion has shown improvement in classification accuracy. The con-

fusion matrix for MODGDF fusion experiment is shown in Table 3.11. 39 excerpts out

of 40 pop music files are correctly classified, but for jazz music more misclassification

is observed. Playing instruments using a synthesizer and drums are the uniqueness of

the pop music, which is effectively mapped by MODGDF/MFCC features. Guitar, base

drums and saxophone are common in jazz and pop, which might have lead to more

misclassification. Moreover, classical music excerpts which employ only instruments

are effectively classified by MODGDF/fusion rather than MFCC alone. It is worth not-

ing that MODGDF based system alone outperforms the MFCC fused system except in

the case of blues. Overall, the experiments show the potential of the fusion of melodic

64

Page 80: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Table 3.11: Confusion matrix of fusion (MODGDF + melodic) experiment.

Blues Classical Jazz Metal Pop %

Blues 23 0 10 6 0 58.97Classical 1 38 0 1 0 95.00

Jazz 3 4 18 0 15 45.00Metal 6 0 2 32 0 80.00Pop 0 0 1 0 39 97.50

features with group delay features in automatic music genre classification.

Summary

Two techniques, both based on MODGDF are proposed for automatically extracting

melody from polyphonic music. In the first approach, the music signal is analysed us-

ing modified group delay functions and prominent pitch sequences are estimated. But

due to stringent constraints on the choice of window function, a method is proposed

in which MODGD analysis is performed on the flattened spectrum with system char-

acteristics cancelled. In MODGD (Source) based method, the power spectrum is first

flattened to yield a signal that is rich in harmonics. The harmonically rich spectrum

is then subjected to modified group delay processing for the estimation of sinusoids

in noise. The window size is selected based on adaptive windowing technique using

the autocorrelation function to capture the dynamic variation of melody. Dynamic pro-

gramming is used to ensure consistency across frames. An important feature of the

proposed algorithm is that it neither requires any substantial prior knowledge of the

structure of musical pitch nor any classification framework. Performance of the algo-

rithms on standard databases do indicate that the proposed algorithm shows promise.

The high level melodic features computed from the melodic contour are fused with

baseline MFCC/MODGD features to improve the classification accuracy in automatic

music genre classification. This is also experimentally validated using GTZAN dataset.

Overall, one can conclude that modified group delay based pitch extraction is promis-

ing.

65

Page 81: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

CHAPTER 4

Two-pitch Tracking in Co-channel Speech

4.1 Introduction

Detecting predominant pitch/multiple pitches that is/are present in speech or music is

a very complex task. In common instances such as in crowded party, meetings, where

many people talk simultaneously with varying intonations, it is very difficult to attend

to a particular voice and it is referred to as “cocktail problem ” (Cherry, 1953). When

a combination of speech utterances from two or more speakers are transmitted, pitch

cues of the individual sources will be weakened by the presence of mutual interfer-

ence. In such ambiguous situations, estimating accurate pitch tracks is a challenging

task. In this chapter, an algorithm based on modified group delay functions is proposed

for two-pitch tracking in co-channel speech. Section 4.2 formulates the problem of

multipitch estimation in speech. State-of-the-art approaches are explained in Section

4.3. Section 4.4 describes the evaluation methodology along with metrics and datasets.

Few applications of multipitch estimation in speech and music are listed in Section 4.5.

The proposed algorithm for two-pitch tracking using modified group delay functions

is given in Section 4.6. The performance of the proposed algorithm is evaluated using

standard metrics in Section 4.7 followed by results and analysis in Section 4.8. Finally,

the promise of the algorithm in multipitch environment is experimentally shown using

three-speaker mixed speech in Section 4.9.

4.2 Multipitch Estimation

Multipitch analysis is the task of analyzing the pitch content of polyphonic audio such

as multi-talker speech, polyphonic music and multi-bird songs. It includes estimating

the frequency and number of pitches at each time frame, and organizing the pitches

according to the sources. Several factors make multiple voice f0 estimation more chal-

lenging than single voice f0 estimation. Mutual overlap between voices weaken their

pitch cues and make it difficult to extract the individual estimates.

Page 82: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

90 100 110 120 130 140 150

50

100

150

200

250

300

350

Frame Index

Pitch

(Hz)

Pitch1Pitch2WS−Pitch

Figure 4.1: Monopitch estimation algorithms fails at mixed speech example.

Wavesurfer pitch is plotted for a two-speaker mixed speech. Individual ref-

erence pitches are also shown.

According to MIREX, multipitch analysis can be approached at three levels. 1

• Level 1 - Multi-pitch estimation is to collectively estimate pitch values of all con-

current sources at each individual time frame without determining their sources.

• Level 2 - Note tracking is to estimate continuous pitch segments that typically

correspond to individual notes or syllables.

• Level 3 - Multipitch estimation and streaming is to estimate pitches and stream

them into a single pitch trajectory over an entire conversation or music perfor-

mance for each of the concurrent sources.

Mathematically, the multi-pitch estimation problem can be formulated as follows

(Christensen et al., 2008): Consider a signal consisting of several, say K, sets of har-

monics with fundamental frequencies ωk, for k = 1, . . . , K, that is corrupted by an

additive white Gaussian noise ω[n], having variance σ2, for n = 0, . . . , N − 1, i.e.,

x[n] =K∑

k=1

L∑

l=1

ak,lejωkln + ω[n] (4.1)

where ak,l = Ak,lejφk,l is the complex amplitude of the lth harmonic of the source

with Ak,l > 0, φk,l being the amplitude and the phase of the lth harmonic of the kth

source, respectively. The model in Equation (4.1) is known as the harmonic sinusoidal

model. The task is to estimate the individual pitch estimates ωk in the mixture signal.

Most of the pitch detection algorithms are designed to estimate pitch in single speaker

cases. Monopitch estimation algorithms fail in estimating multiple pitches in mixed

1http://www.ece.rochester.edu/ zduan/multipitch/

67

Page 83: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

speech. The pitch estimated for a mixed speech utterance, using Wavesurfer algorithm

(W.S-URL, 2012) is shown in Figure 4.1. It fails to track individual pitches correctly in

the presence of strong interference from another speaker. Multiple f0 estimation sys-

tems can be broadly classified according to their mid-level representation (time domain,

spectral domain, auditory filter banks) or by the way they can estimate the interactions

between sources (iterative and joint estimation methods) (Ibnanez, 2010).

4.3 Multipitch estimation algorithms : State-of-the-art

approaches

Numerous methods have been reported for multipitch estimation in speech and music

(Li et al., 2008; Nishimoto et al., 2007; Wu and Wang, 2003). The research on multi-

pitch estimation was initiated through separating co-channel speech signals date back

to 1970 (Shields, 1970). All the approaches can be classified into one of the three

classes, spectral, temporal and spectro-temporal. The existing strategies include itera-

tive estimation or joint estimation, in time/frequency domain. In iterative estimation, a

single-voice algorithm is applied to estimate the f0 of one voice, and that information is

then used to suppress the voice from the mixture so that the fundamental frequencies of

the other voices can be estimated. Suppression of the other voices, may in turn be used

to refine the estimate of the first voice. In joint estimation, all the voices are estimated

at the same time. Whatever be the approach, there exists a compromise between the

computational cost and robustness in the performance. Joint estimation does the source

separation better than iterative estimation but with high computational complexity (Yeh,

2008). This section reviews various methods in detail.

Iterative estimation

In iterative estimation, f0 of the predominant source is computed in the first pass, the

estimated source and its harmonics are suppressed in the second pass, and the next can-

didate is pursued. This process is repeated until the termination requirement is met.

Iterative estimation assumes that at each iteration, there exists a predominant source

such that the extraction of one single f0 is reliable when the remaining partials are frag-

mentary. Schroeder’s histogram based approach is employed for multipitch estimation

68

Page 84: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

in (Kemp, T. and Waibel, A. , 1999). In this work, peak picking in summary autocorre-

lation function (SACF) is used as f0 candidate in the first pass. In the second pass, it is

cancelled from the autocorrelation function (ACF). Iterative cancellation is also applied

in the spectral domain for multipitch estimation (Berenguer et al., 2005), in which bi-

nary masks around matched harmonics are employed to annihilate the estimated partials

from the spectrum. In Klapuri’s work (Klapuri, 2008), signals are first compressed and

half wave rectified. Harmonic matching in the summary magnitude spectrum results in

the predominant pitch followed by its attenuation. Spectral smoothness principle was

proposed as an efficient new mechanism in multipitch estimation in (P.Klapuri, 2001).

Spectral smoothness refers to the expectation that the spectral envelopes of real sound

sources tend to be continuous. The experimental framework of the spectral-smoothness

based model is given in Figure 4.2.

A graphical model based frame work is employed in (Jordan, 2004) for pitch esti-

mation. Maximization of likelihood in a spline smoothing model leads to predominant

pitch detection in the proposed work. In (Abeysekera, 2004), two dimensional distribu-

tion is computed using bispectrum for pitch estimation. The estimated pitch component

is annihilated from the mixed source. Ming Li et al. proposed an iterative estima-

tion method grounded on the subharmonic summation method and spectral cancellation

framework (Li et al., 2008). Iterative strategy in spectral domain is attempted in (Kemp,

T. and Waibel, A. , 1999). First, f0 derived from spectral peaks is removed from the

spectrum and a second f0 is estimated from the residual. In de Chevegne’s work, both

joint cancellation and iterative cancellations in time domain are studied (de Cheveigne,

1993).

Figure 4.2: Iterative multipitch estimation (Klapuri, 2008)

69

Page 85: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Joint estimation

In joint estimation, multiple fundamental frequencies are jointly estimated without any

cancellation framework. A two way mismatch method to estimate pitches is proposed

in Maher and Beauchamp (Maher and Beauchamp, 1994). The algorithm searches for

the pair of fundamental frequencies that minimizes frequency discrepancies between

the harmonic models and the observed peaks. It is the mismatch from the predicted

to measured and the mismatch from the measured to the predominant. Each match is

weighted by the amplitudes of the observed peaks. In this way the algorithm minimizes

the residual by the base match. Some of the algorithms follow a correlogram based

approach, in which channel-lag representation of auto-correlation function in the audi-

tory channel is further processed. Enhanced summary autocorrelation function (ESAF)

based pitch estimation is used in the proposed work of Karjalainnen and Tolonen (Tolo-

nen and Karjalainen, 2000).

The work proposed by Zhang et al. (Zhang et al., 2008) uses weighted summary

auto-correlation function (SACF) in which modified amplitude of ACF models the re-

lationship between f0 and frequency of channel. In (de Cheveigne, 1993; Tolonen and

Karjalainen, 2000), temporal approaches are used with joint estimation strategy. Meddis

and Hewitt extended their single pitch perception model to multiple source separation

in (Meddis and Hewitt, 1991). The correlogram based algorithm proposed by Wu et al.

(Wu and Wang, 2003) uses a unitary model of pitch perception to estimate the pitch of

multiple speakers. The input signal is decomposed into sub-bands using a gammatone

filterbank and the framewise normalized autocorrelation function is computed for each

channel. The peaks selected from all the channels are used to compute the likelihoods

of pitch periodicities and these likelihoods are modeled by an HMM to generate the

pitch trajectories.

2-D processing of speech in multipitch estimation is proposed by Tianyu et al.

(Wang and Quatieri, 2009). The localized time-frequency regions in the narrow band

spectrogram are analyzed to extract pitch candidates in the work. Multi-region analy-

sis combined with the pitch candidates results in the final pitch estimates (Wang and

Quatieri, 2009). M.H Radfar et al. proposed a joint estimation method called MP-

Tracker in spectral domain (Radfar et al., 2011). Pitches are estimated by minimizing

the log spectral distortion between the mixture spectrum and parametric spectra with re-

70

Page 86: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

spect to the underlying frequency components to obtain estimates of the pitch frequen-

cies. Periodicity features are extracted through an auditory front-end and less corrupted

channels in (Jin and Wang, 2011). A summary of few algorithms are listed in the Table

4.1.

Table 4.1: Multipitch estimation-Summarization of few algorithms

S.No. Year Name of Author Algorithms used

1 1999 P.J. Walmsley et al. (Walmsley

et al., 1999)

Polyphonic pitch tracking using joint

Bayesian estimation

2 2001 A.P. Klappuri (P.Klapuri, 2001) Spectral smoothness principle

3 2003 A.P. Klappuri (Klapuri, 2003) Based on harmonicity and spectral

smoothness

4 2008 A. Klappuri (Klapuri, 2008) Using an auditory model

5 2009 R.Bedeau et al. (Badeau et al.,

2009)

Expectation maximization algorithm

6 2010 E. Vincent et al. (Vincent et al.,

2010)

Adaptive harmonic spectral decomposi-

tion

7 2011 Q. Huang et al. (Huang and Wang,

2011)

Based on multi-length windows harmonic

model

8 2011 Wohlmayr et al. (Wohlmayr et al.,

2011)

Using factorial hidden Markov approach

9 2012 John Xi Zhang et al. (Zhang et al.,

2012)

Subspace analysis along with time space

model

10 2012 Daniele Giacobello et

al.(Giacobello et al., 2012)

Combination of sparsity and linear pre-

diction framework

11 2013 S.I. Adalbjornsson (Adalbjornsson

et al., 2013)

Using block sparsity

12 2014 Zhiyano Duan et al.(Pardo et al.,

2010)

Constrained clustering approach

4.4 Evaluation methodology

The guidelines for evaluating the performance of monopitch estimation can be found

in (Rabiner et al., 1976). Since there are no guidelines for the performance evaluation

in the case of multipitch tracking, usually the guidelines of single pitch tracking is

extended for multipitch estimation. Speech mixtures are obtained by mixing utterances

at 0 dB. The utterances are chosen from any of the monopitch databases. The ground

truth is computed using one of the efficient mono-pitch tracking algorithms available.

Datasets

Datasets, which are used in the multipitch experiments include GRID (Cooke et al.,

2006), ATR Database (A.Kurematsu et al., 1995), PTDB-TUG Database (Petriker et al.,

71

Page 87: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

2011), Simple4All (Suni et al., 2014), NOIZEUS and Keele database (Plante et al.,

1995).

Evaluation metrics

The metrics used for the evaluation of the performance of various multipitch estimation

algorithms are listed below.

• Accuracy: Accuracy10 and Accuracy20 correspond to the percentage of frames

at which pitch deviation is less than 10% and 20% with respect to the reference,

respectively. A gross error occurs if the detected pitch is not within the specified

threshold with respect to the reference pitch.

• Standard deviation of the fine pitch errors (Efs) : The standard deviation of the

fine pitch error is a measure of the accuracy of pitch detection during voiced

intervals. The standard deviation of the pitch detection (Efs) is given by:

Efs =

(1

N

(ps − p′s)2 − e2 (4.2)

where ps is the standard pitch, p′s is the detected pitch,N is the number of correct

pitch frames and e is the mean of the fine pitch error. e is given by:

e =1

N

(ps − p′s) (4.3)

• Transition error: The algorithms proposed in (Wu and Wang, 2003; Jin and Wang,

2011) use misclassification as a criterion along with gross errors. In the error

calculation, Ex−−>y denotes the error rate of time frames where x pitch periods

are misclassified as y pitch periods. The gross error, Egs is the percentage of

frames where the detected pitch differs from the true pitch by more than 20%.

4.5 Applications of multipitch estimation

The estimation of the fundamental frequency, or the pitch of audio signals has a wide

range of applications in CASA, prosody analysis, source separation and speaker iden-

tification (de Cheveigne, 1993; Murthy and Yegnanarayana, 2011). In music context,

multipitch estimation is inevitable in applications such as the extraction of “predomi-

nant fo” (Salamon and Gomez, 2012), computation of bass line (Goto and Hayamizu,

1999), content-based indexing of audio databases (Tao Li et al., 2003) and automatic

72

Page 88: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

transcription (Ryynanen and Klapuri, 2008a). It should be noted that interactive mu-

sic applications demand highly robust real time pitch estimation algorithms in all as-

pects (Poliner et al., 2007). Multipitch estimation is widely used in source separation

(P.Klapuri, 2001; Wang and Loizou, 2006).

In (Wang and Loizou, 2006), a single channel speech separation method based on

binary spectrum mask estimation is proposed. The long-short frame associated har-

monic model (LSAH) is used to analyze the spectrum for a speech mixture. Then mul-

tipitch estimation and binary spectrum mask estimation are performed for each speech

source. In (Shao and Wang, 2003), a new usable speech extraction method is proposed

to improve speaker identification performance under the co-channel situation based on

pitch information obtained from a robust multipitch tracking algorithm. Source sepa-

ration through multipitch estimation is addressed by spectral smoothness principle in

(P.Klapuri, 2001).

Multipitch estimation has also established its role in music information retrieval ap-

plications. In melody extraction task proposed by Salamon et al. (Salamon and Gomez,

2012), several pitch contours are first created using salience function and then melody is

estimated based on contour characteristics. In (Salamon et al., 2012a), a method is pro-

posed for tonic identification for Indian classical music based on a multipitch analysis

of the audio signal. Unlike approaches that identify the tonic from a single predomi-

nant pitch track, multipitch representation is used to construct a pitch histogram of the

audio excerpt from which the tonic is identified (Salamon et al., 2012a). Multipitch

estimation task has also become a central tool in musical scene analysis (Christensen

and Jakobsson, 2009).

4.6 Modified group delay based two-pitch tracking in

co-channel speech

Majority of the pitch tracking methods are usually limited to clean speech and give a

degraded performance in the presence of other speakers or noise. An algorithm based

on modified group delay functions is proposed and evaluated on various datasets in this

work.

73

Page 89: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

4.6.1 Theory of pitch detection using modified group delay func-

tions

The monopitch estimation discussed in Section 3.9 can be extended to estimate multi-

ple pitches in multi-talker environment. In the case of multiple speakers, the flattened

power spectrum contains the excitation information of all the speakers.

Consider a speech signal with two different pitches corresponding to that of two

different voices. The source spectrum of this signal (assuming no linear phase) is

E(z) = 1 + z−To + z−T1 + z−2To + z−2T1 + z−3T0 + z−3T1 + ... (4.4)

Assuming that a frame of speech contains no more than two pitch periods, and that there

is no linear phase, the power spectrum of the source is given by:

E(z)E∗(z) = (1+ z−To + z−T1 + z−2To + z−2T1)(1+ zTo + zT1 + z2To + z2T1) (4.5)

Substituting z = ejω,

| E(ejω) |2= 5 + 4 cos(ωTo) + 4 cos(ωT1)+

2 cos(ω2To) + 2 cos(ω2T1) + 2 cos(ω(To − 2T1))+

2 cos(ω(T1 − 2T0)) + 2 cos(ω(2(T1 − T0))) + 2 cos(ω(T1 − T0)) (4.6)

By introducing a parameter γ (0 < γ ≤ 1 ), to control the flatness of the spectrum, we

have

| E(ejω) |2γ= (5 + 4 cos(ωTo) + 4 cos(ωT1)+

2 cos(ω2To) + 2 cos(ω2T1) + 2 cos(ω(To − 2T1))+

2 cos(ω(T1 − 2T0)) + 2 cos(ω(2(T1 − T0))) + 2 cos(ω(T1 − T0)))γ (4.7)

This only results in a composite signal which still is a sum of sinusoids. The number of

74

Page 90: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

sinusoids and the frequencies of the sinusoids are related to the pitch of the individual

sources. If the spectral components corresponding to the periodic component are em-

phasised, the problem of pitch extraction reduces to that of the estimation of sinusoids

in the frequency domain. We now replace ω by n and To, T1 by ωo, ω1, respectively in

Equation (4.7), and remove the dc component to obtain a signal which is ideally a sum

of sinusoids that corresponds to the excitation components of the mixture.

s[n] = a cos(nωo) + b cos(nω1) + c cos(n2ωo)

+ d cos(n2ω1) + e cosn(ω0 − 2ω1) + f cosn(ω1 − 2ω0)

+ g cosn2(ω1 − ω0) + h cosn(ω1 − ω0) + · · ·+ . . .

(4.8)

where a, b, c, d, e, f, g, h are constants.

This signal is subjected to modified group delay processing, which results in peaks

at multiples of the various pitches present in the speech mixture. The procedure to map

these peak locations to constituent pitch trajectories is explained in the next section.

4.6.2 Proposed system description

The block diagram of the proposed system is given in Figure 4.3. The dotted block

represents the MODGD (Source) based predominant pitch estimation algorithm dis-

cussed in Chapter 3. The proposed algorithm computes constituent pitch frequencies

from the mixed speech through iterative estimation and cancellation method. First, the

flattened spectrum is frame-wise analysed by MODGD algorithm. As discussed in Sec-

tion 3.9, peaks can be observed in the MODGD domain at locations corresponding to

the multiples of all the pitch periods and its algebraic combinations. As discussed in the

prominent pitch estimation algorithm, the location of the prominent peak in the range

[Pmin, Pmax] in the MODGD feature space is mapped to the first candidate pitch esti-

mate, where Pmin, Pmax represent minimum and maximum pitch, respectively. In the

second pass, the estimated pitch component and its harmonics are annihilated from the

flattened power spectrum. The residual signal is again subjected to MODGD analysis. In

the post processing phase, pitch grouping followed by removal of pitch outliers results

in the individual pitch trajectories.

75

Page 91: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Figure 4.3: Block diagram of the proposed method. Dotted block represents the

MODGD (Source) based predominant pitch estimation algorithm.

4.6.3 Estimation of constituent pitches

Once the prominent pitch is estimated from the flattened spectrum in the first pass, the

objective is to estimate the second pitch candidate. The proposed algorithm uses comb

filters to annihilate the estimated pitch and its harmonics from the flattened spectrum in

the second pass.

Comb filtering

Comb filters are widely used in many speech processing applications such as speech

enhancement, pitch detection and speaker recognition (Jin et al., 2010; Laskowski and

Jinn, 2010). It has already been shown that the optimum-comb technique is a fast

and useful technique for the extraction of pitch period data from continuous speech

in (Moorer, 1974). Adaptive comb filter (ACF) for harmonic signal enhancement and

spectral estimation is proposed in (Porat and Nehorai, 1986). In (Tan and Alwan, 2011),

a signal to noise ratio weighted correlogram based pitch estimation algorithm is de-

scribed. A bank of comb filters operates in each of the low, mid and high frequency

bands to capture different set of harmonics. Comb filters are also utilized for musical

pitch estimation and the discrimination of different musical instruments (Miwa et al.,

2000). The proposed algorithm uses comb filters to annihilate the estimated pitch and

76

Page 92: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

its harmonics from the flattened spectrum in the first pass. If a signal is periodic, it sug-

gests that some basic waveform repeats with a certain frequency. Mathematically, we

can express it as x[n] ≈ x[n−D] where D is the repetition or pitch period. A measure

of periodicity can be obtained using a metric e(n) defined as

e[n] = x[n]− αx[n−D] (4.9)

where α is a constant, introduced to allow for some variation of the amplitude.

In Z domain,

E(z) = (1− αz−D)X(z) (4.10)

The FIR comb filter transfer function is represented as :

H(z) =E(z)

X(z)= 1− αz−D (4.11)

which shows that the process of matching of a signal by a delayed version of itself is a

filtering problem. Notching comb filters are filters that cancel out signal components at

certain frequencies. In the complex case, these typically have the following form

H(z) =1 + βz−1

1 + ρβz−1=

P (z)

P (ρ−1z)(4.12)

where β is a complex coefficient and -1 < ρ < 1 is real. Since the periodic signal is

comprised of possibly many harmonics, we can use Lk such notch filters having notches

at frequencies Ψi. Such a filter can be factorized into the following form.

P (z) = ΠLk

i=1(1− ejΨiz−1) (4.13)

P (z) = 1 + β1z−1 + β2z

−2 + · · ·+ βLkz−Lk (4.14)

which has zeros on the unit circles at angles corresponding to the desired frequencies.

The comb effect results from phase cancellation and reinforcement between the

delayed and undelayed signal. The magnitude response of the comb filter is

| H(ejω) |=√

(1 + α2) + 2α cos(ωD) (4.15)

Magnitude response and pole-zero plot of comb filter is shown in Figure 4.4. In the

77

Page 93: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9−60

−50

−40

−30

−20

−10

0

Normalized Frequency (×π rad/sample)M

ag

nitu

de

(d

B)

Magnitude Response (dB)

−1 −0.5 0 0.5 1

−1

−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

1

Real Part

Ima

gin

ary

Pa

rt

Figure 4.4: Notch filter magnitude response and pole-zero plot

proposed approach, comb filter is used to annihilate the predominant fundamental fre-

quency component obtained in the first pass from a composite flattened power spectrum

which constitutes multiple excitations.

Removal of partials

Consider a speech mixture of two synthetic speech signals with fundamental frequen-

cies of 200 Hz and 280 Hz. One frame of such a mixed speech is shown in Figure 4.5

(a). Flattened spectrum of corresponding frame is shown in Figure 4.5 (b). Modified

group delay function computed for the synthetic mixture frame is shown in Figure 4.5

(c). The peaks in MODGD feature space, correspond to the pitch candidates present

in the speech mixture and its integral multiples. In the first pass, the prominent peak

in the MODGD feature space is mapped to first pitch estimate followed by the annihi-

lation of it from the residual spectrum. The red color contour in Figure 4.5 (d) is the

computed MODGD for the second pass. The individual pitch tracks computed through

the steps discussed are shown in Figure 4.5 (e) along with reference pitches. Similarly,

another natural audio mixture example is shown in Figure 4.6. Figure 4.6 (a) shows the

MODGD plot for a real audio frame and in Figure 4.6 (b), pitch estimates of the audio

segment are shown. The modified group delay functions obtained in the first pass, and

in the second pass are illustrated in the figure. It is obvious from the figure that the pitch

78

Page 94: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

component corresponding to the predominant peak is annihilated in the second pass.

0 50 100 150 200 250 300−1

−0.5

0

0.5

1

Time(ms)

Am

plitu

de

(a)

1000 2000 3000 4000

0

1

2

3

4

5

6

7

8

9

Frequency(Hz)

Am

plitu

de

(b)

10 20 30 40 500

20

40

60

80

100

120

140

160

180

Time(ms)

MO

DG

D(m

s)

(c)

To

T1

2To

Peakcorrespondsto 2T1

10 20 30 40 500

50

100

150

200

250

300

350

400

450

Time(ms)

MO

DG

D(m

s)

(d)

0 50 100 150 200 250 300180

200

220

240

260

280

300

Time(ms)

Fre

qu

en

cy(H

z)

(e)

REF−Fo(1)

Ref−Fo(2)

MODGD−Fo(1)

MODGD−Fo(2)

Figure 4.5: (a) A frame of synthetic mixture, (b) Flattened spectrum of frame in (a), (c)

MODGD plot in the first pass. (d) MODGD of (a) after the application of comb

filter (second pass), (e) Pitch extracted for the mixed synthetic speech segment.

4.6.4 Pitch trajectory estimation by grouping

At the end of the pitch estimation phase, two pitch candidates per frame are computed.

In the pitch grouping stage, these candidates are grouped into trajectories which com-

prise of continuous smooth individual tracks. Since pitch crossing is not considered,

out of two candidates per frame, high pitch values are consolidated to form one trajec-

tory, while smaller values to the other. Alternatively, dynamic programming based pitch

grouping can also be employed. In that case, the relative closeness of the distance be-

79

Page 95: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

1 1.5 2 2.5 30

0.01

0.02

0.03

0.04

Time(ms)

MO

DG

D(m

s)

(a)

MODGD−First Pass

MODGD−Second Pass

140 145 150 155 160 165 170 175 180

50

100

150

200

250

300

350

400

Frame Index

Fre

quen

cy(H

z)

(b)

Fo(1)−RefFo(2)−RefMODGD Fo(1)MODGD Fo(2)

Figure 4.6: (a) MODGD plot for a frame in first pass and residual spectrum (dotted line)

for a real speech mixture, (b) Presence of stray values in the pitch contour

are marked in circle.

tween peaks in two consecutive frames is used to compute the optimal path. Transition

cost is computed as the absolute difference in distance between the current and previ-

ous frame. The optimal path is selected by minimizing the transition cost across frames

using back tracking. The transition cost Ct(cj/cj−1) between the pitch candidates cj

and cj−1 of consecutive frames is given by (Veldhuis, 2000)

Ct(cj/cj−1) =| Lj − Lj−1 | (4.16)

where Lj , Lj−1 are peak locations in consecutive frames. The optimal pitch sequence

(c1...cM ) is computed using dynamic programming with candidates c1 in the first and

cM in the M th frame in a block by minimizing the transition cost function (Veldhuis,

2000). Transition cost TC(c1...cM) of pitch candidates c1 to cM is computed by

TC(c1...cM) =

j=M∑

j=2

Ct(cj/cj−1) (4.17)

80

Page 96: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

The optimal sequence of pitch markers is determined by back tracking from the can-

didate cM in the M th frame in a block to c1 in the first frame. If the pitch detection

algorithm computes any spurious candidate, dynamic programming may result in erro-

neous pitch tracks. Continuity of pitch contours is used as a criteria to disambiguate the

pitch values.

100 200 300 400 500

0.1

0.2

0.3

0.4

0.5

0.6

0.7

(a)

Time(ms)

Harm

onic

Energ

y

0 100 200 300 40050

100

150

200

250

300

(b)

Time(ms)

Freq

uenc

y(Hz)

Fo(1)

Fo(2)

Fo(1)−Ref

Fo(2)−Ref

1 2 3

0 100 200 300 4000

50

100

150

200

250

Time(ms)

Freq

uenc

y(Hz)

(c)

Fo(1)

Fo(2)

Fo(1)−Ref

Fo(2)−Ref

Figure 4.7: (a) Harmonic energy of a segment, (b) Initial pitch estimates of the segment

in (a), (c) Final pitch trajectory in the post processing stage.

4.6.5 Postprocessing

The accuracy in pitch estimation is improved by a post processing stage in which the

first task is to identify the segments where one speaker is alone present. A soft thresh-

old on harmonic energy is employed to identify these segments. Harmonic energy is

computed from the estimated fundamental frequency and its harmonics. Multiples of

81

Page 97: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

fundamental frequency are obtained by searching for local maxima with 3% tolerance.

Harmonic energy of a signal x[n] is computed by (Ramakrishnan et al., 2008)

En =

kNf0∑

k=kf0

| X [k] |2, (4.18)

where X [k], k, f0 represent the Fourier transform magnitude, bin number, fundamen-

tal frequency, respectively. kf0 and kNf0 represent bins corresponding to fundamental

frequency and its N th multiple. Segments which are detected as single speaker frames

in the voicing detection stage are processed again for monopitch estimation using the

MODGD algorithm. If the pitch estimated follows a path, the estimated sequence is

appended to the existing pitch contour by ensuring consistency. This is illustrated in

Figure 4.7. Figure 4.7 (a) shows harmonic energy plotted for a mixed speech segment.

Regions 1 and 3 in Figure 4.7 (b) consists of only one speaker speech, these regions

have be processed using monopitch estimation algorithm. The appended pitch tracks

obtained are shown in Figure 4.7 (c).

As part of smoothening the curve, stray values are removed by framing appropriate

rules that refine pitch estimates. For example, let ft and ft+1 be the pitch candidates of

consecutive frames in a pitch track after the grouping stage. If ft+1 lies outside the range

[ ft - ρ , ft+ ρ ], this is treated as a spurious pitch estimate and will be interpolated using

previous and successive pitch values (Radfar et al., 2011). We use linear interpolation

for replacing missing pitch frequencies; however, other interpolation techniques such as

cubic or spline interpolation could be used. This simple but effective technique reduces

the pitch error considerably. Note that missing pitch frequencies should typically not be

interpolated for segments corresponding to 40 ms (Radfar et al., 2011). The threshold

ρ is set heuristically to 10 Hz. A typical example is shown in Figure 4.6(b). The circled

part indicates the presence of two stray values in the middle of a continuous curve.

The estimated pitch trajectories for a speech mixture consisting of male and female

utterances are shown in Figure 4.8. Figure 4.8 (a) shows the initial pitch estimates and

Figure 4.8 (b) shows the individual pitch trajectories after post processing. The pitch

trajectories estimated using Wu et al. algorithm (D.L.Wang et al., 2003) are also shown

in Figure 4.8 (c) for the same speech mixture.

82

Page 98: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

200 400 600 800 1000 1200 1400

50

100

150

200

250

300

350

(a)

Time(ms)

Freq

uenc

y(H

z)

REF−Fo(1)REF−Fo(2)Fo(1)Fo(2)

200 400 600 800 1000 1200 1400

50

100

150

200

250

300

350

(b)

Time(ms)

Freq

uenc

y(H

z)

REF−Fo(1)REF−Fo(2)Fo(1)Fo(2)

200 400 600 800 1000 1200 1400

100

150

200

250

300

350

Time(ms)

Freq

uenc

y(H

z)

(c)

REF−Fo(1)REF−Fo(2)WWB−Fo(1)WWB−Fo(2)

Figure 4.8: (a) Initial pitch estimates for a speech mixture, (b) Final pitch trajectories

estimated using the proposed algorithm, (c) Pitch trajectories estimated us-

ing WWB algorithm.

4.6.6 Pitch extraction in noisy and reverberant environment

The presence of noise and reverberation in speech poses a problem, even for mono-

pitch estimation. For noise corrupted speech, both the time-domain periodicity and

spectral-domain periodicity are distorted and hence, conventional pitch estimation fails

to a certain extent (Huang and Lee, 2013). Group delay domain representation of speech

makes it relatively immune to noise when compared to that of the short-time magnitude

spectrum (Hegde et al., 2007b; Yegnanarayana and Murthy, 1992). For instance, con-

sider the noisy signal x[n] as the output of the autoregressive process s[n], corrupted

with Gaussian noise ω[n], i.e

x[n] = s[n] + ω[n] (4.19)

Group delay analysis of an autoregressive process in a noisy environment is discussed

in (Yegnanarayana and Murthy, 1992). z-transform of s[n], ignoring the effects of

83

Page 99: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

truncation of the response of an all-pole system is given as

S(z) =GE(z)

A(z)(4.20)

where E(z) is the z-transform of the excitation sequence e[n] and G/A(z)is the z trans-

form of the all-pole system corresponding to the autoregressive process. From Equation

4.19 and Equation 4.20, we get;

X(z) =GE(z) +W (z)A(z)

A(z)=V (z)

A(z)(4.21)

In group delay domain,

τX(ejω) = τV (e

jω)− τA(ejω) (4.22)

As explained in (Yegnanarayana and Murthy, 1992), the noise spikes in τX(ejω)

can be suppressed by multiplying with the estimated zero spectrum. This results in

an estimate of -τA(ejω) which corresponds to the spectral component in the composite

signal. Thus, group delay based approach is very effective in analyzing frequency com-

ponents of a composite signal in the presence of noise. Room reverberation adversely

affects the characteristics of pitch and thus makes the task of pitch determination more

challenging. It causes degradation of the excitation signal (Jin and Wang, 2010). In re-

verberant environments, the speech signal that reaches the microphone is superimposed

with multiple reflected versions of the original speech signal. These superpositions can

be modeled by the convolution of the room impulse response (RIR), that accounts for

individual reflection delays, with the original speech signal (Allen and Berkley, 1979).

Mathematically, the reverberant speech r[n] is obtained as the convolution of speech

signal s[n] and room impulse response h[n] (Thomas et al., 2008).

r[n] = s[n] ∗ h[n] (4.23)

The room impulse response is described as one realization of a non-stationary stochastic

process in Schroeder’s frequency-domain model (Jot et al., 1997) given by,

h[n] = b[n]e−δn (4.24)

where b[n] is a centered stationary Gaussian noise, and δ is related to the reverberation

84

Page 100: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

time Tr. A typical room impulse response used for the experiment is shown in Fig-

ure 4.9. The proposed algorithm is also analysed in a reverberant environment using

simulated impulse response.

0 0.01 0.02 0.03 0.04 0.05 0.06−0.04

−0.02

0

0.02

0.04

0.06

Time(sec)

Ampl

itude

Figure 4.9: Room impulse response

Table 4.2: Category of mixtures for Dataset:1 and 2

Category Speech data

1 Male/Female, Female/Female, Male/Male

2 Male/Female, babble noise

3 Male/Female, white noise

4 Male/Female with reverberation

4.7 Performance evaluation

In the proposed work, focus is given to the multipitch estimation of speech mixture

with two speakers. The performance of the proposed algorithm was evaluated using the

following datasets:

• Dataset-1: The dataset consists of 40 audio files obtained by mixing a subset of

utterances from the Pitch Tracking Database of Graz university of Technology

(PTDB-TUG) (Petriker et al., 2011), Childers database and a few audio sam-

ples from Simple4All speech corpora (Suni et al., 2014). The PTDB-TUG con-

sists of audio recordings with phonetically rich sentences from TIMIT corpus.

Simple4All corpora consists of audio samples from different languages.

• Dataset-2: GRID (Cooke et al., 2006) corpus consists of high-quality audio and

video recordings of 1000 sentences spoken by each of 34 talkers (18 male, 16

85

Page 101: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

female). A subset of 40 audio files are used for generating mixtures for the eval-

uation.

In the experiments, each audio mixture is processed using a Hamming window of

frame length of 30 ms and hop size of 10 ms. As shown in Table 4.2, the interfer-

ences are classified into four categories by considering clean and noise conditions.

The data set contains audio files of cross gender (male/female) and same gender (fe-

male/female, male/male). The test was conducted mainly for a 0 dB target-to-masker

ratio (TMR) which is considered as the most difficult situation in co-channel speech

segregation problem as both talkers equally mask each other. In categories 2 and 3,

speech is obtained by mixing the category 1 speech data with babble noise (5 dB SNR)

and white noise (10 dB SNR). Category 4 interferences comprising of simulated re-

verberant speech utterances. The performance is also evaluated with speech mixture

of male and female voices with +3dB and -3dB TMR. Reverberant speech is generated

using simulated room acoustics using the model proposed in (Allen and Berkley, 1979).

The simulation uses a reverberation time of T60 = 200ms.

Table 4.3: Performance evaluation on Dataset-1

Category Accuracy10(in%)

MODGD WWB JIN

Male-Female 84.58 76.71 81.04

Female-Female 75.02 60.78 70.31

Male-Male 73.58 66.01 70.01

Male-Female,Babble noise(10 dB SNR) 72.12 60.17 73.43

Male-Female,Babble noise (5 dB SNR) 70.12 52.00 66.72

Male-Female,White noise (10 dB SNR) 73.13 62.97 76.21

Male-Female,White noise (5 dB SNR) 68.05 55.24 66.50

Male-Female with reverberation 63.05 62.70 65.28

Table 4.4: Performance evaluation on Dataset-2

Category Accuracy10(in%)

MODGD WWB JIN

Male-Female 82.65 78.92 78.95

Female-Female 75.01 76.74 76.81

Male-Male 70.30 50.84 73.08

Male-Female,Babble noise (10 dB SNR) 74.32 56.25 72.74

Male-Female,Babble noise (5 dB SNR) 71.08 50.30 65.08

Male-Female,White noise (10 dB SNR) 66.50 66.34 71.64

Male-Female,White noise (5 dB SNR) 59.20 55.28 60.64

Male-Female with reverberation 68.00 64.01 70.38

The performance is evaluated only for voiced frames. The reference frequency of

an unvoiced frame is considered as 0 Hz. To evaluate the performance of our algo-

rithm, a reference pitch contour corresponding to true individual pitch is required. We

86

Page 102: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

computed the reference pitch of clean speech using Wavesurfer (W.S-URL, 2012). The

performance is quantitatively assessed by measuring two types of metrics: accuracy and

standard deviation of the fine pitch errors Efs.

4.8 Results analysis and discussions

The performance of the proposed algorithm is evaluated on three different speech mix-

tures, namely, same gender (M/M, F/F) and cross gender (M/F). In addition to the clean

speech condition, the performance is also evaluated in noisy and reverberant conditions.

WWB algorithm (D.L.Wang et al., 2003) and Jin et al. algorithm (Jin and Wang, 2011)

are used for objective comparison. The algorithm of Wu, Wang, and Brown is referred

to as the WWB algorithm. WWB algorithm integrates a channel-peak selection method

and HMM for forming continuous pitch tracks. WWB framework computes final pitch

estimates in three stages: auditory front-end-processing, pitch statistical modelling and

HMM tracking. Jin et al. algorithm, designed specially to tackle reverberant noises is

similar to WWB algorithm but different in channel selection and pitch scoring strategy.

In (D.L.Wang et al., 2003; Jin and Wang, 2011), half of the corpus is used to estimate

the model parameters for further analysis. Another important fact about the experiments

reported in (D.L.Wang et al., 2003) is that they are focused on speech mixtures with one

dominant speaker. Both algorithms report considerable transition errors, in which pitch

estimates of speaker-1 are misclassified as the pitch estimates of speaker-2.

For a fair comparison, the WWB and Jin et al. algorithm outputs are grouped in the

post processing stage to ensure no transition errors occur (Jin and Wang, 2011). The

grouping is performed using an approach similar to the approach proposed in (Rad-

far et al., 2011). The results obtained through the quantitative evaluation are listed in

Tables 4.3 - 4.8. Tables 4.3 and 4.4 compare the pitch accuracies with 10% tolerance

for dataset-1 and dataset-2, respectively. Jin et al. algorithm and MODGD algorithm

show similar performance with MODGD having a slight edge. It was also observed

that WWB algorithm failed to pick one of the pitch estimates in many frames. The

proposed method reports accuracies of 84.58% and 82.65% for dataset-1 and dataset-2

respectively in clean cross-gender mixtures. It is worth noting that, in babble and white

noise conditions, the performance of MODGD system is at par with that of Jin et al.

algorithm while it is superior in performance when compared with that of WWB algo-

87

Page 103: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

rithm. From Tables 4.3 and 4.4, it is seen that as SNR decreases from 10 dB to 5 dB,

the proposed algorithm is more robust than WWB and Jin et al. algorithms.

When the gender mixture is made up of the same sex, if the pitch values are too

close, the performance of the proposed algorithm is affected due to the filtering opera-

tion. Figure 4.10 shows a situation where the speech mixture comprises of speakers with

close individual pitches. During the cancellation process, the second pitch is also elim-

inated and results in a spurious pitch estimate during the next iteration. In the proposed

group delay based system, both noise and source introduce zeroes that are close to the

unit circle in the z domain (Murthy, 1991; Hegde, Rajesh M., 2005). The fundamental

difference is that source zeroes are periodic while noise zeroes are aperiodic. This is

the primary reason for the better performance of group delay in noisy environments as

compared to other model based methods. Tables 4.5 and 4.6 compare the standard de-

Table 4.5: Comparison of Efs corresponding to Accuracy10 (in Hz) (Dataset-1)

Category Efs

MODGD WWB JIN

Male-Female 4.80 1.81 3.12

Female-Female 5.43 2.12 4.05

Male-Male 4.82 1.52 2.75

Male-Female,Babble noise(10 dB SNR) 5.40 2.17 3.07

Male-Female,Babble noise (5 dB SNR) 6.12 3.12 3.47

Male-Female, White noise (10 dB SNR) 5.47 2.01 3.35

Male-Female, White noise (5 dB SNR) 6.44 3.70 3.70

Male-Female with reverberation 4.98 2.87 3.76

Table 4.6: Comparison of Efs corresponding to Accuracy10 (in Hz) (Dataset-2)

Category Efs

MODGD WWB JIN

Male-Female 3.48 1.59 2.37

Female-Female 3.77 2.17 3.28

Male-Male 3.78 1.43 2.45

Male-Female, Babble noise (10 dB SNR) 3.72 1.61 2.64

Male-Female, Babble noise (5 dB SNR) 4.58 1.83 3.16

Male-Female, White noise (10 dB SNR) 4.10 1.91 3.10

Male-Female, White noise (5 dB SNR) 5.55 1.95 3.18

Male-Female with reverberation 4.34 2.55 3.00

viation of fine pitch error (Efs) for dataset-1 and dataset-2, respectively. While WWB

and Jin et al. algorithms report Efs in the range 2 - 4.5 Hz across categories, the pro-

posed algorithm reports higher error. This is primarily due to the resolution that results

from the length of the flattened signal. The performance of the proposed system with

varying TMR is shown in Table 4.7 and Table 4.8. The analysis shows a similar trend

in the performance of both MODGD algorithm and Jin et al. algorithm. A considerable

88

Page 104: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

variation in accuracy is reported in the case of WWB algorithm as TMR decreases from

-3dB to 3dB, but Jin et al. and the proposed MODGD algorithm show consistently

good performance.

Table 4.7: Comparison of accuracy for different TMR : Database:1

Category MODGD WWB JIN

Accuracy10(in%) Efs Accuracy10(in%) Efs Accuracy10(in%) Efs

Male-Female 0dB 84.58 4.80 76.71 1.81 81.04 3.12

Male-Female -3dB 83.32 4.93 72.90 1.80 79.65 3.74

Male-Female +3dB 85.52 4.85 79.40 1.71 82.21 3.42

Table 4.8: Comparison of accuracy for different TMR : Database:2

Category MODGD WWB JIN

Accuracy10(in%) Efs Accuracy10(in%) Efs Accuracy10(in%) Efs

Male-Female 0dB 82.65 3.48 78.92 1.59 78.95 2.37

Male-Female -3dB 82.72 3.46 71.02 1.73 76.71 2.05

Male-Female +3dB 82.26 3.49 74.37 1.61 77.61 2.10

15 20 25 30 35 40 45

50

100

150

200

250

300

350

Frame Index

Pitc

h (H

z)

PITCH1PITCH2REFPITCH1REFPITCH2

Figure 4.10: Speakers with two closed pitch trajectories

4.9 Pitch estimation : more than two speakers

The iterative method can be extended for pitch estimation of mixed speech with more

than two speakers. This is illustrated with an example of mixed speech with three

speakers. Consider a speech mixture of three synthetic speech signals with fundamental

frequencies of 100 Hz, 200 Hz and 280 Hz. Modified group delay function computed

for a frame of speech mixture is shown in Figure 4.11 (a). The peaks in MODGD

feature space, correspond to the pitch candidates present in the speech mixture and its

integral multiples. In the first pass, the prominent peak in the MODGD feature space

89

Page 105: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

is mapped to first pitch estimate followed by the annihilation of it from the residual

spectrum. The red color contour in Figure 4.11 (a) is the computed MODGD for the

second pass. The green color contour in Figure 4.11 (a) is the computed MODGD for

the third pass. The individual pitch estimates computed, in all the iterations are plotted

in Figure 4.11 (b) along with reference pitches. The individual pitches estimated using

the proposed method for a three speaker natural mixed speech is shown in Figure 4.12.

20 40 60 80 100 120 140

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

1.8

2

2.2

x 10−5

Time (ms)

MO

DG

D (s

ec)

(a)

MODGD−First PassMODGD−Second PassMODGD−Third Pass

0 5 10 15 20 25 3050

100

150

200

250

300

Frame Index

Pitc

h (H

z)

(b)

REF Pitch1REF Pitch2REF Pitch3MODGD Pitch1MODGD Pitch2MODGD Pitch3

Figure 4.11: (a) Modified group delay function computed for a speech frame, (b) Pitch

contours computed for the speech segment.

80 90 100 110 120 130

50

100

150

200

250

Frame Index

Pitc

h (H

z)

REF Pitch1REF Pitch2REF Pitch3MODGD Pitch1MODGD Pitch2MODGD Pitch3

Figure 4.12: Pitch contours computed for a natural speech mixture with three speakers.

90

Page 106: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Table 4.9: Pitch estimation accuracy for three speaker mixed speech.

Sl.No Speech MixtureAccuracy

(%)

1 Two females, One male 62.72

2 Two males, One female 58.72

3 Three females 50.47

4 Three males 46.53

The performance of the proposed group delay based iterative method was evaluated

using natural three-speaker mixed speech created using GRID subset. Four categories

of speech mixtures were considered for the experiment, which include (2 females,1

male), (2 males, 1 female), (3 females) and (3 males) speech mixtures. The results are

tabulated in Table 4.9. The overall accuracy of 54.61% is reported for the entire subset.

It is worth noting that, in three speaker case, the experiment is conducted only for pitch

estimation, not for tracking. Speaker specific characteristics are needed for accurate

tracking of individual pitch contours. Due to iterative cancellation using comb filters, if

one of the frequencies in the mixture is a multiple of the other, it may be attenuated on

the fly.

Summary

Phase based approach for two-pitch tracking is presented, and it yields competitive per-

formance as compared to other state-of-the-art approaches. The flattened spectrum is

processed using MODGD algorithm to estimate the predominant pitch in each frame in

the first pass. Then the estimated pitch and its harmonics are filtered out using comb

filter. In the second pass, the residual spectrum is again analysed using MODGD al-

gorithm to estimate the second candidate pitch. The pitch grouping stage followed by

the post processing step result in the final pitch trajectories. The performance of the

proposed algorithm was evaluated on mixed speech with cross gender (Female, Male),

same gender (Male/Male, Female/Female) mixtures. It does not require pre-training on

source models from isolated recordings. The experiment can be extended for multip-

itch estimation, but for pitch tracking, additional information such as speaker-specific

characteristics are needed.

91

Page 107: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

CHAPTER 5

The Source-MODGD Coefficient Features (SMCF) in

Speech/Music Analysis

5.1 Introduction

Over the years, researchers in acoustics have employed many feature extraction tech-

niques to parametrize the speech signal in the front-end of numerous speech based

applications. Parametric representation has an important role in speech processing. The

objective of feature extraction is two-fold: a) compression, and b) discrimination. It

should also be invariant to irrelevant factors such as speaker variations, accent differ-

ences, speaking rates and background noises (Xiong, 2009). Most of the time, the

feature extraction methods are application specific, features performing well in some

applications, may not perform well in others; for example the feature extraction meth-

ods used for speech or speaker recognition are quite different to those used in MIR

applications. Feature extraction methods in time, frequency and cepstral domain have

been successfully utilized in various speech based applications. As discussed earlier,

group delay based features have also been extensively used in speech and music pro-

cessing applications. In this chapter, a variant of group delay feature is proposed and

the effectiveness of the proposed feature is evaluated for speech-music detection and

estimating number of speakers in mixed speech. Section 5.2 introduces a new feature

derived from the flattened spectrum and lists the steps to compute the same. The task of

estimating number of speakers in a mixed speech using new feature is explained in Sec-

tion 5.3, along with dataset and performance evaluation. Speech-music classification

and detection using the proposed feature is described in Section 5.4.

5.2 Source-MODGD coefficient features(SMCF)

The success of MFCC, combined with its robust and cost-effective computation, turned

it into a standard choice in speech recognition applications. The cepstral representa-

Page 108: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

tion of the speech spectrum provides a good representation of the spectral properties

of the signal. Cepstral features are widely used in speech recognition, speech analysis

and speech enhancement applications. Linear prediction cepstral coefficients (LPCCs),

Bark frequency cepstral coefficient (BFCC), auditory periphery model based gamma-

tone frequency cepstral coefficient (GFCC), have been utilized in many speech process-

ing applications.

Earlier, the MODGDF has been effectively employed in speech recognition, speaker

recognition (Hegde, Rajesh M., 2005), language recognition and emotion recognition

(Dey et al., 2011). Apart from the MODGDF feature (Hegde, Rajesh M., 2005), the

proposed feature is derived from the flattened power spectrum. Group delay analysis on

the flattened power spectrum leads to excitation components by annihilating the system

characteristics.

Since the flattened spectrum consists of source information of multiple speakers,

peaks at multiples of pitch periods and its algebraic combinations can be seen in MODGD

feature space. Since the musical frame consists of many pitch components due to ac-

companying instruments other than the singer, the MODGD feature space will have

peaks at multiples of each of the pitch periods. This can be observed in Figure 5.1. Fig-

ure 5.1 (a) shows MODGD on flattened power spectrum for a music frame and Figure

5.1 (b) shows MODGD on flattened power spectrum for a speech frame. This charac-

teristics can be used as a criteria to distinguish music from speech. The modified group

delay functions derived from the flattened spectrum is converted to useful features using

DCT. The new feature representation, source-modified group delay coefficient features

(SMCF), c[n] are obtained by (Murthy and Rao, 2003),

c[n] =

k=Nf∑

k=0

τsm[k] cos(n(2k + 1)π

Nf

) (5.1)

where Nf is the discrete Fourier Transform order and τsm[k] is the source modified

group delay.

Steps to compute Source-MODGD Coefficient features (SMCF) are summarized

below.

• Frame blocking the speech/music signal at a frame size of 20 ms and frame shift

of 10 ms.

93

Page 109: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

5 10 15

0.005

0.01

0.015

0.02

0.025

0.03

(a)

Time (ms)

MO

DG

D (

sec)

5 10 15 200

0.05

0.1

0.15

(b)

Time (ms)

MO

DG

D (

sec)

Figure 5.1: (a) MODGD on flattened power spectrum of music frame, (b) MODGD on

flattened power spectrum of a speech frame

• Speech/music spectrum is flattened using root cesptral smoothing to annihilate

the system characteristics.

• Apply MODGD algorithm on the flattened power spectrum to compute modified

group delay function.

• Compute DCT on MODGD spectra to get the SMCF vectors.

5.3 Estimating number of speakers in multipitch envi-

ronment

Meetings with a number of individuals, can lead to strong interference from several

speakers. The information about the number of speakers in such a situation is useful

in blind source separation (Emiya et al., 2008), and pitch trajectory estimation in mul-

tipitch environment (Wu and Wang, 2003). Estimating the number of speakers plays a

crucial role in speaker diarization.

5.3.1 Related work

During radio talk shows, at some point it happens that all the guests speak simulta-

neously with varying intonation. The ability to track the locations of intermittently

speaking multiple speakers in the presence of background noise and reverberation is of

great interest due to the vast number of potential applications (Quinlan et al., 2009).

Hershey et al, made use of a model-based speaker identification (SID) module, called

Iroquois (Hershey et al., 2010) to identify speakers existing in the mixture. In (Shao

94

Page 110: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

et al., 2010), CASA based approach is employed to detect the number of speakers in

the speech mixture. MAP criterion is also proposed to detect number of speakers at

frame level in model based single channel source separation (SCSS) (Mowlaee et al.,

2010). Several sensor based methods were proposed in the literature for detection of the

number of sources whose mixed signals are collected from multiple sensors. Multiple

sensor based methods rely on the time delays of arrival of speech signals between the

two microphones for a given speaker (Swamy et al., 2007; Kumar et al., 2011).

Pitch prediction features (Lewis and Ramachandran, 2001), the temporal evolu-

tion of the 2-D Average Magnitude Difference Function (AMDF) (Vishnubhotla and Espy-

Wilson, 2008), modulation characteristics (Arai, 2003), the statistical characteristic of

the 7th Mel coefficient (Sayoud and Ouamour, 2010) are also utilized in estimating

number of speakers. In (Lewis and Ramachandran, 2001), pitch prediction feature

(PPF) is used in identifying temporal regions or frames as being either one-speaker or

two-speaker speech. A method for multi-speaker tracking using distributed nodes with

microphone arrays is described in (Plinge and Fin, 2014). Since the modulation peak

decreases as number of speakers increases, algorithm in (Arai, 2003) uses the region of

the modulation frequency between 2 and 8 Hz to estimate the number of speakers. In

(Emiya et al., 2008), distance between a speech mix and a single speaker reference is

used as the metric to compute number of speakers in a speech mix.

5.3.2 Proposed system

In the proposed method, we start from the computation of SMCF from the mixed speech.

In the feature extraction phase, 20 dimensional SMCF vectors are frame wise computed.

Gaussian mixture model based classifier is used in the classification phase. In the simple

Gaussian classifier, each probability density function (PDF) is assumed to be a mul-

tidimensional Gaussian distribution whose parameters are estimated using the training

set. The iterative expectation-maximization (EM) algorithm is used to estimate the pa-

rameters of each Gaussian component and the mixture weights (Munoz-Exposito et al.,

2007). In the training phase, feature vectors computed from the training set are used to

build models for one-speaker case, two-speaker case and three-speaker case. 64 com-

ponent GMM is used to model the different classes of speech mixtures. 60% files of the

dataset is used for training and the rest for testing. During the testing phase, the clas-

95

Page 111: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

sifier evaluates the log-likelihoods of the unknown speech mixture data against these

models. The model that gives the maximum accumulated likelihood is declared as the

correct match.

5.3.3 Performance evaluation

The performance of the proposed algorithm is evaluated on speech mixtures generated

by the subset of GRID dataset. A subset of 40 audio files are used for generating mix-

tures for the evaluation. Additive mixtures of the audio files are created using audacity.

The test is conducted on speech mixtures, generated with 0 dB target to interference

ratio (TIR). The results are tabulated as a confusion matrix in Table 5.1. The results

show that the proposed feature is very effective in discriminating different speakers in

a mixed speech. Overall accuracy of the estimation of the number of speakers in the

multipitch environment is 74.66%. As the number of speakers increases, more misclas-

sification may be observed. The feature can be made robust using delta and delta-delta

coefficients to tackle such situations.

Table 5.1: Confusion matrix for number of speaker estimation task. SP-1, SP-2 and

SP-3 denote one speaker case, two speaker case and three speaker case, re-

spectively. Class wise accuracy is given as the last column entry.

SP-1 SP-2 SP-3 Accuracy(%)

SP-1 23 2 1 89SP-2 3 15 8 58SP-3 1 5 20 77

5.4 Speech/music classification and detection

Another potential application of the proposed feature is in audio indexing. In video and

audio databases, archiving is an important task. Presently such archives are indexed

manually and can only be accessed through the tedious effort of human experts, whose

availability is limited. The important task that is to be carried out is to split the audio

file into silence/music/crowd noise segments.

Low bit-rate audio coding can also benefit from speech/music classification. Sepa-

rate codec designs are used in digital encoding of speech and music. Design of a multi-

96

Page 112: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

mode coder that can accommodate different signals can be efficiently implemented, if

speech/music classifier gives appropriate decision on an audio segment (El-Maleh et al.,

2000). Segregating the signal into speech and music segments is an obvious first step

before applying speech-specific or music-specific algorithm. Speech-Music classifica-

tion is defined as the binary problem of classifying pre-segmented audio data to the

speech or music class. The detection task consists of finding segments of music and

speech in a signal and classifying each segment as music or speech. Speech-music de-

tection can be used to segment recordings, which are typically several minutes long in

duration.

5.4.1 Related work

Saunders uses statistics of the energy contour and zero-crossing rate to discriminate

speech and music (Saunders, 1996). Eric Scheirer and Malcolm Slaney addressed au-

dio segmentation by utilizing various combinations of 13 features (Scheirer and Slaney,

1997). Pfeiffer et al (Pfeiffer and Effelsberg, 1996) use perceptual criteria by matching

characteristics such as amplitude, pitch and frequency. A supervised tree-based vec-

tor quantizer trained to maximize mutual information (MMI) was attempted in (Foote,

1997). The proposed system in (Samouelian, 1997), uses information theory to con-

struct a decision tree from several different TV programs. During the training phase,

the feature extraction framework extracts features from the continuous audio signal

on a frame-by-frame basis and uses these features to train a decision tree (Quinlan,

1993). Low frequency modulation (E. Scheirer, 1997), entropy and dynamism features

(E. Scheirer, 1997) have also been employed for the audio indexing task.

5.4.2 Speech/music classification

The proposed feature can be effectively utilized for audio indexing in multimedia ap-

plications. In the proposed approach, 40 dimensional feature vectors comprising SMCF

and delta-SMCF are computed from the audio files and classified using a trained GMM.

In Figure 5.2, 20 dimensional SMCF feature vectors for entire frames of speech/music

segment are mapped into a two dimensional space. It can be seen that SMCF features

are effective in differentiating speech from music.

The GTZAN music/speech data collection (DB1) is used for performance evalua-

97

Page 113: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

tion of the proposed task. The dataset consists of 120 tracks, each 30 seconds long.

Each class (music/speech) has 60 examples. The tracks are all 22050 Hz Mono 16-bit

audio files in .wav format. The files were collected from a variety of sources including

personal CDs, radio, microphone recordings, in order to represent a variety of recording

conditions. In speech/music classification task, an audio file is classified into speech/

music using GMM framework. 60% of the data is used for training the GMM and the

rest for testing during classification stage. The decision to classify an audio file is taken

based on the likelihood of entire frames in the utterance. Results are tabulated in Table

5.2. All speech files are identified correctly. The proposed system reports an overall

accuracy of 95%.

−1.5 −1 −0.5 0 0.5 1

x 104

−2.5

−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

x 104

SpeechMusic

Figure 5.2: Two dimensional visualization of speech-music data with SMCF feature us-

ing Sammon mapping

Table 5.2: Confusion matrix for audio indexing task. Class wise accuracy is given as

the last column entry.

Speech Music %

Speech 26 0 100Music 2 18 90

5.4.3 Speech/music detection

In speech/music detection, the earlier experiment is extended in such a way that system

should mark speech/music segments in an audio material. In speech/music detection,

experiment is performed on 4 broadcasting materials each of 4 minutes duration (DB2),

using a GMM classifier, trained using the SMCF. 30 minutes of speech/music data, col-

98

Page 114: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Table 5.3: Results on Speech/Music detection on audio recordings. Correctly indexed

speech/music frames in each recording is entered in %. Global accuracy is

given as the entry in the last column.

Speech (%) Music (%) Global (%)

Audio recording 1 95.09 74.00 84.54Audio recording 2 83.80 90.52 87.16Audio recording 3 76.38 88.34 82.36Audio recording 4 88.74 78.00 83.37

lected from T.V musical shows are used for training the model. The data is manually

annotated.

Frame level evaluation

In frame level, the percentage of the audio frames which are correctly recognized is

estimated (Theodorou et al., 2012). Frame-based evaluation is carried out on 10 ms

segments. It has been observed that silence frames or frames with less pitch content in

both the speech and music segments cause overlapping in feature space. To overcome

such situations, a frame is classified into speech/music based on the cumulative decision

from the neighboring frames up to 0.25 s segment length. To improve the performance

accuracy of the speech/music detection, a post-decision processing step is implemented

to minimize switching (upto 1 s) between the two classes in a long speech/music seg-

ment. Since switching of small segments in between long speech/music is less prob-

able, rules are formulated to avoid such switching in decision making. The preceding

and succeeding decisions are used for decision processing of the present segment. Re-

sults are tabulated in Table 5.3. From the table, it can be seen that the proposed method

could attain global accuracy greater than 80% for all the broadcasting materials.

Event level evaluation:

Events 1 will be evaluated on an onset-only basis as well as an onset-offset basis using

the precision, recall, and F-Measure. The performance metrics are selected as per the

MIREX standards used for speech/music evaluation task in 2015 2. The performance

metrics are given by :

1 An event is a transition from speech to music or music to speech2www.music-ir.org/mirex/wiki/2015:Music/Speech-Classification-and-DetectionEvaluation

99

Page 115: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

0 20 40 60 80 100 120 140 160 180−1

−0.5

0

0.5

1

Time (sec)

Am

plitu

de

(a)

0 20 40 60 80 100 120 140 160 1801

1.5

2

Time (sec)

Spe

ech/

Mus

ic

(b)

SMCFReference

Figure 5.3: Performance evaluation of an audio sample, 1 stands for speech and 2 stands

for music

• fb_F: Frame level F-measure.

• Onset-only event-based F-measure-500ms tolerance (eb_F_500ms) : A ground

truth segment is assumed to be correctly detected if the system identifies the right

class (Music/Speech) and the detected segment’s onset is within a 1000ms range

( +/- 500ms) of the onset of the ground truth segment or within 20% of the ground

truth segment’s length.

• Onset-offset event-based F-measure-500ms tolerance (eb_Foff_500ms) : A

ground truth segment is assumed to be correctly detected if the system identifies

the right class (Music/Speech), and the detected segment’s onset is within +/-

500ms of the onset of the ground truth segment, and the detected segment’s offset

is either within +/- 500ms of the offset of the ground truth segment or within 20%

of the ground truth segment’s length.

• Onset-only event-based F-measure-1s tolerance (eb_F_1s) : A ground truth

segment is assumed to be correctly detected if the system identifies the right class

(Music/Speech) and the detected segment’s onset is within ( +/- 1000ms) of the

onset of the ground truth segment or within 20% of the ground truth segment’s

length.

• Onset-offset event-based F-measure-1s tolerance (eb_Foff_1s) : A ground

truth segment is assumed to be correctly detected if the system identifies the right

class (Music/Speech), and the detected segment’s onset is within +/- 1000ms of

the onset of the ground truth segment, and the detected segment’s offset is either

within +/- 1000ms of the offset of the ground truth segment or within 20% of the

100

Page 116: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

ground truth segment’s length.

The results of speech/music detection are shown in Table 5.4. The algorithm pro-

posed by Panag et al. (C. Panagiotakis, 2005) and MFCC based system are also shown

for comparison. In (C. Panagiotakis, 2005), two signal characteristics are used: the am-

plitude, measured by the root mean square (RMS), and the mean frequency, measured

by the average density of zero-crossings. F-measure of the proposed SMCF based sys-

tem is higher than MFCC based system but less as compared to the scheme of Panag et

al. Figure 5.3, shows an example of speech/music detection in an audio file, using the

proposed method.

Table 5.4: Performance comparison of Speech/Music detection in database, DB2

Method fb_F eb_F_500ms eb_Foff_500ms eb_F_1s eb_Foff_1s

Panag et al. 0.8385 0.7040 0.6640 0.7040 0.6640

MFCC 0.6036 0.2690 0.2325 0.2690 0.2325

SMCF 0.6255 0.2821 0.1980 0.2821 0.1980

Summary

A variant of group delay feature derived from the flattened spectrum is introduced in

this chapter. The power spectrum is flattened using cepstral smoothing to annihilate

the system characteristics. The flattened power spectrum is analyzed using MODGD

algorithm followed by DCT, to convert to SMCF features. The potential of the new

feature is experimentally evaluated on diverse datasets by two experiments, namely

audio indexing task and estimating the number of speakers in mixed speech. The results

demonstrate the potential of the new feature in speech/music signal processing and

content based information retrieval applications.

101

Page 117: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

CHAPTER 6

Summary and Conclusions

6.1 Summary

The group delay function which is defined as the negative derivative of the unwrapped

short-time phase spectrum, can be computed directly from the speech signal, without

unwrapping the short-time phase spectrum. The group delay function has been effec-

tively used to extract various source and system parameters. These features have also

been gainfully converted to appropriate features for speech synthesis/recognition.

Three tasks in speech/music processing have been addressed using modified group

delay functions. They are the predominant melodic pitch estimation in polyphonic mu-

sic, multipitch estimation in mixed speech and speech/music detection in audio record-

ings. The features extracted from the computed melodic contour is also used in au-

tomatic music genre classification. A new feature is derived from the flattened power

spectrum and is effectively used in speech/music discrimination and estimation of num-

ber of speakers in mixed speech.

In melody extraction, two paradigms are proposed. In MODGD (Direct), modified

group delay analysis is performed directly on the music signal to compute predominant

melodic pitch. In this approach, the proper selection of cepstral window emphasizes

the peak corresponding to pitch in the group delay spectrum. The search space in this

algorithm is limited by the first formant frequency. For high pitched voices, the method

ends-up with erroneous pitch estimate, where a low formant can be confused for the

pitch frequency. To overcome the stringent constraints in the first approach, group delay

analysis on flattened spectrum MODGD (Source) is proposed. The spectrum flattening

annihilates the system characteristics and MODGD analysis is performed on the flat-

tened spectrum. Transient analysis is incorporated to capture the dynamic variation of

melody. The consistency is ensured by dynamic programming. Median filtering is per-

formed to get a refined melodic pitch contour in the post-processing phase. Automatic

music genre classification using fusion of melodic features and group delay features

Page 118: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

is also addressed. Results show the promise of the complementary nature of melodic

feature in automatic music genre classification.

In the second part, the thesis addresses the problem of multipitch estimation using

group delay function in clean and noisy conditions. Iterative estimation and cancellation

technique is adopted in the proposed method. The predominant pitch is estimated and

the estimated component along with harmonics are cancelled out from the flattened

spectrum using comb filters. Pitch grouping and median filtering is performed in the

post-processing stage to compute the individual pitch trajectories.

In the third part, a variant of group delay feature is introduced. The new feature

representation is derived from the flattened power spectrum after decorrelating using

discrete cosine transform. This feature is used in two tasks, namely, estimation of

number of speakers in mixed speech and speech-music detection.

The major contributions of the thesis are as follows:

• Two algorithms based on modified group delay functions are proposed for melodic

pitch estimation in polyphonic music.

• The complementary nature of high level features derived from melodic pitch con-

tour is studied extensively for automatic music genre classification.

• A new approach based on modified group delay processing, exploiting the high

resolution property is proposed for two-pitch tracking.

• A variant of group delay feature is introduced and two applications of the new

feature, namely estimating number of speaker in mixed speech and speech/music

detection are addressed.

6.2 Criticism

The following are the criticisms regarding the work presented herein.

• We have observed significant gap in raw pitch accuracy and raw chroma accuracy

in many audio files. Efficient octave error minimization techniques should be

incorporated in the post processing phase to reduce the chroma error in melodic

pitch estimation.

• In multipitch estimation, pitch trajectory estimation in the case of cross trajectory

is not addressed. Since dynamic programming fails at the point of pitch cross-

ing, methods utilizing speaker characteristics should be incorporated to track the

103

Page 119: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

pitch trajectories. In addition, pitch estimation for more than two speakers is ad-

dressed, but for tracking, speaker-specific characteristics should also be employed

in multipitch environment.

104

Page 120: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

APPENDIX A

Mel-frequency cepstral coefficients

Mel Frequency Cepstral Coefficients (MFCCs) are a feature widely used in automatic

speech and speaker recognition. It was introduced by Davis and Mermelstein in the

1980’s, and have been state-of-the-art ever since. According to psychophysical studies,

human perception of the frequency content of sound follows subjectively defined non-

linear scale called the mel scale. The subjective pitch in mels, fmel corresponding to f

can be computed by,

fmel = 2595 log10

(

1 +f

700

)

(A.1)

This leads to the definition of MFCC, a baseline acoustic feature set for speech and

speaker recognition applications.

Let {y[n]}Nn=1 represent a frame of speech that is pre-emphasized and Hamming

windowed. First y[n] is converted into frequency domain by an Ms-point (DFT) which

leads to the energy spectrum.

|Y [k]|2 =

∣∣∣∣∣

N∑

n=1

y[n].e(−j2πnk

Ms)

∣∣∣∣∣

2

(A.2)

where l ≤ k ≤Ms.

A filter bank, uniformly spaced in mel scale is imposed on the spectrum calculated

in Equation A.2. The outputs eiQi=1 of the Mel-scaled band-pass filters can be calculated

by a weighted summation between respective filter response ψi[k] and energy spectrum

|Y [k]|2 as

ei =

Ms2∑

k=1

|Y [k]|2 ψi[k] (A.3)

Finally, Discrete Cosine Transform (DCT) is taken on the log filter bank energies

log(e[i])Qi=1 and the final MFCC coefficients Cm can be written as

Cm =

√2

Q

Q−1∑

l=0

log(e[l + 1]) cos

[

m

(2l − 1

2

Q

]

(A.4)

Page 121: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

where 0 ≤ m ≤ Rcep − 1, and Rcep is the desired number of cepstral coefficients.

Normally, Q = 20 and 10 to 30 cepstral coefficients are taken for speech processing

applications. The first coefficient C0 is discarded because it represents spectral energy.

106

Page 122: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

APPENDIX B

Melody extraction vamp plug-in

A graphical user interface (GUI) for automatic melody extraction from polyphonic mu-

sic is inevitable in many music processing applications. Melody based reference tem-

plates required for the searchable database in query by humming systems must be ex-

tracted from polyphonic soundtracks.

Publicly available pitch detection interfaces, like PRAAT 1 and the Aubio pitch de-

tector 2 for the Sonic Visualizer program 3 have been designed for monophonic audio

and cannot be expected to perform acceptably well on polyphonic audio. However, a

few polyphonic transcription tools also exist. The polyphonic transcription VAMP plu-

gin 4 has been designed exclusively for guitar and piano music and outputs MIDI notes.

Melodyne’s Direct Note Access (DNA) 5 attempts to isolate simultaneously played mu-

sical notes but is not freely available for evaluation. MELODIA (Salamon and Gomez,

2012), is a newly proposed GUI with more user friendly approaches for melody extrac-

tion from polyphonic music.

A plug-in is developed based on the proposed method of melody extraction as part

of the work. In the proposed method, the music signal is flattened to annihilate sys-

tem characteristics and the residual signal is analyzed using group delay algorithm to

estimate the predominant pitch contour. Plug-in can be used to visualise the prominent

melody of an audio recording, using Sonic Visualiser vamp host. It allows the user to

load an audio file, extract the melody and visualise its pitch sequences. The output is

the melodic contour estimated with hop-size specified. Each row of the output contains

a time stamp and the corresponding frequency of the melody in Hertz. Two snapshots

of the interface are provided in Figures B.1 and B.2. It consists of a waveform viewer,

a spectrogram viewer, a menu bar, controls for audio viewing, scrolling and playback

and a parameter window.

1http://www.praat.org2http://aubio.org3http://www.sonicvisualiser.org4http://isophonics.net/QMVampPlugins5http://www.celemony.com/cms/index.php

Page 123: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Figure B.1: Selection of plug-in on Sonic Visualiser

108

Page 124: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

Figure B.2: Melodic sequences of proposed algorithm along with MELODIA refer-

ence.

109

Page 125: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

REFERENCES

1. Abeysekera, S. S. (2004). Multiple pitch estimation of poly-phonic audio signals in a

frequency-lag domain using the bispectrum. Proceedings of International Symposium

on Circuits and Systems, 3(911), 469–472.

2. Adalbjornsson, S. I., A. Jakobsson, , and M. G.Christensen (2013). Estimating mul-

tiple pitches using block sparsity. Proceedings of IEEE Int. Conf. Acoust., Speech, and

Signal Processing, 6220–6224.

3. A.Kurematsu, K.Takeda, Y.Sagisaka, S. Katagiri, H.Kuwabara, and K.Shikano

(1995). ATR japanese speech database as a tool of speech recognition and synthesis.

Speech Communication, 9(4), 357–363.

4. Allen, J. B. and D. A. Berkley (1979). Image method for efficiently simulating small-

room acoustics. J. Acoust. Soc. Amer, 65(4), 943–950.

5. Anand Joseph, M., S. Guruprasad, and B. Yegnanarayana (2006). Extracting for-

mants from short segments of speech using group delay functions. Proceedings of Int.

Conf. Spoken Language Processing, 1009–1012.

6. Arai, T. (2003). Estimating number of speakers by the modulation characteristics of

speech. Proceedings of IEEE Int. Conf. Acoust., Speech, and Signal Processing, 2,

197–200.

7. Arora, V. and L. Behera (2013). On-line melody extraction from polyphonic audio

using harmonic cluster tracking. IEEE Transactions on Audio Speech and Language

Processing, 21(3), 520–530.

8. Aryafar, K., S. Jafarpour, and A. Shokoufandeh (2012). Music genre classification

using sparsity-eager support vector machines. Proceedings of International Conference

on Pattern Recognition (ICPR), 1526–1529.

9. Badeau, R., V. Emiya, and B. David (2009). Expectation maximization algorithm for

multi pitch estimation and separation of overlapping harmonic spectra. Proceedings of

IEEE Int. Conf. Acoust., Speech, and Signal Processing, 11(6), 3073–3076.

10. Barbedo, J. G. A. and A. Lopes (2007). Automatic genre classification of musical

signals. EURASIP Journal on Advances in Signal Processing, Article ID 64960.

11. Bastys, A., A. Kisel, and B. Salna (2010). The use of group delay features of linear

prediction model for speaker recognition. INFORMATICA, 21, 1–12.

12. Bello, J. P., L. Daudet, S. Abdallah, C. Duxbury, M. Davies, and M. B.Sandler

(2005). A tutorial on onset detection in music signals. IEEE Transactions on Speech

and Audio Processing, 13(5), 1035–1047.

13. Bello, J. P., G. Monti, and M. Sandler (2000). Techniques for automatic music tran-

scription. Proceedings of International Society for Music Information Retrieval Con-

ference (ISMIR 2000).

110

Page 126: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

14. Bellur, A. (2013). Automatic Tonic Identification in Indian Classical Music. M.S.thesis,

Indian Institute of Technology, Department of Electrical Engg., Madras, India.

15. Bellur, A. and H. A. Murthy (2013). A novel application of group delay function for

identifying tonic in carnatic music. Proceedings of 21st Europian Signal Processing

Conference, 1–5.

16. Benetos, E., S. Dixon, D. Giannoulis, H. Kirchhoff, and A. Klapuri (2012). Auto-

matic music transcription: Breaking the glass ceiling. Proceedings of 13th International

Society for Music Information Retrieval Conference, 379–384.

17. Berenguer, O., F. Casajus, and S. Torres (2005). Multiple piano note identification

using spectral matching method with derived patterns. Journal of Audio Engineering

Soc., 53, 32–43.

18. Bittner, R. M., J. Salamon, S. Essid, and J. P. Bello (2005). Melody extraction by

contour classification. Proceedings of 6th International Society for Music Information

Retrieval Conference, 500–506.

19. Bittner, R. M., J. Salamon, M. Tierney, M. Mauch, C. Cannam, and J. P. Bello

(2014). Medleydb: A multitrack dataset for annotation-intensive MIR research. Pro-

ceedings of the International Society for Music Information Retrieval (ISMIR) Confer-

ence, Taipei, Taiwan.

20. Bozkurt, B., L. Couvreur, and T. Dutoit (2007). Chirp group delay analysis of speech

signals. Speech Communication, 49(3), 159–176.

21. Brossier, P. M. (2005). Fast melody extraction using aubio. Extended abstract, 4th

Music Inform. Retrieval Evaluation eXchange (MIREX), 325–333.

22. C. Panagiotakis, G. T. (2005). A speech/music discriminator based on RMS and Zero-

crossings. IEEE Transactions on Multimedia, 7(1), 155–166.

23. Cancela, P. (2008). Tracking melody in polyphonic audio. Extended abstract, 4th

Music Inform. Retrieval Evaluation eXchange (MIREX).

24. Cao, C., M. Li, J. Liu, and Y. Yan (2007). Singing melody extraction in polyphonic

music by harmonic tracking. Proceedings International Society for Music Information

Retrieval, 373–374.

25. Casey, M. A., R. Veltkamp, M. Goto, M. Leman, C. Rhodes, and M. Slaney (2008).

Content-based music information retrieval: Current directions and future challenges.

Proceedings of the IEEE, 668–696.

26. Cherry, E. C. (1953). Some experiments on the recognition of speech, with one and

with two ears. J. Acoust.Soc.Amer, 25, 975–79.

27. Cheveigne, A. D. and H. Kawahara (2002.). Yin, a fundamental frequency estimator

for speech and music. J. Acoust. Soc. Amer., 111(4):1917–1930.

28. Christensen, M. and A. Jakobsson (2009). Multi-pitch estimation. Synthesis lectures

on speech and audio processing, B. Juang, Ed. SanRafael, CA: Morgan and Claypool.

29. Christensen, M., P. Stoica, A. Jakobsson, and S. H. Jensen (2008). Multi-pitch esti-

mation. Signal Processing, 88(4), 972–983.

111

Page 127: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

30. Cooke, M., J. Barker, S. Cunningham, and X. Shao (2006). An audio-visual corpus

for speech perception and automatic speech recognition. J. Acoust. Soc. Amer, 120,

2421–2424.

31. de Cheveigne, A. (1993). Separation of concurrent harmonic sounds: Fundamental

frequency estimation and a time domain cancellation model for auditory processing.

Journal of the Acoust. Soc. Am., 93(6), 3271–3290.

32. Deivapalan, P., M. Jha, R. Guttikonda, and H. A.Murthy (2008). DONLabel: An

automatic labeling tool for indian languages,. Proceedings of National Conference on

Communication, 263–266.

33. Dey, S., R. Rajan, R. Padmanabhan, and H. Murthy (2011). Feature diversity for

emotion, language and speaker verification. Proceedings of National Conference on

Communication(NCC), 1 – 5.

34. D.L.Wang, M.Wu, and G. Brown (2003). A multipitch tracking algorithm for noisy

speech. IEEE Transaction on Speech and Audio Signal Processing, 11, 229–241.

35. Downie, J. S. (2004). The scientific evaluation of music information retrieval systems:

Foundations and future. Computer Music Journal, 28, 12–23.

36. Dressler, K. (2011). An auditory streaming approach for melody extraction from poly-

phonic music. Proceedings of International Society for Music Information Retrieval

Conference, 19–24.

37. Durrieu, J., G. Richard, B. David, and C. Fevotte (2010). Source/filter model for un-

supervised main melody extraction from polyphonic audio signals. IEEE Transactions

on Audio, Speech, and Language Processing, 18(3), 564–575.

38. Dutta, S. and H. A. Murthy (2014). Discovering typical motifs of a raga from one-

liners of songs in carnatic music. Proceedings of 15th International Society for Music

Information Retrieval Conference (ISMIR), 397–402.

39. E. Scheirer, M. S. (1997). Construction and evaluation of a robust multi-feature

speech/music discriminator. Proceedings of IEEE International Conference on Acous-

tics, Speech, and Signal Processing, 1, 1331–1334.

40. E.Gomez, S.Streich, B. Ong, R. P. Paiva, S. Tappert, J. Batke, G. Poliner, D. Ellis,

and J. Bello (2006). A quantitative comparison of different approaches for melody

extraction from polyphonic audio recordings. Report ,Uniersity of Pompiu Fabra.

41. El-Maleh, K., M. Klein, G. Petrucci, and P. Kabalo (2000). Speech/music discrim-

ination for multimedia applications. Proceedings of IEEE Int. Conf. Acoust., Speech,

and Signal Processing, 2445–2448.

42. Ellis, D. P. and G. E. Poliner (2006). Classification-based melody transcription. Ma-

chine Learning Journal, 65(2-3), 439–456.

43. Emiya, V., R. Badeau, and B. David (2008). Automatic transcription of piano music

based on HMM tracking of jointly-estimated pitches. Proceedings of 16th European

Signal Processing Confeence (EUSIPCO-2008).

44. Field, S. E. (1998). Conceptual and representational issues in melodic comparison.

Computing in musicology: A directory of research, 11, 3–64.

112

Page 128: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

45. Foote, J. (1997). A similarity measure for automatic audio classification. Proceedings

of Spring Symposium on Intelligent Integration and Use of Text, Image, Video and Audio

Corpora, 1–7.

46. Gerhard, D. (2003). Pitch extraction and fundamental frequency:history and current

techniques. Technical Report TR-CS 2003-06.

47. G.Hu and D.L.Wang (2010). A tandem algorithm for pitch estimation and voiced

speech segregation. IEEE Transactions on Audio Speech Language Processing, 18(8),

2067–2079.

48. Giacobello, D., M. G. Christensen, M. N. Murthi, S. H. Jensen, and M. Moonen

(2012). Sparse linear prediction and its applications to speech processing. IEEE Trans-

actions on Audio, Speech, and Language Processing,, 20(5), 1644–1657.

49. Gomez, S., F. Canadas, J. S. J. Bonada, P. Vera, and P. Cabanas (2012). Predom-

inant fundamental frequency estimation vs singing voice separation for the automatic

transcription of accompanied flamenco singing. Proceedings of 13th International So-

ciety for Music Information Retrieval Conference, 601–606.

50. Goto, M. and S. Hayamizu (1999). A real-time music scene description system: De-

tecting melody and bass lines in audio signals. Working Notes of Workshop on Compu-

tational Auditory Scene Analysis, 31–40.

51. Gulati, S., A. Bellur, J. Salamon, H. Ranjani, I. Vignesh, H. A. Murthy, and

X. Serra (2014). Automatic tonic identification in indian art music:approaches and

evaluation. Journal of New Music Research, 43(1), 53–71.

52. Hari Krishnan P, S., R. Padmanabhan, and H. A. Murthy (2006). Robust voice

activity detection using group delay functions. Proceedings of IEEE International Con-

ference on Industrial Technology, 2603–2607.

53. Hegde, R. M., H. A. Murthy, and V. R. R. Gadde (2007a). Significance of joint fea-

tures derived from the modified group delay function in speech processing. EURASIP

Journal on Advances in Signal Processing, 2007, 1–3.

54. Hegde, R. M., H. A. Murthy, and V. R. R. Gadde (2007b). Significance of the mod-

ified group delay features in speech recognition. IEEE International Transactions on

Audio,Speech and Language Processing, 15, 190–202.

55. Hegde, Rajesh M. (2005). Fourier transform based features for speech recognition.

Ph.D Dissertation, Indian Institute of Technology Madras, Department of Computer

Science and Engg., Madras, India.

56. Hershey, J. R., S. J. Rennie, P. A. Olsen, and T. T. Kristjansson (2010). Super-human

multi-talker speech recognition: A graphical modeling approach. Elsevier Computer

Speech and Language, 24(1), 45–66.

57. Holzapfel, A. and Y. Stylianou (2008a). Beat tracking using group delay based onset

detection. Proceedings of International Society for Music Information Retrieval, 653–

658.

58. Holzapfel, A. and Y. Stylianou (2008b). Musical genre classification using nonneg-

ative matrix factorization-based features. IEEE Transactions on Audio, Speech, and

Language Processing, 16, 424–434.

113

Page 129: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

59. Hsu, C., D. Wang, and J. R. Jang (2011). A trend estimation algorithm for singing

pitch detection in musical recordings. Proceedings of IEEE Int. Conf. Acoust., Speech,

and Signal Processing, 393–396.

60. Hsu, C. L., L. Y. Chen, J. S. R. Jang, and H. J. Li (2009). Singing pitch extraction

fom monaural polyphonic songs by contextuual audio modeling and singing harmonic

enhancement. Proceedings of the 10th International Society for Music Information

Retrieval Conference, 201–206.

61. Hsu, C. L. and J. S. Jang (2010a). Singing pitch extraction by voice vibrato/tremolo

estimation and instrument partial deletion. Proceedings of International Society for

Music Information Retrieval, 525–530.

62. Hsu, C. L. and J. S. R. Jang (2010b). Singing pitch extraction by voice vibrato /

tremolo estimation and instrument partial deletion. Proceedings of the 11th Interna-

tional Society for Music Information Retrieval Conference, 525–530.

63. Hsu, C.-L., D. Wang, J.-S. R. Jang, and K. Hu (2012). A tandem algorithm for singing

pitch extraction and voice separation from music accompaniment. IEEE Trans. Audio,

Speech, and Language Processing, 20, 1482–1491.

64. Huang, F. and T. Lee (2013). Pitch estimation in noisy speech using accumulated peak

spectrum and sparce estimation technique. Proceedings of IEEE Int. Conf. Acoust.,

Speech, and Signal Processing, 21(1), 99–109.

65. Huang, Q. and D. Wang (2011). Multipitch estimation for speech mixture based on

multi-length windows harmonic mode. Proceedings of Fourth International Joint Con-

ference on Computational Sciences and Optimization (CSO), 345–348.

66. Ibnanez, A. P. (2010). Computationally effcient methods for polyphonic music tran-

scription. Ph.D Thesis, Departament de Llenguatges-i-Systemes Informatics, Univer-

sitst d’Alacant.

67. Ishwar, V. (2014). Pitch estimation of the predominant vocal melody from heterophonic

music audio recordings. Master Thesis, Universitat Pompeu Fabra, Barcelona, 2014.

68. Ishwar, V., S. Dutta, A. Bellur, and H. A. Murthy (2013). Motif spotting in an alapana

in carnatic music. Proceedings of 14th International Society for Music Information

Retrieval Conference (ISMIR), 499–504.

69. Jayesh, M. and C. Ramalingam (2014). An improved chirp group delay based algo-

rithm for estimating the vocal tract response. Proceedings of European Signal Process-

ing Conference (EUSIPCO), 2295–2299.

70. Jean-Marc and Christophe Ris, Developement of a speech recognizer using a hybrid

HMM/MLP system. volume 4. Belgium, 1999.

71. Jin, W., X. Liu, M. S. Scordilis, and L. Han (2010). Speech enhancement using

harmonic emphasis and adaptive comb filtering. IEEE Trans. on Audio Speech and

Language Processing, 18(2), 356–368.

72. Jin, Z. and D. Wang (2010). A multipitch tracking algorithm for noisy and reverberant

speech. Proceedings of IEEE Int. Conf. Acoust., Speech, and Signal Processing, 4218

–4221.

114

Page 130: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

73. Jin, Z. and D. Wang (2011). HMM-based multipitch tracking for noisy and reverberant

speech. IEEE Trans. on Audio,Speech, Lang. Process, 19(5), 1091–1102.

74. Jo, S. and C. D. Yoo (2010). Melody extraction from polyphonic audio based on par-

ticle filter. Proceedings of 11 th International Society for Music Information Retrieval

Conference (ISMIR), 357–362.

75. Jones, D. and T. Parks (1990). A high-resolution data-adaptive time-frequency repre-

sentation. IEEE Trans. Acoust., Speech, Lang. Process, 38(12), 2127–2135.

76. Joo, S., S. Jo, and C. D. Yoo (2010). Melody extraction from polyphonic audio signal

MIREX-2010. Extended Abstarct, MIREX 2010, Audio Melody Extraction Task.

77. Jordan, M. (2004). Graphical models. Statistical Science, 19, 140–155.

78. Jot, J.-M., L. Cerveau, and O. Warusfel (1997). Analysis and synthesis of room

reverberation based on a statistical time-frequency model. Proceedings of 103rd AES

Convention.

79. Kemp, T. and Waibel, A. (1999). Unsupervised training of a speech recognizer:

Recent experiments. Proceedings of Eurospeech, 2725–2728.

80. Klapuri, A. (2003). Multiple fundamental frequency estimation based on harmonicity

and spectral smoothness. IEEE Transaction on Speech and Audio Processing, 11(6),

804–816.

81. Klapuri, A. (2008). Multipitch analysis of polyphonic music and speech signals using

an auditory model. IEEE Trans. Audio Speech and Language Processing, 16(2), 255–

266.

82. Koduri, G. K., J. Serra, and X. Serra (2012). Characterization of intonation in carnatic

music by parametrizing pitch histograms. Proceedings of 13th International Society for

Music Information Retrieval Conference (ISMIR 2012), 199–204.

83. Kum, S., C. Oh, and J. Nam (2016). Melody extraction on vocal segments using multi-

column deep neural networks. Proceedings of 17th International Society for Music

Information Retrieval Conference (ISMIR), 819–825.

84. Kumar, J. C., R. Janakiraman, and H. A. Murthy, KL divergence based feature

switching in the linguistic search space for automatic speech recognition. In Proceed-

ings of National Conference on Communication(NCC). 2010.

85. Kumar, J. C. and H. A. Murthy (2009). Entropy based measures for incorporating

feature stream diversity in the linguistic search space for syllable based automatic an-

notated recognizer. Proceedings of National Conference on Communication, 286–289.

86. Kumar, P. (2015). High Resolution Property of Group Delay and its Application to

Musical Onset Detection on Carnatic Percussion Instruments. M.Tech dissertation,

Indian Institute of Technology, Department of Electrical Engg., Madras, India.

87. Kumar, P. A. M., J. Sebastian, and H. A. Murthy (2015). Musical onset detection on

carnatic percussion instruments. Proceedings of National Conference on Communica-

tion, 1–6.

115

Page 131: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

88. Kumar, P. V. A., L. Balakrishna, C. Prakash, and S. V. Gangashetty (2011). Bessel

features for estimating number of speakers from multispeaker speech signals. Proceed-

ings of International Conference on Systems, Signals and Image Processing (IWSSIP),

1–4.

89. Laskowski, K. and Q. Jinn (2010). Modeling prosody for speaker recognition – why

estimating pitch may be a red herring. Proceedings of Odyssey 2010.

90. Lewis, M. A. and R. P. Ramachandran (2001). Cochannel speaker count labelling

based on the use of cepstral and pitch prediction derived features. Pattern Recognition,

34, 499–507.

91. Li, M., C. Cao, D. Wang, P. Lu, Q. Fu, and Y. Yan (2008). Cochannel speech sepa-

ration using multi-pitch estimation and model based voiced sequential grouping. Pro-

ceedings of Int. Conf. Spoken Language Processing.

92. Li, T., M. Ogihara, and Q. Li (2003). A comparative study on content-based music

genre classification. Proceedings of International Conference on Research and Devel-

opment in Informaion Retrieval, 282–289.

93. Li, Y. and D. Wang (2007). Separation of singing voice from music accompaniment for

monaural recordings. IEEE Transactions on Audio, Speech, and Language Processing,

15(4), 1475–1487.

94. Machado Rocha, B. M. (2011). Genre classification based on predominant melodic

pitch contours. Master Thesis ,Uniersity of Pompiu Fabra.

95. Madhumurthy, K. V. and B. Yegnanarayana (1989). Effectiveness of representation

of signals through group delay functions. Signal Processing, 17, 141–150.

96. Maher, R. and J. Beauchamp (1994). Fundamental frequency estimation of musical

signals using two-way mismatch procedure. J. Acoust. Soc. Amer., 19, 2254–2263.

97. Mauch, M. and S. Dixon (2014). PYIN: A fundamental frequency estimator using

probabilistic threshold distributions. Proceedings of IEEE Int. Conf. Acoust., Speech,

and Signal Processing, 659–663.

98. Meddis, R. and M. Hewitt (1991). Virtual pitch and phase sensitivity of a computer

model of the auditory periphery. I: Pitch identification. J. Acoust. Soc. Am., 4(2), 2866–

2882.

99. Ming, G., W. Jinfang, L. Dongxin, and L. Chang (2013). Depression detection using

the derivative features of group delay and delta phase spectrum. Proceedings of IEEE

International Conference on Advance Computing(IACC), 1275–1278.

100. Miwa, T., Y. Tadokoro, and T. Saito (2000). Musical pitch estimation and discrimi-

nation of musical instruments using comb filters for transcription. Proceedings of 42nd

Midwest Symposium on Circuits and Systems, 1, 105–108.

101. Moorer, J. A. (1974). The optimum comb method of pitch period analysis of con-

tinuous digitized speech. IEEE Trans. Acoust. Speech Signal Processing, ASSP-22,

330–338.

116

Page 132: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

102. Mowlaee, P., M. G. Christensen, Z. Tan, and S. H. Jensen (2010). A MAP crite-

rion for detecting the number of speakers at frame level in model-based single channel

speech separation. Rec. Asilomar Conf. Signals, Syst.,Comput., 538–541.

103. Munoz-Exposito, J. E., S. Garcia-Galan, N. Ruiz-Reyes, and P. Vera-Candeas

(2007). Adaptive network-based fuzzy inference system vs. other classification algo-

rithms for warped LPC-based speech/music discrimination,. Engineering Applications

of Artificial Intelligence, 20(6), 783–793.

104. Murthy, H. A. (1991). Algorithms for Processing Fourier Transform Phase of Signals.

Ph.D dissertation, Indian Institute of Technology, Department of Computer Science and

Engg., Madras, India.

105. Murthy, H. A. (1994). Pitch extraction from root cepstrum. Proceedings of the 3rd

International Conference on Spoken Language Processing, 1055–1058.

106. Murthy, H. A. and G. V. R. Rao (2003). The modified group delay function and its

application to phoneme recognition. Proceedings of IEEE Int. Conf. Acoust., Speech,

and Signal Processing, 1, 68–71.

107. Murthy, H. A. and B. Yegnanarayana (1991.). Formant estimation from group delay

functions. Speech Communication, 209–221.

108. Murthy, H. A. and B. Yegnanarayana (1991a). Formant extraction from minimum

phase group delay function. Speech Communication, 10, 209–221.

109. Murthy, H. A. and B. Yegnanarayana (1991b). Speech processing using group delay

functions. Signal Processing, 22, 259–267.

110. Murthy, H. A. and B. Yegnanarayana (2011). Group delay functions and its applica-

tion to speech processing. Sadhana, 36(5), 745–782.

111. Murty, K. S. R. and B. Yegnanarayana (2008). Epoch extraction from speech signals.

IEEE International Transactions on Audio,Speech and Language Processing, 16(8),

1602–1613.

112. Nagarajan, T. and H. A. Murthy (2006). Language identification using acoustic log-

likelihood of syllable-like units. Speech Communication, 48, pp.913–926.

113. Nagarajan, T., H. A. Murthy, and R. M. Hegde (2003). Segmentation of speech into

syllable-like units. Proceedings of EUROSPEECH, 2893–2896.

114. Nishimoto, T., H. Kameoka, and S. Sagayama (2007). A multipitch analyzer based on

harmonic temporal structured clustering. IEEE Trans. on Audio Speech and Language

Processing, 15(3), 982–994.

115. N.Scaringella, G.Zoia, and MIyenK (2006). Automatic genre classification of music

content. IEEE Signal Processing Magazine, 23(2), 133–141.

116. Ono, N., K. Miyamoto, J. L. Roux, H. Kameoka, and S. Sagayama (2008). Sep-

aration of a monaural audio signal into harmonic/percussive components by comple-

mentary diffusion on spectrogram. Proceedings of the European Signal Processing

Conference, 25–29.

117

Page 133: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

117. Oppenheim, A. V. and R. W. Schafer, Discrete Time Signal Processing. Prentice Hall,

Inc, New Jersey, 1990.

118. Padmanabahan, R. (2012). Studies on Voice Activity Discussion and Feature Diversity

for speaker recognition. Ph.D dissertation, Indian Institute of Technology, Department

of Computer Science and Engg., Madras, India.

119. Padmanabhan, R. and H. A. Murthy (2010). Acoustic feature diversity and speaker

verification. Proceedings of Int. Conf. Spoken Language Processing, 2010–2113.

120. Painter, T. and A. Spanias (2000). Perceptual coding of digital audio. Proceeding of

IEEE, 88(4), 451–515.

121. Paiva, R., T. Mendes, and A. Cardoso (2006). Melody detection in polyphonic music

signals. Comput. Music J., 30(4), 80–98.

122. Paliwal, K. K. (2003). Usefulness of phase in speech processing. Proceedings of IPSJ

Spoken Language Processing Workshop, 1–6.

123. Pardo, B., Z. Duan, and C. Zhang (2010). Multiple fundamental frequency estimation

by modeling spectral peaks and non-peak regions. IEEE Trans. Audio Speech Language

Processing, 18(8), 2121–2131.

124. Petriker, G., M. Wohlmayr, S. Petrik, and F. Pernkopf (2011). A pitch tracking cor-

pus with evaluation on multipitch tracking scenario. Proceedings of Int. Conf. Spoken

Language Processing, 1509–1512.

125. Pfeiffer, F. S., S. and W. Effelsberg (1996). Automatic audio content analysis. Pro-

ceedings of ACM International Conference on Multimedia,, 21–30.

126. Pikrakis, A., F. Gomez, S. Oramas, J. Bnez, J. Mora, F. Escobar, E. Gomez, and

J. Salamon (2012). Tracking melodic patterns in flamenco singing by analyzing poly-

phonic music recordings. Proceedings of International Society for Music Information

Retrieval Conference (ISMIR 2012), 421–426.

127. P.Klapuri, A. (2001). Multipitch estimation and sound separation by the spectral

smoothness principle. Proceedings of IEEE International Conference on Acoustics,

Speech and Signal Processing (ICASSP), 2001, 5, 3381–3384.

128. Plante, F., G.Meyer, and W.A.Ainsworth (1995). A pitch extraction reference

database. Proceedings of Fourth European Conference on Speech Communication and

Technology, 1–5.

129. Plinge, A. and G. A. Fin (2014). Multi-speaker tracking using multiple distributed

microphone arrays. Proceedings of IEEE Int. Conf. Acoust., Speech, and Signal Pro-

cessing, 614 – 618.

130. Poliner, G. and D. Ellis (2005). A classification approach to melody transcription.

Proceedings of International Conference on Music Information Retrieval, 161–166.

131. Poliner, G., D. Ellis, A. Ehmann, E. Gomez, S. Streich, and B. Ong (2007). Melody

transcription from music audio:approaches and evaluation. Proceedings of the IEEE

Int. Conf. on Audio, Speech and Language Processing, 15(4), 1247–1256.

118

Page 134: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

132. Porat, B. and A. Nehorai (1986). Adaptive comb filtering for harmonic signal en-

hancement. IEEE Trans. on Acoust., Speech, Signal Processing, 34(5), 1124–1138.

133. Pradhan, A., S. A. Shanmugam, A. Prakash, V. Kamakoti, and H. A. Murthy

(2013). Syllable based statistical text to speech system. Proceedings of European

Signal Processing Conference, 1–5.

134. Prasad, V. K., T. Nagarajan, and H. A. Murthy (2004). Automatic segmentation of

continuous speech using minimum phase group delay functions. Speech Communica-

tion, 42, pp.429–446.

135. Qin, J., H. Lin, and X. Liu (2011). Query by humming systems using melody matching

model based on the genetic algorithm. Journal of Software, 6(12), 2416–2420.

136. Quinlan, A., M. Kawamoto, Y. Matsusaka, H. Asoh, and F. Asano (2009). Tracking

intermittently speaking multiple speakers using a particle filter. EURASIP Journal on

Audio, Speech, and Music Processing, 1–11.

137. Quinlan, J. R., C4.5 Programs for Machine Learning. 1993.

138. Rabiner, L., M.J.Cheng, A.E.Rosenberg, and A.McGonegal (1976). A comparative

study of several pitch estimation algorithms. Proceedings of IEEE Int. Conf. Acoust.,

Speech, and Signal Processing, 23, 552–557.

139. Radfar, M., R. Dansereau, W. Chan, and W. Wong (2011). MPtracker: A new multi-

pitch detection and separation algorithm for mixed speech signals. Proceedings of IEEE

Int. Conf. Acoust., Speech, and Signal Processing, 4468 –4471.

140. Rajan, P., T. Kinnunen, C. Hanilci, J. Pohjalainen, and P. Alku (2013). Using group

delay functions from all-pole models for speaker recognition. Proceedings of Int. Conf.

Spoken Language Processing, 2489–2493.

141. Rajan, R. and H. A. Murthy (2013). Group delay based melody monopitch extraction

from music. Proceedings of the IEEE International Conference on Audio, Speech and

Signal Processing, 186–190.

142. Ramakrishnan, S., V. Rao, and P. Rao (2008). Singing voice detection in north Indian

classical music. Proceedings of the National Conference on Communications, 1–5.

143. Rao, K. S., S. R. M. Prasanna, and B. Yegnanarayana (2007). Determination of

instants of significant excitation in speech using Hilbert envelope and group delay func-

tion. IEEE Signal Processing Letters, 14(10), 762–765.

144. Rao, P. and S. Shandilya (2004). On the detection of melodic pitch in a percussive

background. J. Audio Eng Soc, 52(4), 378–391.

145. Rao, V., P. Gaddipati, and P. Rao (2012). Signal-driven window length adaptation

for sinusoid detection in polyphonic music. IEEE Transactions on Audio Speech and

Language Processing, 20(1), 342–348.

146. Rao, V. and P. Rao (2010). Vocal melody extraction in the presence of pitched accom-

paniment in polyphonic music. IEEE Transactions on Audio, Speech, and Language

Processing, 18(8), 2145–2154.

119

Page 135: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

147. Rasipuram, R., R. M. Hegde, and H. A. Murthy (2008a). Incorporating acoustic

feature diversity into the linguistic search space for syllable based speech recognition.

Proceedings of European Conference on Signal Processing, 1–5.

148. Rasipuram, R., R. M. Hegde, and H. A. Murthy (2008b). Significance of group

delay based acoustic features in the linguistic feature space for syllable recognition.

Proceedings of Int. Conf. Spoken Language Processing, 1–5.

149. Richard, G., J. L. Durrieu, and B. David (2008). Singer melody extraction in poly-

phonic signals using source separation methods. Proceedings of the IEEE International

Conference on Acoustics, Speech and Signal Processing, 169–172.

150. R.M.Hegde, H. Murthy, and G. Rao (2004). Application of the modified group delay

function to speaker identification and discrimination. Proceedings of IEEE Int. Conf.

Acoust., Speech, and Signal Processing, 517–520.

151. Ryynanen, M. and A. Klapuri (2008a). Automatic transcription of melody ,base

line,and chords in polyphonic music. Computer Music Journal, 32(3), 72–86.

152. Ryynanen, M. and A. Klapuri (2008b). Query by humming of midi and audio using

locality sensitive hashing. Proceedings of IEEE Int. Conf. Acoust., Speech, and Signal

Processing, 2249–2252.

153. Ryynanen, M., T. Virtanen, J. Paulus, and A. Klapuri (2008). Accompaniment sep-

aration and karaoke application based on automatic melody transcription. Proceedings

of IEEE International Conference on Multimedia and Expo, 1417–1420.

154. Salamon, J. and E. Gomez (2012). Melody extraction from polyphonic music signals

using pitch contours characteristics. IEEE Transactions on Audio Speech and Language

Processing, 20(6), 1759–1770.

155. Salamon, J., E. Gomez, D. Ellis, and G. Richard (2015). Melody extraction from

music recordings. IEEE Signal Processing Society.

156. Salamon, J., E. Gomez, D. P. Ellis, and G. Richard (2014a). Melody extraction from

polyphonic music signals. IEEE Signal Processing Magazine, 118–134.

157. Salamon, J., E. Gomez, D. P. W. Ellis, and G. Richard (2014b). Melody extraction

from polyphonic music signals: Approaches, applications and challenges. IEEE Signal

Processing Magazine, 31(2), 118–134.

158. Salamon, J., S. Gulati, and X. Serra (2012a). A multipitch approach to tonic tdenti-

fication in indian classical music. Proceedings of The International Society for Music

Information Retrieval (ISMIR) Conference, 499–504.

159. Salamon, J., B. Rocha, and E. Gomez (2012b). Musical genre classification using

melody features extracted from polyphonic music signals. Proceedings of IEEE Int.

Conf. Acoust., Speech, and Signal Processing, 81 – 84.

160. Salamon, J., J. Serra, and E. Gomez (2013). Tonal representations for music retrieval:

From version identification to query-by-humming. International Journal of Multimedia

Information Retrieval, special issue on Hybrid Music Information Retrieval, 2(1), 45–

58.

120

Page 136: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

161. Salamon, J. and J. Urbano (2012). Current challenges in the evaluation of predom-

inant melody extraction algorithms. Proceedings of International Society for Music

Information Retrieval Conference (ISMIR), 289–294.

162. Samouelian, A. (1997). Frame level phoneme classification using inductive learning.

Computer Speech and Language, 11, 161–186.

163. Sarada, G. L., A. Lakshmi, H. A. Murthy, and T. Nagarajan (2009). Automatic tran-

scription of continuous speech into syllable-like units for indian languages. Sadhana,

34, Part 2, pp.221–233.

164. Saunders, J. (1996). Real-time discrimination of broadcast speech/ music. Proceed-

ings of the IEEE International Conference on Acoustics, Speech and Signal Processing

(ICASSP), 2, 993–996.

165. Sayoud, H. and S. Ouamour (2010). Proposal of a new confidence parameter esti-

mating the number of speakers-an experimental investigation. Journal of Information

Hiding and Multimedia Signal Processing, 1(2), 101–109.

166. Scheirer, E. (2000). Music listening systems. Ph.D thesis., MIT .

167. Scheirer, E. and M. Slaney (1997). Construction and evaluation of a robust multifea-

ture speech/music discriminator. Proceedings of the IEEE International Conference on

Acoustics, Speech and Signal Processing (ICASSP),, 2, 1331–1334.

168. Sebastian, J., M. G. Kumar, Y. S. Sreekar, V. R. Rajeev, H. A. Murthy, and S. Mri-

ganka (2017). GDSpike: An accurate spike estimation algorithm from noisy calcium

fluorescence signal. Proceedings of IEEE Int. Conf. Acoust., Speech, and Signal Pro-

cessing.

169. Sebastian, J., P. A. M. Kumar, and H. A. Murthy (2015a). An analysis of the high

resolution property of group delay function with applications to audio signal processing.

Speech Comunication, 81, 42–53.

170. Sebastian, J., P. A. M. Kumar, and H. A. Murthy (2015b). Pitch estimation from

speech using grating compression transform on modified group-delay-gram. Proceed-

ings of National Conference on Communication(NCC), 1–6.

171. Sethu, V., E. Ambikairajah, and J. Epps (2007). Group delay features for emotion

detection. Proceedings of 8th Conference of the International Speech Communication

Association, 2273–2276.

172. Shan, M. K., F. F. Kuo, and M. F. Chen (2002). Music style mining and classification

by melody. Proceedings of IEEE International Conference on Multimedia and Expo,

97– 100.

173. Shanmugam, S. A. and H. A. Murthy (2014). Group delay based phone segmentation

for HTS. Proceedings of National Conference on Communication, 1–6.

174. Shao, Y., S. Srinivasan, Z. Jin, and D. Wang (2010). A computational auditory scene

analysis system for speech segregation and robust speech recognition. Computer Speech

and Language, 24(1), 77–93.

121

Page 137: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

175. Shao, Y. and D. Wang (2003). Co-channel speaker identification using speech extrac-

tion based on multipitch tracking. Proceedings of IEEE Int. Conf. Acoust., Speech, and

Signal Processing, 2, 205–208.

176. Shields, J. (1970). Separation of added speech signls by digital comb filteing. Ph.D

Thesis, MIT .

177. Simsekli, U. (2010). Automatic music genre classification using bass lines. Proceed-

ings of International Conference on Pattern Recognition, 4137 – 4140.

178. Song, J., S. Y. Bae, and K. Yoon (2002). Mid-level music melody representation

of polyphonic audio for query-by-humming system. Proceedings of 3rd International

Conference on Music Information Retrieval, 133–139.

179. Su, Y. (2012). Group delay variations in microwave filters and equalization methodolo-

gies. Master’s Thesis in Microtechnology and Nanoscience.

180. Suni, A., T. Raitio, D. Gowda, R. Karhila, M. Gibson, and O. Watts. (2014). The

simple4all entry to the blizzard challenge 2014. Proceedings of the Blizzard Challenge

2014 Workshop,September 2014, Singapore.

181. Swamy, R. K., K. S. R. Murty, and B. Yegnanarayana (2007). Determining number

of speakers from multispeaker speech signals using excitation source information. IEEE

Signal Processing Lett, 14(7), 481–484.

182. Tachibana, H., T. Ono, N. Ono, and S. Sagayama (2010). Melody line estimation

in homophonic music audio signals based on temporal-variability of melodic source.

Proceedings of IEEE International Conference Acoustic, Speech, Signal Processing,

425–428.

183. Tan, L. and A. Alwan (2011). Noise-robust f0 estimation using SNR-weighted sum-

mary correlograms from multi-band comb filters. Proceedings of IEEE Int. Conf.

Acoust., Speech, and Signal Processing, 4464–4467.

184. Tao Li, A., M. Ogihara, and Q. Li (2003). A comparative study on content-based

music genre classification. Proceedings of the International ACM SIGIR Conference

on Research and Development in Information Retrieval, 1759–1770.

185. Theodorou, T., I. Mporas, and N. Fakotakis (2012). Automatic sound classification

of radio broadcast news. International Journal of Signal Processing, Image Processing

and Pattern Recognition, 5(1), 37–48.

186. Thiruvaran, T., E. Ambikairajah, and J. Epps (2007). Group delay features for

speaker recognition. Proceedings of 6th International Conference on Information, Com-

munications and Signal Processing, 1–5.

187. Thomas, S., S. Ganapathy, and H. Hermansky (2008). Recognition of reverberant

speech using frequency domain linear prediction. IEEE Signal Processing Letters, 15,

681–684.

188. Thornburg, H. (2005). Detection and modeling of transient audio signals with prior

information. Ph.D Thesis, Standford University.

189. Tolonen, T. and M. Karjalainen (2000). A computationally efficient multipitch anal-

ysis model. IEEE Transaction Speech and Audio Processing, 8(6), 708–716.

122

Page 138: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

190. Tralie, C. J. and P. Bendich (2015). Cover song identification with timbral shape

sequences. arXiv preprint arXiv:1507.05143 , 2015.

191. Tzanetakis, G. and P. Cook (2002). Musical genre classification of audio signals.

IEEE Transactions on Speech and Audio Processing, 10(5), 293–302.

192. Tzanetakis, G., G. Essl, and P. Cook (2001). Automatic musical genre classification

of audio signals. Proceedings of International Society for Music Information Retrieval

Conference(ISMIR), 205–210.

193. Tzu, C. Y., M. J. Wu, J. S. R. Jang, W. L. Chang, and I. B. Liao (2012). A hybrid

approach to singing pitch extraction based on trend estimation and Hidden Markov

Models. IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP) Kyoto,

Japan, 457–460.

194. UlaBagci and E. Erzin (2007). Musical genres using inter-genre similarity. IEEE

Signal Processing Letters, 14(8), 521–524.

195. Veldhuis, R. (2000). Consistant pitch marking. Proc. Sixth International Conf.on

Spoken Language Processing, 3, 207–210.

196. Vijayan, K., V. Kumar, and K. S. R. Murty (2014). Feature extraction from ana-

lytic phase of speech signals for speaker verification. Proceedings of Fifteenth Annual

Conference of the International Speech Communication Association, 1658–1662.

197. Vincent, E., N. Bertin, and R. Badeau (2010). Adaptive harmonic spectral decompo-

sition for multiple pitch estimation. IEEE Transaction on Speech and Audio Processing,

18(3), 528–537.

198. Vinodh, M., A. Bellur, K. B. Narayan, M. D. Thakare, A. Susan, N. Suthakar, and

H. A. Murthy (2010). Using polysyllabic units for text to speech synthesis in indian

languages. Proceedings of National Conference on Communication, 1–5.

199. Vishnubhotla, S. and C. Espy-Wilson (2008). An algorithm for multi-pitch tracking

in co-channel speech. Proceedings of Int. Conf. Spoken Language Processing.

200. Walmsley, P. J., S. J. Godsill, and P. J. W. Rayner (1999). Polyphonic pitch tracking

using joint bayesian estimation of multiple frame parameters. Proceedings of 1999

IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 119–

122.

201. Wang, D. and G. J. Brown (2006). Computational Auditory Scene Analysis:Principles,

Algorithms, and Applications. John Wiley and sons, (ISBN:0471741094).

202. Wang, D. and P. C. Loizou (2006). Single channel speech source separation based on

binary spectrum mask estimation. University of Texas at Dallas-Poster.

203. Wang, T. T. and T. F. Quatieri (2009). 2-D processing of speech for multi-pitch anal-

ysis. Proceedings of Int. Conf. Spoken Language Processing.

204. Wiering, F. (2006). Can humans benefit from music information retrieval? Proceedings

of the 4th international conference on Adaptive multimedia retrieval, 82–94.

123

Page 139: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

205. Wohlmayr, M., M. Stark, and F. Pernkopf (2011). A probabilistic interaction model

for multipitch tracking with factorial Hidden Markov Models. IEEE Trans. Audio,

Speech, Lang. Process., 19(4), 799–810.

206. W.S-URL (2012). http://www.speech.kth.se/wavesurfer/. URL.

207. Wu, M. and D. Wang (2003). A multipitch tracking algorithm for noisy speech. IEEE

Trans. Speech Audio Processing, 11(3), 229–241.

208. Wulfing, J. and M. Riedmille (2012). Unsupervised learning of local features for music

classifications. Proceedings of International Society for Music Information Retrieval

conference (ISMIR), 139–144.

209. Xiong, X. (2009). Robust Speech Features and Acoustic Models For Speech Recogni-

tion,. Ph.D dissertation, School of Computer Engineering,Nanyag Technological Uni-

versity.

210. Xu, C., N. C. Maddage, X. Shao, F. Cao, and Q. Tian (2003). Musical genre classi-

fication using support vector machines. Proceedings of IEEE International Conference

on Acoustics, Speech, and Signal Processing, 5, 429–432.

211. Yegnanarayana, B., M. A. Joseph, V. G. Suryakanth, and N. Dhananjaya (2011).

Decomposition of speech signals for analysis of aperiodic components of excitation.

Proceedings of IEEE Int. Conf. Acoust., Speech, and Signal Processing, 5396–5399.

212. Yegnanarayana, B. and H. A. Murthy (1992). Significance of group delay functions

in spectrum estimation. IEEE Trans. Signal Processing, 40(9), 2281–2289.

213. Yegnanarayana, B., H. A. Murthy, and V. R. Ramachandran (1991). Processing of

noisy speech using modified group delay functions. Proceedings of the IEEE Int. Conf.

on Audio, Speech and Signal Processing, 945–948.

214. Yegnanarayana, B., D. K. Saikia, and T. R. Krishan (1984). Significance of group

delay functions in signal reconstruction from spectral magnitude or phase. IEEE Trans.

Acoustics Speech and Signal Processing, 32(3), 610–623.

215. Yeh, C. (2008). A multiple fundamental frequency estimation of polyphonic record-

ings. Ph.D Thesis, University of Paris, IRCAM.

216. Yip, P. and K. R. Rao, Discrete Cosine Transform: Algorithms, Advantages and Appli-

catons. Academic Press, USA, 1997.

217. Yoon, J. Y., C. J. Song, S. P. Lee, and H. Park (2011). Extracting predominant melody

of polyphonic music based on harmonic structure. Extended abstract, MIREX-2011,

Audio melody extraction task, 1–2.

218. Zatorre, R. J. and S. R. Baum (2012). Musical melody and speech intonation: Singing

a different tune. Public Libarary of Science, 10(7), 1–6.

219. Zhang, J. X., M. Christensen, H. Jensen, and M. Moonen (2012). Joint DOA and

multi-pitch estimation based on subspace techniques. EURASIP Journal on Advances

in Signal Processing, 1–5.

124

Page 140: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

220. Zhang, X., W. Liu, P. Li, , and B. Xu (2008). Multipitch detection based on weighted

summary correlogram. Proceedings of International Symposium on Chinese Spoken

Language Processing(ISCSLP), 1–4.

221. Zhu, D. and K. K. Paliwal (2004). Product of power spectrum and group delay function

for speech recognition. Proceedings of IEEE Int. Conf. Acoust., Speech, and Signal

Processing, 572–575.

125

Page 141: Estimation of Pitch in Speech and Music Signals Using Modified …€¦ · THESIS CERTIFICATE This is to certify that the thesis titled, “Estimation of Pitch in Speech and Music

LIST OF PUBLICATIONS

1. Rajeev Rajan, Manaswi Misra and Hema A. Murthy ,“Melody extraction from

music using modified group delay functions”, International Journal of Speech

Technology, Volume 20, Issue 1, pp 185–204, March 2017.

2. Rajeev Rajan and Hema A. Murthy, “Two-pitch tracking in co-channel speech

using modified group delay functions”, Speech Communication, Volume 89,

pp 37–46, May 2017.

3. Rajeev Rajan and Hema A. Murthy, “Music genre classification using roup

delay and melodic Features ”, In Procee. of Communications (NCC) , National

Conference on, pp 1–5, Indian Institute of Technology, Chennai, March 2017.

4. Rajeev Rajan and Hema A. Murthy “Group delay based melody monopitch

extraction from music”, In Procee. Acoustics, Speech and Signal Processing

(ICASSP) , IEEE International Conference on, pp 186–190, Vancouver, Canada,

May 2013.

5. Rajeev Rajan and Hema A. Murthy ,“Melodic pitch extraction from music sig-

nals using modified group delay functions”, In Procee. of Communications

(NCC), National Conference on, pp 1–5, Indian Institute of Technology, Delhi,

February 2013.

126