Expressive speech can be synthesized using acoustic feature modeling by mapping the spectral and fundamental frequency parameters between neutral speech and target emotions based on context. Speaker and text-independent emotion conversion are challenging modeling problems in this paradigm. In this paper, spectral mapping using an i-vector-based framework of fixed dimensions is proposed for the speaker-independent emotion conversion, considering the entire problem in the utterance domain, rather than the existing approaches using frame-level processing. The high dimensionality of i-vectors and reduced utterances for i-vector training necessitate the use of Probabilistic Linear Discriminant Analysis (PLDA) to derive the emotion dependent latent vector. The i-vector setup does not require parallel data or alignment procedures at any stage of training. F 0 training is conducted on a multilayer feed-forward neural network using limited aligned seed parallel data. The framework is tested on three different languages (datasets) viz. German (EmoDB), Telugu (IITKGP), and English (SAVEE). The proposed approach delivered superior performance compared to the baseline under both clean and noisy data conditions considered for analysis. Under clean data conditions, the proposed model was found to perform better than the baseline with a Mel Cepstral Distortion as low as 3.8 (fear), an F 0-RMSE of 26.31 (happiness), and a Perceptual Evaluation of Speech Quality (PESQ) of 3.64 (anger) across datasets. Subjective testing provided a maximum CMOS of 4.10 (anger), 4.44 (fear), and 3.43 (happiness). INDEX TERMS CV-GMM, speech emotion, feed-forward ANN, i-vector, MFCC, PLDA. I. INTRODUCTION Emotions form a prominent para-linguistic element of human communication, consisting of speech, facial expressions, gestures, body language, etc. Among these, speech is the most readily accessible information source containing datapoints such as the message conveyed, speaker identity, gender, emotion, and speaker state of health. Emotions animate our speech and is essential for effective dialogue delivery in human-machine interaction and socio-cultural relationships. Furthermore, expressive speech synthesis finds applications The associate editor coordinating the review of this manuscript and approving it for publication was Kathiravan Srinivasan. in story-telling, speaking aids for the disabled [1]-[6], video games and speech-to-speech translators [7] to name a few. An expression synthesis system is normally added as a post-processing stage in text-to-speech synthesis (TTS) systems. There is often a need for a TTS synthesizer tested across multiple languages. This case is particularly relevant in multilingual countries such as India, where 22 official languages [8] exist along-with several other unofficial languages. Effective training of prosodic and spectral parameters from multiple languages is particularly useful in designing affective speech-to-speech translators in lowresource languages. Human-like dialogue delivery often encounters spontaneou...