• Login
    View Item 
    •   Research Bank Home
    • Unitec Institute of Technology
    • Study Areas
    • Computing
    • Computing Conference Papers
    • View Item
    •   Research Bank Home
    • Unitec Institute of Technology
    • Study Areas
    • Computing
    • Computing Conference Papers
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Phonated speech reconstruction using twin mapping models

    Sharifzadeh, Hamid; HajiRassouliha, Amir; McLoughlin, I.V.; Ardekani, Iman; Allen, Jacqueline E.

    Thumbnail
    Share
    View fulltext online
    isspit-HRSH-Final.pdf (6.650Mb)
    Date
    2015-12
    Citation:
    Sharifzadeh, H. R., HajiRassouliha, A., McLoughlin, I. V., Ardekani, I. T., & Allen, J. E. (2015, December) Phonated Speech Reconstruction Using Twin Mapping Models. IEEE (Ed.), Proceedings of the 15th IEEE International Symposium on Signal Processing and Information Technology (pp.1-6)
    Permanent link to Research Bank record:
    https://hdl.handle.net/10652/3352
    Abstract
    Computational speech reconstruction algorithms have the ultimate aim of returning natural sounding speech to aphonic and dysphonic individuals. These algorithms can also be used by unimpaired speakers for communicating sensitive or private information. When the glottis loses function due to disease or surgery, aphonic and dysphonic patients retain the power of vocal tract modulation to some degree but they are unable to speak anything more than hoarse whispers without prosthetic aid. While whispering can be seen as a natural and secondary aspect of speech communications for most people, it becomes the primary mechanism of communications for those who have impaired voice production mechanisms, such as laryngectomees. In this paper, by considering the current limitations of speech reconstruction methods, a novel algorithm for converting whispers to normal speech is proposed and the efficiency of the algorithm is discussed. The proposed algorithm relies upon twin mapping models and makes use of artificially generated whispers (called whisperised speech) to regenerate natural phonated speech from whispers. Through a training-based approach, the mapping models exploit whisperised speech to overcome frame to frame time alignment problem in the speech reconstruction process.
    Keywords:
    speech reconstruction, impaired speech, aphonic patients, dysphonic patients, voice production, computational speech reconstruction algorithms
    ANZSRC Field of Research:
    200402 Computational Linguistics
    Copyright Holder:
    IEEE Communications Society

    Copyright Notice:
    All rights reserved
    Available Online at:
    http://ece.adu.ac.ae/ISSPIT2015/index.html
    Rights:
    This digital work is protected by copyright. It may be consulted by you, provided you comply with the provisions of the Act and the following conditions of use. These documents or images may be used for research or private study purposes. Whether they can be used for any other purpose depends upon the Copyright Notice above. You will recognise the author's and publishers rights and give due acknowledgement where appropriate.
    Metadata
    Show detailed record
    This item appears in
    • Computing Conference Papers [150]

    Te Pūkenga

    Research Bank is part of Te Pūkenga - New Zealand Institute of Skills and Technology

    • About Te Pūkenga
    • Privacy Notice

    Copyright ©2022 Te Pūkenga

    Usage

    Downloads, last 12 months
    44
     
     

    Usage Statistics

    For this itemFor the Research Bank

    Share

    About

    About Research BankContact us

    Help for authors  

    How to add research

    Register for updates  

    LoginRegister

    Browse Research Bank  

    EverywhereInstitutionsStudy AreaAuthorDateSubjectTitleType of researchSupervisorCollaboratorThis CollectionStudy AreaAuthorDateSubjectTitleType of researchSupervisorCollaborator

    Te Pūkenga

    Research Bank is part of Te Pūkenga - New Zealand Institute of Skills and Technology

    • About Te Pūkenga
    • Privacy Notice

    Copyright ©2022 Te Pūkenga