Jenei et al., 2021 - Google Patents
Separation of several illnesses using correlation structures with convolutional neural networksJenei et al., 2021
View PDF- Document ID
- 6535839999290460136
- Author
- Jenei A
- Kiss G
- Tulics M
- Sztahó D
- Publication year
- Publication venue
- Acta Polytechnica Hungarica
External Links
Snippet
There is already a lot of research in the literature on the binary separation of healthy people and people with some illnesses that affects speech. However, there are only a few examinations where more illnesses are recognized together. The examination of the latter is …
- 201000010099 disease 0 title abstract description 50
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/66—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification
- G10L17/26—Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification
- G10L17/04—Training, enrolment or model building
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/065—Adaptation
- G10L15/07—Adaptation to the speaker
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/90—Pitch determination of speech signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/93—Discriminating between voiced and unvoiced parts of speech signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wroge et al. | Parkinson’s disease diagnosis using machine learning and voice | |
Mekyska et al. | Robust and complex approach of pathological speech signal analysis | |
Amato et al. | An algorithm for Parkinson’s disease speech classification based on isolated words analysis | |
Travieso et al. | Detection of different voice diseases based on the nonlinear characterization of speech signals | |
Jenei et al. | Separation of several illnesses using correlation structures with convolutional neural networks | |
Upadhya et al. | Thomson Multitaper MFCC and PLP voice features for early detection of Parkinson disease | |
Dahmani et al. | Vocal folds pathologies classification using Naïve Bayes Networks | |
Joshy et al. | Dysarthria severity classification using multi-head attention and multi-task learning | |
Aversano et al. | A machine learning approach for early detection of parkinson’s disease using acoustic traces | |
Miao et al. | Fusing features of speech for depression classification based on higher-order spectral analysis | |
Verde et al. | A lightweight machine learning approach to detect depression from speech analysis | |
Eskıdere et al. | Detection of Parkinson's disease from vocal features using random subspace classifier ensemble | |
Mittal et al. | Deep learning approach for voice pathology detection and classification | |
Majda-Zdancewicz et al. | Deep learning vs feature engineering in the assessment of voice signals for diagnosis in Parkinson’s disease | |
Kulkarni et al. | Child cry classification-an analysis of features and models | |
Costantini et al. | Machine learning based voice analysis in spasmodic dysphonia: an investigation of most relevant features from specific vocal tasks | |
Kodali et al. | Automatic classification of the severity level of Parkinson’s disease: A comparison of speaking tasks, features, and classifiers | |
Orozco-Arroyave et al. | Automatic detection of hypernasal speech of children with cleft lip and palate from spanish vowels and words using classical measures and nonlinear analysis | |
Zhu et al. | How generalizable and interpretable are speech-based covid-19 detection systems?: A comparative analysis and new system proposal | |
Sztahó et al. | Automatic discrimination of several types of speech pathologies | |
Jenei et al. | Possibilities of recognizing depression with convolutional networks applied in correlation structure | |
Romero et al. | Exploring transformer-based language recognition using phonotactic information | |
Patil et al. | Selection of classifiers for depression detection using acoustic features | |
Sztahó et al. | Automatic separation of various disease types by correlation structure of time shifted speech features | |
Gao et al. | Intelligent diagnosis approach for depression using vocal source features |