Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleMarch 2017
Rotation invariance through structured sparsity for robust hyperspectral image classification
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 6205–6209https://doi.org/10.1109/ICASSP.2017.7953349Sparse representation based classification has gained popularity with geospatial image analysis in general and hyperspectral image analysis in particular. A central idea with such classification approaches is that a test pixel (spectral reflectance vector)...
- research-articleMarch 2017
A novel ensemble classifier of hyperspectral and LiDAR data using morphological features
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 6185–6189https://doi.org/10.1109/ICASSP.2017.7953345Due to the benefits and limitation of different remote sensing sensors, fusion of the features from multiple sensors, such as hyperspectral and light detection and ranging (LiDAR) is an effective method for land cover mapping. In this paper, we propose a ...
- research-articleMarch 2017
End-to-end joint learning of natural language understanding and dialogue manager
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 5690–5694https://doi.org/10.1109/ICASSP.2017.7953246Natural language understanding and dialogue policy learning are both essential in conversational systems that predict the next system actions in response to a current user utterance. Conventional approaches aggregate separate models of natural language ...
- research-articleMarch 2017
e-vectors: JFA and i-vectors revisited
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 5435–5439https://doi.org/10.1109/ICASSP.2017.7953195Systems based on i-vectors represent the current state-of-the-art in text-independent speaker recognition. In this work we introduce a new compact representation of a speech segment, similar to the speaker factors of Joint Factor Analysis (JFA) and to i-...
- research-articleMarch 2017
TristouNet: Triplet loss for speaker turn embedding
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 5430–5434https://doi.org/10.1109/ICASSP.2017.7953194TristouNet is a neural network architecture based on Long Short-Term Memory recurrent networks, meant to project speech sequences into a fixed-dimensional euclidean space. Thanks to the triplet loss paradigm used for training, the resulting sequence ...
-
- research-articleMarch 2017
Incremental adaptation using active learning for acoustic emotion recognition
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 5160–5164https://doi.org/10.1109/ICASSP.2017.7953140The performance of speech emotion classifiers greatly degrade when the training conditions do not match the testing conditions. This problem is observed in cross-corpora evaluations, even when the corpora are similar. The lack of generalization is ...
- research-articleMarch 2017
A PLLR and multi-stage Staircase Regression framework for speech-based emotion prediction
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 5145–5149https://doi.org/10.1109/ICASSP.2017.7953137Continuous prediction of dimensional emotions (e.g. arousal and valence) has attracted increasing research interest recently. When processing emotional speech signals, phonetic features have been rarely used due to the assumption that phonetic variability ...
- research-articleMarch 2017
Biologically inspired speech emotion recognition
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 5135–5139https://doi.org/10.1109/ICASSP.2017.7953135Conventional feature-based classification methods do not apply well to automatic recognition of speech emotions, mostly because the precise set of spectral and prosodic features that is required to identify the emotional state of a speaker has not been ...
- research-articleMarch 2017
Effective emotion recognition in movie audio tracks
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 5120–5124https://doi.org/10.1109/ICASSP.2017.7953132This paper addresses the problem of speech emotion recognition from movie audio tracks. The recently collected Acted Facial Expression in the Wild 5.0 database is used. The aim is to discriminate among angry, happy, and neutral. We extract a relatively ...
- research-articleMarch 2017
Multi-task deep neural network with shared hidden layers: Breaking down the wall between emotion representations
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 4990–4994https://doi.org/10.1109/ICASSP.2017.7953106Emotion representations are psychological constructs for modelling, analysing, and recognising emotion, being one essential element of affect. Due to its complexity, the boundaries between different emotion concepts are often fuzzy, which is also ...
- research-articleMarch 2017
End-to-end spoofing detection with raw waveform CLDNNS
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 4860–4864https://doi.org/10.1109/ICASSP.2017.7953080Albeit recent progress in speaker verification generates powerful models, malicious attacks in the form of spoofed speech, are generally not coped with. Recent results in ASVSpoof2015 and BTAS2016 challenges indicate that spoof-aware features are a ...
- research-articleMarch 2017
Stimulated training for automatic speech recognition and keyword search in limited resource conditions
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 4830–4834https://doi.org/10.1109/ICASSP.2017.7953074Training neural network acoustic models on limited quantities of data is a challenging task. A number of techniques have been proposed to improve generalisation. This paper investigates one such technique called stimulated training. It enables standard ...
- research-articleMarch 2017
A new chaotic feature for EEG classification based seizure diagnosis
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 4651–4655https://doi.org/10.1109/ICASSP.2017.7953038Seeking effective measures to characterize the chaotic patterns of EEG signals for seizure diagnosis is a long-term endeavor in the literature. We propose to count the number of zero-crossing (ZC) points on Poincaré surface as a feature when the ...
- research-articleMarch 2017
Linear Discriminant Analysis with few training data
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 4626–4630https://doi.org/10.1109/ICASSP.2017.7953033Statistically-optimal Linear Discriminant Analysis (LDA) is formulated as a maximization that involves the nominal statistics of the classes to be discriminated. In practice, however, these nominal statistics are unknown and estimated from a collection of ...
- research-articleMarch 2017
Bernoulli filter based algorithm for joint target tracking and classification in a cluttered environment
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 4396–4400https://doi.org/10.1109/ICASSP.2017.7952987In this paper, single-target tracking using radar measurements is addressed. Recently, algorithms based on Bernoulli random finite sets have proved efficient in a cluttered environment. However, in Bayesian approaches, the choice of the motion model ...
- research-articleMarch 2017
Deep multi-view models for glitch classification
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 2931–2935https://doi.org/10.1109/ICASSP.2017.7952693Non-cosmic, non-Gaussian disturbances known as “glitches”, show up in gravitational-wave data of the Advanced Laser Interferometer Gravitational-wave Observatory, or aLIGO. In this paper, we propose a deep multi-view convolutional neural ...
- research-articleMarch 2017
Dirichlet Mixture Matching Projection for supervised linear dimensionality reduction of proportional data
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 2806–2810https://doi.org/10.1109/ICASSP.2017.7952668An effective novel algorithm to reduce the dimensionality of labeled proportional data is presented which uses an optimal linear projection to project the data into a low-dimensional space. Assuming that each class of the projected data is generated by a ...
- research-articleMarch 2017
Dynamic Probabilistic Linear Discriminant Analysis for video classification
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 2781–2785https://doi.org/10.1109/ICASSP.2017.7952663Component Analysis (CA) comprises of statistical techniques that decompose signals into appropriate latent components, relevant to a task-at-hand (e.g., clustering, segmentation, classification). Recently, an explosion of research in CA has been witnessed,...
- research-articleMarch 2017
Speech emotion recognition with ensemble learning methods
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 2756–2760https://doi.org/10.1109/ICASSP.2017.7952658In this paper, we propose to apply ensemble learning methods on neural networks to improve the performance of speech emotion recognition tasks. The basic idea is to first divide unbalanced data set into balanced subsets and then combine the predictions of ...
- research-articleMarch 2017
Learning representations of emotional speech with deep convolutional generative adversarial networks
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)Pages 2746–2750https://doi.org/10.1109/ICASSP.2017.7952656Automatically assessing emotional valence in human speech has historically been a difficult task for machine learning algorithms. The subtle changes in the voice of the speaker that are indicative of positive or negative emotional states are often “...