US20070066914A1 - Method and System for Detecting and Classifying Mental States - Google Patents
Method and System for Detecting and Classifying Mental States Download PDFInfo
- Publication number
- US20070066914A1 US20070066914A1 US11/531,238 US53123806A US2007066914A1 US 20070066914 A1 US20070066914 A1 US 20070066914A1 US 53123806 A US53123806 A US 53123806A US 2007066914 A1 US2007066914 A1 US 2007066914A1
- Authority
- US
- United States
- Prior art keywords
- signal
- bio
- mental state
- representations
- features
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/165—Evaluating the state of mind, e.g. depression, anxiety
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/372—Analysis of electroencephalograms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7253—Details of waveform analysis characterised by using transforms
- A61B5/7257—Details of waveform analysis characterised by using transforms using Fourier transforms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
- A61B5/7267—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
Definitions
- the present invention relates generally to the detection and classification of the mental state of a human.
- the invention is suitable for use in an electronic entertainment or other platforms in which electroencephalograph (EEG) data is collected and analyzed in order to determine a subject's response to stimuli, such as an emotional response, or to measure the mental state of a user when they are consciously focusing on a task, image or willed experience, in order to provide control signals to that platform.
- EEG electroencephalograph
- the method and system for detecting and classifying mental states would also be desirable for the method and system for detecting and classifying mental states to be suitable for use in real time applications, with a minimum of time being required to train and develop a usable interactive system. It would also be desirable to provide a method and system for detecting and classifying mental states that ameliorate or overcome one or more disadvantages of known detection and classification methods and systems.
- one aspect of the present invention provides a method of detecting and classifying mental states.
- the method comprises the steps of receiving bio-signals from one or more bio-signal detectors; generating multiple different representations of each bio-signal; determining the value of one or more features of the each bio-signal representation; and comparing the feature values to one or more than one mental state signature, each mental state signature defining reference feature values indicative of a predetermined mental state.
- the step of generating multiple different representations of each bio-signal may comprise the step of dividing the bio-signals into different epochs.
- the step of generating multiple different representations of each bio-signal further comprises the step of generating representations of the bio-signal epochs into one or more different domains.
- Each bio-signal epoch may be divided into one or more than one of different frequency, temporal and spatial domain representations.
- the different frequency domain representations may be obtained by dividing each bio-signal epoch into distinguishable frequency bands.
- the different temporal domain representations may be obtained by dividing each bio-signal epoch into a plurality of time segments. In one or more embodiments, the time segments in each epoch temporally overlap but in other embodiments the time segments in each epoch do not temporally overlap.
- the different spatial domain representations may be obtained by dividing each bio-signal epoch into a plurality of spatially distinguishable channels. Each channel may be derived from a different bio-signal detector.
- the step of determining the value of one or more features of the each bio-signal representation may comprise determining values of features of individual bio-signal representations.
- the feature may comprise, for example, signal power of one or more than one bio-signal representations, signal power of one or more than one spatially distinguishable channels, a change in signal power of one or more than one bio-signal representations or a change in signal power of one or more than one spatially distinguishable channels.
- the step of determining the value of one or more features of the each bio-signal representation may comprise determining values of features between different bio-signal representations. At least coherence or correlation may be detected between different bio-signal representations.
- One or more than one feature may comprise the correlation or coherence between signal power in different spatially distinguishable channels.
- One or more than one feature may comprises correlation or coherence between changes in signal power in different frequency bands.
- the step of determining the value of one or more features of the each bio-signal representation may comprise applying one or more transforms to the different bio-signal representations, such as a Fourier Transform, wavelet transform or other linear or non-linear mathematical transform.
- the step of comparing the feature values to one or more than one mental state signature may comprise using a neural network to classify whether the measured feature values are indicative of the presence of a predefined mental state.
- the step of comparing the feature values to one or more than one mental state signature may comprises performing a distance measure to measure the similarity between the measured feature values and the reference features values to classify whether the measured feature values are indicative of the presence of a predefined mental state.
- the mental state may be an emotional state, but may also result from mental focus on a task, image or other willed experience.
- Another aspect of the invention provides a method of creating a signature for use in a method of detecting and classifying mental states as described above.
- the method may comprise the steps of eliciting a desired mental state from a user; determining the features of the bio-signal representations that most significantly indicate the presence of the desired mental state by the user; and generating the signature from a combination of those features.
- the desired mental state need not be predefined.
- the method may further include the step of using feature values that are determined when the desired mental state is elicited from the user to update the signature for that mental state.
- Yet another aspect of the invention provides an apparatus for detecting and classifying mental states, comprising a processor and associated memory device for carrying out a method as described above.
- a further aspect of the invention provides a computer program product, tangibly stored on machine readable medium, the product comprising instructions operable to cause a processor to carry out a method as described above.
- a still further aspect of the invention provides a computer program product comprising instructions operable to cause a processor to carry out a method as described above.
- FIG. 1 is a schematic diagram of an apparatus for detecting and classifying mental states in accordance with the present invention
- FIG. 2 is a schematic diagram illustrating the position of bio-signal detectors in the form of scalp electrodes forming part of a head set used in the apparatus shown in FIG. 1 ;
- FIGS. 3 and 4 are flow charts illustrating the broad functional steps performed during detection and classification of mental states by the apparatus shown in FIG. 1 ;
- FIG. 5 is a graphical representation of bio-signals processed by the apparatus of FIG. 1 and the transformation of those bio-signals.
- the mental states can be deliberative or non-deliberative.
- deliberative mental states occur when a subject consciously focuses on a task, image or willed experience.
- non-deliberative mental states are mental states, such as emotions, preference, or sensations, which lack the subjective quality of a volitional act.
- the apparatus 100 includes a head set 102 of bio-signal detectors capable of detecting various bio-signals from a subject, particularly electrical signals produced by the body, such as electroencephalograph (EEG) signals, electrooculograph (EOG) signals, electromyograph (EMG) signal, or like signals.
- EEG electroencephalograph
- EEG electrooculograph
- EMG electromyograph
- the apparatus 100 is capable of detection of at least some mental states (both deliberative and non-deliberative) using solely electrical signals, particularly EEG signals, from the subject, and without direct measurement of other physiological processes, such as heart rate, blood pressure, respiration or galvanic skin response, as would be obtained by a heart rate monitor, blood pressure monitor, and the like.
- the mental states that can be detected and classified are more specific than the gross correlation of brain activity of a subject, e.g., as being awake or in a type of sleep (such as REM or a stage of non-REM sleep), conventionally measured using EEG signals.
- a type of sleep such as REM or a stage of non-REM sleep
- specific emotions, such as excitement, or specific willed tasks, such as a command to push or pull an object can be detected.
- the headset 102 includes a series of scalp electrodes for capturing EEG signals from a subject or user.
- EEG signals measured and used by the apparatus 100 can include signals outside the frequency ranges of theta, alpha and beta waves (4-30 Hz), that are commonly analysed in research systems.
- the scalp electrodes may directly contact the scalp or alternately may be of the non-contact type that do not require direct placement on the scalp.
- the headset is generally portable and non-constraining.
- the electrical fluctuations detected over the scalp by the series of scalp sensors are attributed largely to the activity of brain tissue located at or near the skull.
- the source is the electrical activity of the cerebral cortex, a significant portion of which lies on the outer surface of the brain below the scalp.
- the scalp electrodes pick up electrical signals naturally produced by the brain and make it possible to observe electrical impulses across the surface of the brain.
- the headset 102 includes several scalp electrodes, in other embodiments only one or more scalp electrodes, e.g., sixteen electrodes, may be used in the headset.
- Each of the signals detected by the headset 102 of electrodes is fed through a sensor interface 104 , which can include an amplifier to boost signal strength and a filter to remove noise, and then digitized by the analogue to digital converter 106 . Digitized samples of the signal captured by each of the scalp sensors are stored during operation of the apparatus 100 in a data buffer 108 for subsequent processing.
- the apparatus 100 further includes a processing system 109 including a digital signal processor 112 , a co-processing device 110 and associated memory device for storing a series of instructions (otherwise known as a computer program or computer control logic) to cause the processing system 109 to perform desired functional steps.
- the memory includes a series of instructions defining at least one algorithm 114 to be performed by the digital signal processor 112 for detecting and classifying a predetermined type of mental state.
- a corresponding control signal is transmitted in this exemplary embodiment to an input/output interface 116 for transmission via a wireless transmission device 118 to a platform 120 for use as a control input by electronic entertainment applications, programs, simulators or the like.
- the apparatus 100 also enables the generation of signatures for mental states. This can be important since some signatures can define a mental state that can be used across a population. These signatures are then used by the processing system 109 for classification and detection of the mental state for users other than the subject from whom the signatures were generated.
- the algorithms are implemented in software and the series of instructions is stored in the memory of the processing system, e.g., in the memory of the processing system 109 .
- the series of instructions causes the processing system 109 to perform the functions of the invention as described herein.
- the instructions Prior to being loaded into the memory, the instructions can be tangibly embodied in a machine readable storage device, such as a computer disk or memory card, or in a propagated signal.
- the algorithms are implemented primarily in hardware using, for example, hardware components such as application specific integrated circuits (ASICs). Implementation of the hardware state machine so as to perform the functions described herein will be apparent to persons skilled in the relevant art.
- the algorithms are implemented using a combination of software and hardware.
- an FPGA field programmable gate array
- the processing functions could be performed by a single processor.
- the buffer 108 could be eliminated or replaced by a multiplexer (MUX), and the data stored directly in the memory of the processing system.
- MUX could be placed before the A/D converter stage so that only a single A/D converter is needed.
- the connection between the apparatus 100 and the platform 120 can be wired rather than wireless.
- the apparatus 100 can include a headset assembly that includes the headset, a MUX, A/D converter(s) before or after the MUX, a wireless transmission device, a battery for power supply, and a microcontroller to control battery use, send data from the MUX or A/D converter to the wireless chip, and the like.
- the apparatus can also include a separate processor unit that includes a wireless receiver to receive data from the headset assembly, and the processing system, e.g., the digital signal processor and the co-processor.
- the processor unit can be connected to the platform by a wired or wireless connection.
- the apparatus can include a head set assembly as described above, the platform can include a wireless receiver to receive data from the headset assembly, and a digital signal processor dedicated to detection of mental states can be integrated directly into the platform.
- the apparatus can include a head set assembly as described above, the platform can include a wireless receiver to receive data from the headset assembly, and the mental state detection algorithms are performed in the platform by the same processor, e.g., a general purpose digital processor, that executes the application, programs, simulators or the like.
- FIG. 2 illustrates one example of a positioning system 200 of the scalp electrodes forming part of the headset 102 .
- the system 200 of electrode placement shown in FIG. 2 is referred to as the “10-20” system and is based on the relationship between the location of an electrode and the underlying area of cerebral cortex.
- Each point on the electrode placement system 200 indicates a possible scalp electrode position.
- Each site is indicated by a letter to identify the lobe and a number or other letter to identify the hemisphere location.
- the letters F, T, C, P, and O stand for Frontal, Temporal, Central, Parietal and Occipital. Even numbers referred to the right hemisphere and odd numbers refer to the left hemisphere.
- the letter Z refers to an electrode place on the mid-line.
- the mid-line is a line along the scalp on the sagittal plane originating at the nasion and ending at the inion at the back of the head
- the “10” and “20” refer to percentages of the mid-line division.
- the mid-line is divided into 7 positions, namely, Nasion, Fpz, Fz, Cz, Pz, Oz and Inion, and the angular intervals between adjacent positions are 10%, 20%, 20%, 20%, 20% and 10% of the mid-line length respectively.
- the headset 102 including scalp electrodes positioned according to the system 200 , are placed on the head of a subject in order to detect EEG signals.
- FIG. 3 shows a series of steps carried out by the apparatus 100 during the capture of those EEG signals and subsequent data preparation operations carried out by the processing system 109 .
- the EEG signals are captured and then digitised using the analogue to digital converters 106 .
- the data samples are stored in the data buffer 108 .
- the EEG signals detected by the headset 102 may have a range of characteristics, but for the purposes of illustration typical characteristics are as follows: Amplitude 10 - 4000 ⁇ V, Frequency Range 0.16 - 256 Hz and Sampling Rate 128 - 2048 Hz.
- the data samples are conditioned for subsequent analysis.
- Sources of possible noise that are desired to be eliminated from the data samples include external interference introduced in signal collection, storage and retrieval.
- examples of external interference include power line signals at 50/60 Hz and high frequency noise originating from switching circuits residing in the EEG acquisition hardware.
- a typical operation carried out during this conditioning step is the removal of baselines via high pass filters. Additional checks are performed to ensure that data samples are not collected when a poor quality signal is detected from the headset 102 . Signal quality information can be fed back to a user to help them to take corrective action.
- EEG signals consist, in this example, of measurements of the electrical potential at numerous locations on a user's scalp. These signals can be represented as a set of observations X n of some “signal sources” sm where n ⁇ [1:N], m ⁇ [1:M], n is channel index, N is number of channels, m is source index, M is number of sources. If there exists a set of transfer functions F and G that describe the relationship between S m and X n , one can then identify with a certain level of confidence which sources or components have a distinct impact on observation X n , and their characteristics.
- ICA Independent Component Analysis
- the EEG signals are converted, in steps 306 , 308 and 310 , into different representations that facilitate the detection and classification of the mental state of a user of the headset 102 .
- the data samples are firstly divided into equal length time segments within epochs, at step 306 . While in the exemplary embodiment illustrated in FIG. 5 there are seven time segments of equal duration within the epoch, in another embodiment the number and length of the time segments may be altered. Furthermore, in another embodiment, time segments may not be of equal duration and may or may not overlap within an epoch.
- the length of each epoch can vary dynamically depending on events in the detection system such as artefact removal or signature updating. However, in general, an epoch is selected to be sufficiently long that a change in mental state, if one occurs, can be reliably detected.
- FIG. 5 is a graphical illustration of EEG signals detected from the 32 electrodes in the headset 102 .
- Three epochs 500 , 502 and 504 are shown each with 2 seconds before and 2 seconds after the onset of a change in the mental state of a user.
- the baseline before the event is limited to 2 seconds whereas the portion after the event (EEG signal containing emotional response) varies, depending on the current emotion that is being detected.
- the processing system 109 divides the epochs 500 , 502 and 504 into time segments.
- the epoch 500 is divided into 1 second long segments 506 to 518 , each of which overlap by half a second.
- a 4 second long epoch would then yield 7 segments.
- the processing system 109 then acts in steps 308 and 310 to transform the EEG signal into the different representations so that the value of one or more features of each EEG signal representation can be calculated and collated at step 312 .
- the EEG signal can be converted from the time domain (signal intensity as a function of time) into the frequency domain (signal intensity as a function of frequency).
- the EEG signals are band-passed (during transform to frequency domain) with low and high cut-off frequencies of 0.16 and 256 Hz, respectively.
- the EEG signal can be converted into a differential domain (marginal changes in signal intensity as a function of time).
- the frequency domain can also be converted into a differential domain (marginal changes in signal intensity as a function of frequency), although this may require comparison of frequency spectrums from different time segments.
- step 312 the value of one or more features of each EEG signal representation can be calculated (or collected from previous steps if the transform generated scalar values), and the various values assembled to provide a multi-dimensional representation of the mental state of the subject. In addition to values calculated from transformed representations of the EEG signal, some values could be calculated from the original EEG signals.
- the aggregate signal power in each of a plurality of frequency bands can be calculated.
- seven frequency bands are used with the following frequency ranges: ⁇ (2-4Hz), ⁇ (4-8Hz), ⁇ 1( 8-10Hz), ⁇ 2( 10-13Hz), ⁇ 1( 13-20Hz), ⁇ 2( 20-30Hz) and ⁇ (30-45).
- the signal power in each of these frequency bands is calculated.
- the signal power can be calculated for various combinations of channels or bands. For example, the total signal power for each spatial channel (each electrode) across all frequency bands could be determined, or the total signal power for a given frequency band across all channels could be determined.
- both the number of and ranges of the frequency bands may be different to the exemplary embodiment depending notably on the particular application or detection method employed.
- the frequency bands could overlap.
- features other than aggregate signal power such as the real component, phase, peak frequency, or average frequency, could be calculated from the frequency domain representation for each frequency band.
- the signal representations are in the time, frequency and spatial domains.
- the multiple different representations can be denoted as x where n, i, j, k are epoch, channel, frequency band, and segment index, respectively. Typical values for these parameters are: i ⁇ [1:32] 32 spatially distinguishable channels (referenced Fp 1 to CPz) j ⁇ [1:7] 7 frequency distinguishable bands (referenced ⁇ to ⁇ )
- Other common features to be calculated by the processing system 109 at step 312 include the signal power in each channel, the marginal changes of the power in each frequency band in each channel, the correlations/coherence between different channels, and the correlations between the marginal changes of the powers in each frequency band.
- the choice between these properties depends on the types of mental state that are desired to distinguish. In general, marginal properties are more important in case of short term emotional burst whereas in a long term mental state, other properties are more significant.
- a variety of techniques can be used to transform the EEG signal into the different representations and to measure the value of the various features of the EEG signal representations.
- traditional frequency decomposition techniques such as Fast Fourier Transform (FFT) and band-pass filtering
- FFT Fast Fourier Transform
- band-pass filtering measures of signal coherence and correlation
- the coherence or correlation values can be collated in step 312 to become part of the multi-dimensional representation of the mental state.
- the correlations/coherence is calculated between different channels, this could also be considered a domain, e.g., a spatial coherence/correlation domain (coherence/correlation as a function of electrode pairs).
- a wavelet transform, dynamical systems analysis or other linear or non-linear mathematical transform may be used in step 310 .
- the FFT is an efficient algorithm of the discrete Fourier transform which reduces the number of computations needed for N data points from 2N 2 to 2N log 2 N. Passing a data channel in time domain through an FFT, will generate a description for that data segment in the complex frequency domain.
- Coherence is a measure of the amount of association or coupling between two different time series.
- a coherence computation can be carried out between two channels a and b, in frequency band Cn, where the Fourier components of channels a and b of frequency f ⁇ are xa ⁇ and xb ⁇ is:
- a coherence computation can be carried out between two channels a and b, in frequency band ⁇ n, where the Fourier components of channels a and b of frequency f ⁇ are X an and X bu is: C ab ⁇ ⁇ ⁇ n ⁇ ⁇ f ⁇ ⁇ ⁇ n ⁇ x a ⁇ ⁇ ⁇ ⁇ X b ⁇ ⁇ ⁇ * ⁇ f ⁇ ⁇ ⁇ n ⁇ x a ⁇ ⁇ ⁇ 2 ⁇ ⁇ f ⁇ ⁇ ⁇ n x b ⁇ ⁇ ⁇ 2
- Correlation is an alternative to coherence to measure the amount of association or coupling between two different time series.
- FIG. 4 shows in the various data processing operations, preferably carried out in real-time, which are then carried out by the processing system 109 .
- the calculated values of one or more features of each signal representation are compared to one or more mental state signatures stored in the memory of the processing system 109 to classify the mental state of the user.
- Each mental state signature defines reference feature values that are indicative of a predetermined mental state.
- a number of techniques can be used by the processing device 109 to match the pattern of the calculated feature values to the mental state signatures.
- a multi layer perceptron neural network can be used to classify whether a signal representation is indicative of a mental state corresponding to a stored signature.
- the processing system 109 can use a standard perceptron with n inputs, one or more hidden layers of m hidden nodes and an output layer with l output nodes. The number of output nodes is determined by how many independent mental states the processing system is trying to recognize. Alternately, the number of networks used may be varied according to the number of mental states being detected.
- F 1 and F 2 are the activation functions that act on the components of the column vectors separately to produce another column vector and Y is the output vector.
- the activation function determines how the node is activated by the inputs.
- the processing system 109 uses a sigmoid function. Other possibilities are a hyperbolic tangent function or even a linear function.
- the weight matrices can be determined either recursively or all at once.
- Distance measures for determining similarity of an unknown sample set to a known one can be used as an alternative technique to the neural network.
- Distances such as the modified Mahalanobis distance, the standardised Euclidean distance and a projection distance can be used to determine the similarity between the calculated feature values and the reference feature values defined by the various mental state signatures to thereby indicate how well a user's mental state reflects each of those signatures.
- the mental state signatures and weights can be predefined. For example, for some mental states, signatures are sufficiently uniform across a human population that once a particular signature is developed (e.g., by deliberately evoking the mental state in test subjects and measuring the resulting signature), this signature can be loaded into the memory and used without calibration by a particular user. On the other hand, for some mental states, signatures are sufficiently non-uniform across the human population that predefined signatures cannot be used or can be used only with limited satisfaction by the subject. In such a case, signatures (and weights) can be generated by the apparatus 100 , as discussed below, for the particular user (e.g., by requesting that the user make a willed effort for some result, and measuring the resulting signature).
- the accuracy of a signature and/or weights that was predetermined from test subjects can be improved by calibration for a particular user.
- the user could be exposed to a stimulus that is expected to produce a particular mental state, the resulting bio-signals compared to a predefined signature.
- the user can be queried regarding the strength of the mental state, and the resulting feedback from the user applied to adjust the weights.
- calibration could be performed by a statistical analysis of the range of stored multi-dimensional representations.
- the user can be requested to make a willed effort for some result, and the multi-dimensional representation of the resulting mental state can be used to adjust the signature or weights.
- the apparatus 100 can also be adapted to generate and update signatures indicative of a user's various mental states.
- data samples of the multiple different representations of the EEG signals generated in steps 300 to 310 are saved by the processing system 109 in memory, preferably for all users of the apparatus 100 .
- An evolving database of data samples is thus created which allows the processing device 109 to progressively improve the accuracy of mental state detection for one or more users of the apparatus 100 .
- one or more statistical techniques are applied to determine how significant each of the features is in characterising different mental states. Different coordinates are given a rating based on how well they differentiate.
- the techniques implemented by the processing system 109 use a hypothesis testing procedure to highlight regions of the brain or brainwave frequencies from the EEG signals, which activate during different mental states. At a simplistic level, this approach typically involves determining whether some averaged (mean) power value for a representation of the EEG signal differs to another, given a set of data samples from a defined time period. Such a “mean difference” test is performed by the processing system 109 for every signal representation.
- the processing system 109 implements an Analysis of Variance (ANOVA) F ratio test to search for differences in activation, combined with a paired Student's T test.
- ANOVA Analysis of Variance
- the T test is functionally equivalent to the one way ANOVA test for two groups, but also allows for a measure of direction of mean difference to be analysed (i.e. whether the mean value of a mental state 1 is larger than the mean value for a mental state 2 , or vice versa).
- t mean ⁇ ⁇ of ⁇ ⁇ mental ⁇ ⁇ state ⁇ ⁇ 1 - mean ⁇ ⁇ of ⁇ ⁇ mental ⁇ ⁇ state ⁇ ⁇ 2 ( variance ⁇ ⁇ of ⁇ ⁇ mental ⁇ ⁇ state ⁇ ⁇ 1 n ⁇ ⁇ for ⁇ ⁇ mental ⁇ ⁇ state ⁇ ⁇ 1 ) + ( variance ⁇ ⁇ of ⁇ ⁇ mental ⁇ ⁇ state ⁇ ⁇ 2 n ⁇ ⁇ for ⁇ ⁇ mental ⁇ ⁇ state ⁇ ⁇ 2 )
- the “n” which makes the denominator in the lower half of the T equation is the number of time series recorded for a particular mental state which make up the means being contrasted in the numerator. (i.e. the number of overlapping or non-overlapping epochs recorded during an update.
- the subsequent t value is used in a variety of ways by the processing system 109 , including the rating of the feature space dimensions to determine the significance level of the many thousands of features that are typically analysed.
- Features may be weighted on a linear or non-linear scale, or in a binary fashion by removing those features which do not meet a certain level of significance.
- the range of t values that will be generated from the many thousands of hypothesis tests during a signature update can be used to give an overall indication to the user of how far separated the detected mental states are during that update.
- the t value is an indication of that particular mean separation for the two actions, and the range of t values across all coordinates provides a metric for how well, on average, all of the coordinates separate.
- the above-mentioned techniques are termed univariate approaches as the processing system 109 performs the analysis for each individual coordinate at a time, and make feature selections decisions based on those individual t test or ANOVA test results. Corrections may be made at step 406 to adjust for the increased chance of probability error due to the use of the mass univariate approach.
- Statistical techniques suitable for this purpose include the following multiplicity correction methods: Bonferroni, False Discovery Rate and Dunn Sidack.
- the processing system 109 can therefore employ such techniques as Discriminant Function Analysis and Multivariate analysis of variance (MANOVA), which not only provides a means to select feature space in a multivariate manner, but also allows the use of eigenvalues created during the analysis to actually classify unknown signal representations in a real-time environment.
- MANOVA Discriminant Function Analysis and Multivariate analysis of variance
- the processing system 109 prepares for classifying incoming real-time data by weighting the coordinates so that those with the greatest significance in detecting a particular mental state are given precedence. This can be carried out by applying adaptive weight preparation, neural network training or statistical weightings.
- the signatures stored in the memory of the processing system 109 are updated or calibrated at step 410 .
- the updating process involves taking data samples, which is added to the evolving database. This data is elicited for the detection of a particular mental state. For example, to update a willed effort mental state, a user is prompted to focus on that willed effort and signal data samples are added to the database and used by the processing system 109 to modify the signature for that detection.
- detections can provide feedback for updating the signatures that define that detection. For example, if a user wants to improve their signature for willing an object to be pushed away, the existing detection can be used to provide feedback as the signature is updated. In that scenario, the user sees the detection improving, which provides reinforcement to the updating process.
- a supervised learning algorithm dynamically takes the update data from step 410 and combines it with the evolving database of recorded data samples to improve the signatures for the mental state that has been updated.
- Signatures may initially be empty or be prepared using historical data from other users which may have been combined to form a reference or universal starting signature.
- the signature for the mental state that has been updated is made available for mental state classification (at step 400 ) as well as signature feedback rating at step 416 .
- a rating is available in real-time which reflects how the mental state detection is progressing.
- the apparatus 100 can therefore provide feedback to a user to enable them to observe the evolution of a signature over time.
- the discussion above has focused on determination of the presence or absence of a particular mental state. However, it is also possible to determine the intensity of that particular mental state.
- the intensity can be determined by measuring the “distance” of the transformed signal from the user to a signature. The greater the distance, the lower the intensity.
- the user can be queried regarding the strength of the mental state. The resulting feedback from the user is applied to adjust the weights to calibrate the distance to the intensity scale.
- the apparatus 100 advantageously enables the online creation of signatures in near real-time.
- the detection of a user's mental state and creation of a signature can be achieved in a few minutes, and then refined over time as the user's signature for that mental state is updated. This can be very important in interactive applications, where a short term result is important as well as incremental improvement over time.
- the apparatus 100 advantageously enables the detection of a mental state having a pregenerated signature (whether predefined or created for the particular user) in real-time.
- a mental state having a pregenerated signature whether predefined or created for the particular user
- the detection of the presence or absence of a user's particular mental state, or the intensity of that particular mental state can be achieved in real-time.
- signatures can be created for mental states that need not be predefined.
- the apparatus 100 can classify mental states that are recorded for, not just mental states that are predefined and elicited via pre-defined stimuli.
- the mental state detection system described herein can utilize a huge number of feature dimensions which cover many spatial areas, frequency ranges and other dimensions.
- the system ranks features by their ability to distinguish a particular mental state, thus highlighting those features that are better able to capture the brain's activity in a given mental state.
- the features selected by the user reflect characteristics of the electrical signals measured on the scalp that are able to distinguish a particular mental state, and therefore reflect how the signals in their particular cortex are manifested on the scalp.
- the user's individual electrical signals that indicate a particular mental state have been identified and stored in a signature. This permits real-time mental state detection or generation within minutes, through algorithms which compensate for the individuality of EEG.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Psychiatry (AREA)
- Veterinary Medicine (AREA)
- Animal Behavior & Ethology (AREA)
- Public Health (AREA)
- General Health & Medical Sciences (AREA)
- Surgery (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Psychology (AREA)
- Artificial Intelligence (AREA)
- Social Psychology (AREA)
- Developmental Disabilities (AREA)
- Hospice & Palliative Care (AREA)
- Child & Adolescent Psychology (AREA)
- Educational Technology (AREA)
- Evolutionary Computation (AREA)
- Fuzzy Systems (AREA)
- Mathematical Physics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physiology (AREA)
- Signal Processing (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)
Abstract
A method of detecting and classifying mental states, comprising the steps of receiving bio-signals from one or more bio-signal detectors; generating multiple different representations of each bio-signal; determining the value of one or more features of the each bio-signal representation; and comparing the feature values to one or more than one mental state signature, each mental state signature defining reference feature values indicative of a predetermined mental state.
Description
- This application is a continuation-in-part of and claims priority to U.S. application Ser. No. 11/225,835, filed on Sep. 12, 2005, which is incorporated by reference.
- The present invention relates generally to the detection and classification of the mental state of a human. The invention is suitable for use in an electronic entertainment or other platforms in which electroencephalograph (EEG) data is collected and analyzed in order to determine a subject's response to stimuli, such as an emotional response, or to measure the mental state of a user when they are consciously focusing on a task, image or willed experience, in order to provide control signals to that platform. It will therefore be convenient to describe the invention in relation to that exemplary but non-limiting application.
- Interactions between humans and machines are usually restricted to the use of cumbersome input devices such as keyboards, joysticks and other manually operable controls. A number of input devices have been developed to assist disabled users in providing commands without requiring the use of manually operable controls. Some of these input devices detect eyeball movement or are voice activated to minimize the physical movement required by a user to operate these input devices. A number of studies have been conducted to determine the feasibility of eliminating physical movement from control inputs by detecting the mental state of a user. Most of these studies have been conducted in the medical sphere to determine the responsiveness of patients to external stimuli in situations where those patients are unable to otherwise communicate with medical staff.
- To date though, attempts to detect the mental state of a user have been rudimentary only, and are unsuited to use in complex environments, such as contemporary software-based gaming or like platforms.
- It would be desirable to provide a method and system for detecting and classifying a range of mental states in a manner that was suitable for use in a variety of contemporary applications. It would also be desirable for that system and method to be adaptable to suit a number of applications, without requiring the use of significant data processing resources.
- It would also be desirable for the method and system for detecting and classifying mental states to be suitable for use in real time applications, with a minimum of time being required to train and develop a usable interactive system. It would also be desirable to provide a method and system for detecting and classifying mental states that ameliorate or overcome one or more disadvantages of known detection and classification methods and systems.
- There also exists a need to provide technology that simplifies man machine interactions. It would be preferable for this technology to be robust, powerful and adaptable to a number of platforms and environments. It would also be desirable for technology to optimize the use of natural human interaction techniques so that the man machine interaction is as natural as possible for a human user.
- With that in mind, one aspect of the present invention provides a method of detecting and classifying mental states. The method comprises the steps of receiving bio-signals from one or more bio-signal detectors; generating multiple different representations of each bio-signal; determining the value of one or more features of the each bio-signal representation; and comparing the feature values to one or more than one mental state signature, each mental state signature defining reference feature values indicative of a predetermined mental state.
- The step of generating multiple different representations of each bio-signal may comprise the step of dividing the bio-signals into different epochs. Preferably, the step of generating multiple different representations of each bio-signal further comprises the step of generating representations of the bio-signal epochs into one or more different domains. Each bio-signal epoch may be divided into one or more than one of different frequency, temporal and spatial domain representations.
- The different frequency domain representations may be obtained by dividing each bio-signal epoch into distinguishable frequency bands. The different temporal domain representations may be obtained by dividing each bio-signal epoch into a plurality of time segments. In one or more embodiments, the time segments in each epoch temporally overlap but in other embodiments the time segments in each epoch do not temporally overlap.
- The different spatial domain representations may be obtained by dividing each bio-signal epoch into a plurality of spatially distinguishable channels. Each channel may be derived from a different bio-signal detector.
- The step of determining the value of one or more features of the each bio-signal representation may comprise determining values of features of individual bio-signal representations. The feature may comprise, for example, signal power of one or more than one bio-signal representations, signal power of one or more than one spatially distinguishable channels, a change in signal power of one or more than one bio-signal representations or a change in signal power of one or more than one spatially distinguishable channels.
- The step of determining the value of one or more features of the each bio-signal representation may comprise determining values of features between different bio-signal representations. At least coherence or correlation may be detected between different bio-signal representations. One or more than one feature may comprise the correlation or coherence between signal power in different spatially distinguishable channels. One or more than one feature may comprises correlation or coherence between changes in signal power in different frequency bands.
- The step of determining the value of one or more features of the each bio-signal representation may comprise applying one or more transforms to the different bio-signal representations, such as a Fourier Transform, wavelet transform or other linear or non-linear mathematical transform.
- The step of comparing the feature values to one or more than one mental state signature may comprise using a neural network to classify whether the measured feature values are indicative of the presence of a predefined mental state.
- The step of comparing the feature values to one or more than one mental state signature may comprises performing a distance measure to measure the similarity between the measured feature values and the reference features values to classify whether the measured feature values are indicative of the presence of a predefined mental state.
- The mental state may be an emotional state, but may also result from mental focus on a task, image or other willed experience.
- Another aspect of the invention provides a method of creating a signature for use in a method of detecting and classifying mental states as described above. The method may comprise the steps of eliciting a desired mental state from a user; determining the features of the bio-signal representations that most significantly indicate the presence of the desired mental state by the user; and generating the signature from a combination of those features. The desired mental state need not be predefined.
- The method may further include the step of using feature values that are determined when the desired mental state is elicited from the user to update the signature for that mental state.
- Yet another aspect of the invention provides an apparatus for detecting and classifying mental states, comprising a processor and associated memory device for carrying out a method as described above.
- A further aspect of the invention provides a computer program product, tangibly stored on machine readable medium, the product comprising instructions operable to cause a processor to carry out a method as described above.
- A still further aspect of the invention provides a computer program product comprising instructions operable to cause a processor to carry out a method as described above.
- These and other features, aspects and advantages of the present invention will become better understood with regard to the following description, appended claims, and accompanying figures which depict various views and embodiments of the device, and some of the steps in certain embodiments of the method of the present invention, where:
-
FIG. 1 is a schematic diagram of an apparatus for detecting and classifying mental states in accordance with the present invention; -
FIG. 2 is a schematic diagram illustrating the position of bio-signal detectors in the form of scalp electrodes forming part of a head set used in the apparatus shown inFIG. 1 ; - FIGS. 3 and 4 are flow charts illustrating the broad functional steps performed during detection and classification of mental states by the apparatus shown in
FIG. 1 ; and -
FIG. 5 is a graphical representation of bio-signals processed by the apparatus ofFIG. 1 and the transformation of those bio-signals. - Turning now to
FIG. 1 , there is shown anapparatus 100 for detecting and classifying mental states. The mental states can be deliberative or non-deliberative. In general, deliberative mental states occur when a subject consciously focuses on a task, image or willed experience. In contrast, non-deliberative mental states are mental states, such as emotions, preference, or sensations, which lack the subjective quality of a volitional act. - The
apparatus 100 includes ahead set 102 of bio-signal detectors capable of detecting various bio-signals from a subject, particularly electrical signals produced by the body, such as electroencephalograph (EEG) signals, electrooculograph (EOG) signals, electromyograph (EMG) signal, or like signals. Theapparatus 100 is capable of detection of at least some mental states (both deliberative and non-deliberative) using solely electrical signals, particularly EEG signals, from the subject, and without direct measurement of other physiological processes, such as heart rate, blood pressure, respiration or galvanic skin response, as would be obtained by a heart rate monitor, blood pressure monitor, and the like. In addition, the mental states that can be detected and classified are more specific than the gross correlation of brain activity of a subject, e.g., as being awake or in a type of sleep (such as REM or a stage of non-REM sleep), conventionally measured using EEG signals. For example, specific emotions, such as excitement, or specific willed tasks, such as a command to push or pull an object, can be detected. - In the exemplary embodiment illustrated in the drawings, the
headset 102 includes a series of scalp electrodes for capturing EEG signals from a subject or user. It should be noted, however, that the EEG signals measured and used by theapparatus 100 can include signals outside the frequency ranges of theta, alpha and beta waves (4-30 Hz), that are commonly analysed in research systems. The scalp electrodes may directly contact the scalp or alternately may be of the non-contact type that do not require direct placement on the scalp. Unlike systems that provide high-resolution 3-D brain scans, e.g., MRI or CAT scans, the headset is generally portable and non-constraining. - The electrical fluctuations detected over the scalp by the series of scalp sensors are attributed largely to the activity of brain tissue located at or near the skull. The source is the electrical activity of the cerebral cortex, a significant portion of which lies on the outer surface of the brain below the scalp. The scalp electrodes pick up electrical signals naturally produced by the brain and make it possible to observe electrical impulses across the surface of the brain. Although in this exemplary embodiment the
headset 102 includes several scalp electrodes, in other embodiments only one or more scalp electrodes, e.g., sixteen electrodes, may be used in the headset. - Each of the signals detected by the
headset 102 of electrodes is fed through asensor interface 104, which can include an amplifier to boost signal strength and a filter to remove noise, and then digitized by the analogue todigital converter 106. Digitized samples of the signal captured by each of the scalp sensors are stored during operation of theapparatus 100 in adata buffer 108 for subsequent processing. - The
apparatus 100 further includes aprocessing system 109 including adigital signal processor 112, aco-processing device 110 and associated memory device for storing a series of instructions (otherwise known as a computer program or computer control logic) to cause theprocessing system 109 to perform desired functional steps. Notably, the memory includes a series of instructions defining at least onealgorithm 114 to be performed by thedigital signal processor 112 for detecting and classifying a predetermined type of mental state. Upon detection of each predefined type of mental state, a corresponding control signal is transmitted in this exemplary embodiment to an input/output interface 116 for transmission via awireless transmission device 118 to aplatform 120 for use as a control input by electronic entertainment applications, programs, simulators or the like. - As well as enabling the classification and detection of mental states, the
apparatus 100 also enables the generation of signatures for mental states. This can be important since some signatures can define a mental state that can be used across a population. These signatures are then used by theprocessing system 109 for classification and detection of the mental state for users other than the subject from whom the signatures were generated. - In one embodiment, the algorithms are implemented in software and the series of instructions is stored in the memory of the processing system, e.g., in the memory of the
processing system 109. The series of instructions causes theprocessing system 109 to perform the functions of the invention as described herein. Prior to being loaded into the memory, the instructions can be tangibly embodied in a machine readable storage device, such as a computer disk or memory card, or in a propagated signal. In another embodiment, the algorithms are implemented primarily in hardware using, for example, hardware components such as application specific integrated circuits (ASICs). Implementation of the hardware state machine so as to perform the functions described herein will be apparent to persons skilled in the relevant art. In yet another embodiment, the algorithms are implemented using a combination of software and hardware. - Other implementations of the
apparatus 100 are possible. Instead of a digital signal processor, an FPGA (field programmable gate array) could be used. Rather than a separate digital signal processor and co-processor, the processing functions could be performed by a single processor. Thebuffer 108 could be eliminated or replaced by a multiplexer (MUX), and the data stored directly in the memory of the processing system. A MUX could be placed before the A/D converter stage so that only a single A/D converter is needed. The connection between theapparatus 100 and theplatform 120 can be wired rather than wireless. - Although the
apparatus 100 is illustrated inFIG. 1 with all processing functions occurring in a single device that is external to the platform, other implementations are possible. For example, the apparatus can include a headset assembly that includes the headset, a MUX, A/D converter(s) before or after the MUX, a wireless transmission device, a battery for power supply, and a microcontroller to control battery use, send data from the MUX or A/D converter to the wireless chip, and the like. The apparatus can also include a separate processor unit that includes a wireless receiver to receive data from the headset assembly, and the processing system, e.g., the digital signal processor and the co-processor. The processor unit can be connected to the platform by a wired or wireless connection. As another example, the apparatus can include a head set assembly as described above, the platform can include a wireless receiver to receive data from the headset assembly, and a digital signal processor dedicated to detection of mental states can be integrated directly into the platform. As yet another example, the apparatus can include a head set assembly as described above, the platform can include a wireless receiver to receive data from the headset assembly, and the mental state detection algorithms are performed in the platform by the same processor, e.g., a general purpose digital processor, that executes the application, programs, simulators or the like. -
FIG. 2 illustrates one example of apositioning system 200 of the scalp electrodes forming part of theheadset 102. Thesystem 200 of electrode placement shown inFIG. 2 is referred to as the “10-20” system and is based on the relationship between the location of an electrode and the underlying area of cerebral cortex. Each point on theelectrode placement system 200 indicates a possible scalp electrode position. Each site is indicated by a letter to identify the lobe and a number or other letter to identify the hemisphere location. The letters F, T, C, P, and O stand for Frontal, Temporal, Central, Parietal and Occipital. Even numbers referred to the right hemisphere and odd numbers refer to the left hemisphere. The letter Z refers to an electrode place on the mid-line. The mid-line is a line along the scalp on the sagittal plane originating at the nasion and ending at the inion at the back of the head The “10” and “20” refer to percentages of the mid-line division. The mid-line is divided into 7 positions, namely, Nasion, Fpz, Fz, Cz, Pz, Oz and Inion, and the angular intervals between adjacent positions are 10%, 20%, 20%, 20%, 20% and 10% of the mid-line length respectively. - In operation, the
headset 102, including scalp electrodes positioned according to thesystem 200, are placed on the head of a subject in order to detect EEG signals.FIG. 3 shows a series of steps carried out by theapparatus 100 during the capture of those EEG signals and subsequent data preparation operations carried out by theprocessing system 109. - At
step 300, the EEG signals are captured and then digitised using the analogue todigital converters 106. The data samples are stored in thedata buffer 108. The EEG signals detected by theheadset 102 may have a range of characteristics, but for the purposes of illustration typical characteristics are as follows: Amplitude 10 - 4000 μV, Frequency Range 0.16 - 256 Hz and Sampling Rate 128 - 2048 Hz. - At
step 302, the data samples are conditioned for subsequent analysis. Sources of possible noise that are desired to be eliminated from the data samples include external interference introduced in signal collection, storage and retrieval. For EEG signals, examples of external interference include power line signals at 50/60 Hz and high frequency noise originating from switching circuits residing in the EEG acquisition hardware. A typical operation carried out during this conditioning step is the removal of baselines via high pass filters. Additional checks are performed to ensure that data samples are not collected when a poor quality signal is detected from theheadset 102. Signal quality information can be fed back to a user to help them to take corrective action. - An
artefact removal step 304 is then carried out to remove signal interference. EEG signals consist, in this example, of measurements of the electrical potential at numerous locations on a user's scalp. These signals can be represented as a set of observations Xn of some “signal sources” sm where nε[1:N], mΕ[1:M], n is channel index, N is number of channels, m is source index, M is number of sources. If there exists a set of transfer functions F and G that describe the relationship between Sm and Xn, one can then identify with a certain level of confidence which sources or components have a distinct impact on observation Xn, and their characteristics. Different techniques such as Independent Component Analysis (ICA) are applied by theapparatus 100 to find components with greatest impact on the amplitude of Xn. These components often result from interference such as power line noise, signal drop outs, and muscle, eye blink, and eye movement artefacts. - The EEG signals are converted, in
steps headset 102. - The data samples are firstly divided into equal length time segments within epochs, at
step 306. While in the exemplary embodiment illustrated inFIG. 5 there are seven time segments of equal duration within the epoch, in another embodiment the number and length of the time segments may be altered. Furthermore, in another embodiment, time segments may not be of equal duration and may or may not overlap within an epoch. The length of each epoch can vary dynamically depending on events in the detection system such as artefact removal or signature updating. However, in general, an epoch is selected to be sufficiently long that a change in mental state, if one occurs, can be reliably detected.FIG. 5 is a graphical illustration of EEG signals detected from the 32 electrodes in theheadset 102. Threeepochs - The
processing system 109 divides theepochs FIG. 5 , theepoch 500 is divided into 1 secondlong segments 506 to 518, each of which overlap by half a second. A 4 second long epoch would then yield 7 segments. - The
processing system 109 then acts insteps step 312. For example, for each time segment and each channel, the EEG signal can be converted from the time domain (signal intensity as a function of time) into the frequency domain (signal intensity as a function of frequency). In an exemplary embodiment, the EEG signals are band-passed (during transform to frequency domain) with low and high cut-off frequencies of 0.16 and 256 Hz, respectively. - As another example, the EEG signal can be converted into a differential domain (marginal changes in signal intensity as a function of time). The frequency domain can also be converted into a differential domain (marginal changes in signal intensity as a function of frequency), although this may require comparison of frequency spectrums from different time segments.
- In
step 312 the value of one or more features of each EEG signal representation can be calculated (or collected from previous steps if the transform generated scalar values), and the various values assembled to provide a multi-dimensional representation of the mental state of the subject. In addition to values calculated from transformed representations of the EEG signal, some values could be calculated from the original EEG signals. - As an example of the calculation of the value of a feature, in the frequency domain, the aggregate signal power in each of a plurality of frequency bands can be calculated. In an exemplary embodiment described herein, seven frequency bands are used with the following frequency ranges: δ(2-4Hz), θ(4-8Hz), α1(8-10Hz), α2(10-13Hz), β1(13-20Hz), β2(20-30Hz) and γ(30-45). The signal power in each of these frequency bands is calculated. In addition, the signal power can be calculated for various combinations of channels or bands. For example, the total signal power for each spatial channel (each electrode) across all frequency bands could be determined, or the total signal power for a given frequency band across all channels could be determined.
- In other embodiments of the invention, both the number of and ranges of the frequency bands may be different to the exemplary embodiment depending notably on the particular application or detection method employed. In addition, the frequency bands could overlap. Furthermore, features other than aggregate signal power, such as the real component, phase, peak frequency, or average frequency, could be calculated from the frequency domain representation for each frequency band.
- In this exemplary embodiment, the signal representations are in the time, frequency and spatial domains. The multiple different representations can be denoted as x where n, i, j, k are epoch, channel, frequency band, and segment index, respectively. Typical values for these parameters are:
iε[1:32] 32 spatially distinguishable channels (referenced Fp1 to CPz)
jε[1:7] 7 frequency distinguishable bands (referenced δto γ) - The operations carried out in step 310-312 often produce a large number of state variables. For example, calculating correlation values for 2 four-second long epochs consisting of 32 channels, using 7 frequency bands gives more than 1 million state variables:
32C2×72×72=1190896 - Since individual EEG signals and combinations of EEG signals from different sensors can be used, as well as wide range of features from a variety of different transform domains, the number of dimensions to be analysed by the
processing system 109 is extremely large. This huge number of dimensions enables theprocessing system 109 to detect a wide range of mental states, since the entire or a significant portion of the cortex and a full range of features are considered in detecting and classifying a mental state. - Other common features to be calculated by the
processing system 109 atstep 312 include the signal power in each channel, the marginal changes of the power in each frequency band in each channel, the correlations/coherence between different channels, and the correlations between the marginal changes of the powers in each frequency band. The choice between these properties depends on the types of mental state that are desired to distinguish. In general, marginal properties are more important in case of short term emotional burst whereas in a long term mental state, other properties are more significant. - A variety of techniques can be used to transform the EEG signal into the different representations and to measure the value of the various features of the EEG signal representations. For example, traditional frequency decomposition techniques, such as Fast Fourier Transform (FFT) and band-pass filtering, can be carried out by the
processing system 109 atstep 308, whilst measures of signal coherence and correlation can be carried out at step 310 (in this later case, the coherence or correlation values can be collated instep 312 to become part of the multi-dimensional representation of the mental state). Assuming that the correlations/coherence is calculated between different channels, this could also be considered a domain, e.g., a spatial coherence/correlation domain (coherence/correlation as a function of electrode pairs). For example, in other embodiments, a wavelet transform, dynamical systems analysis or other linear or non-linear mathematical transform may be used instep 310. - The FFT is an efficient algorithm of the discrete Fourier transform which reduces the number of computations needed for N data points from 2N2 to 2N log2N. Passing a data channel in time domain through an FFT, will generate a description for that data segment in the complex frequency domain.
- Coherence is a measure of the amount of association or coupling between two different time series. Thus, a coherence computation can be carried out between two channels a and b, in frequency band Cn, where the Fourier components of channels a and b of frequency fμ are xaμ and xbμ is:
- Thus, a coherence computation can be carried out between two channels a and b, in frequency band ωn, where the Fourier components of channels a and b of frequency fμ are Xan and Xbu is:
- Correlation is an alternative to coherence to measure the amount of association or coupling between two different time series. For the same assumption as of coherence section above, a correlation rab, computation can be carried out between the signals of two channels Xa(ti) and Xb(ti), is defined as,
where Xai and Xb have already had common band-pass filtering 1010 applied to them. -
FIG. 4 shows in the various data processing operations, preferably carried out in real-time, which are then carried out by theprocessing system 109. Atstep 400, the calculated values of one or more features of each signal representation are compared to one or more mental state signatures stored in the memory of theprocessing system 109 to classify the mental state of the user. Each mental state signature defines reference feature values that are indicative of a predetermined mental state. - A number of techniques can be used by the
processing device 109 to match the pattern of the calculated feature values to the mental state signatures. A multi layer perceptron neural network can be used to classify whether a signal representation is indicative of a mental state corresponding to a stored signature. Theprocessing system 109 can use a standard perceptron with n inputs, one or more hidden layers of m hidden nodes and an output layer with l output nodes. The number of output nodes is determined by how many independent mental states the processing system is trying to recognize. Alternately, the number of networks used may be varied according to the number of mental states being detected. The output vector of the neural network can be expressed as,
Y=F2(W2·F1(W1·X))
where W1 is m by (n+1) weight matrix, W2 is an l by (m+1) weight matrix (the additional column in the weight matrices allows for a bias term to be added) and X=(X1,X2, . . . Xn) is the input vector. F1 and F2 are the activation functions that act on the components of the column vectors separately to produce another column vector and Y is the output vector. The activation function determines how the node is activated by the inputs. Theprocessing system 109 uses a sigmoid function. Other possibilities are a hyperbolic tangent function or even a linear function. The weight matrices can be determined either recursively or all at once. - Distance measures for determining similarity of an unknown sample set to a known one can be used as an alternative technique to the neural network. Distances such as the modified Mahalanobis distance, the standardised Euclidean distance and a projection distance can be used to determine the similarity between the calculated feature values and the reference feature values defined by the various mental state signatures to thereby indicate how well a user's mental state reflects each of those signatures.
- The mental state signatures and weights can be predefined. For example, for some mental states, signatures are sufficiently uniform across a human population that once a particular signature is developed (e.g., by deliberately evoking the mental state in test subjects and measuring the resulting signature), this signature can be loaded into the memory and used without calibration by a particular user. On the other hand, for some mental states, signatures are sufficiently non-uniform across the human population that predefined signatures cannot be used or can be used only with limited satisfaction by the subject. In such a case, signatures (and weights) can be generated by the
apparatus 100, as discussed below, for the particular user (e.g., by requesting that the user make a willed effort for some result, and measuring the resulting signature). Of course, for some mental states the accuracy of a signature and/or weights that was predetermined from test subjects can be improved by calibration for a particular user. For example, to calibrate the subjective intensity of a non-deliberative mental state for a particular user, the user could be exposed to a stimulus that is expected to produce a particular mental state, the resulting bio-signals compared to a predefined signature. The user can be queried regarding the strength of the mental state, and the resulting feedback from the user applied to adjust the weights. Alternatively, calibration could be performed by a statistical analysis of the range of stored multi-dimensional representations. To calibrate a deliberative mental state, the user can be requested to make a willed effort for some result, and the multi-dimensional representation of the resulting mental state can be used to adjust the signature or weights. - The
apparatus 100 can also be adapted to generate and update signatures indicative of a user's various mental states. Atstep 402, data samples of the multiple different representations of the EEG signals generated insteps 300 to 310 are saved by theprocessing system 109 in memory, preferably for all users of theapparatus 100. An evolving database of data samples is thus created which allows theprocessing device 109 to progressively improve the accuracy of mental state detection for one or more users of theapparatus 100. - At
step 404, one or more statistical techniques are applied to determine how significant each of the features is in characterising different mental states. Different coordinates are given a rating based on how well they differentiate. The techniques implemented by theprocessing system 109 use a hypothesis testing procedure to highlight regions of the brain or brainwave frequencies from the EEG signals, which activate during different mental states. At a simplistic level, this approach typically involves determining whether some averaged (mean) power value for a representation of the EEG signal differs to another, given a set of data samples from a defined time period. Such a “mean difference” test is performed by theprocessing system 109 for every signal representation. - Preferably, the
processing system 109 implements an Analysis of Variance (ANOVA) F ratio test to search for differences in activation, combined with a paired Student's T test. The T test is functionally equivalent to the one way ANOVA test for two groups, but also allows for a measure of direction of mean difference to be analysed (i.e. whether the mean value of amental state 1 is larger than the mean value for amental state 2, or vice versa). The general formula for the Student's T test is: - The “n” which makes the denominator in the lower half of the T equation is the number of time series recorded for a particular mental state which make up the means being contrasted in the numerator. (i.e. the number of overlapping or non-overlapping epochs recorded during an update.
- The subsequent t value is used in a variety of ways by the
processing system 109, including the rating of the feature space dimensions to determine the significance level of the many thousands of features that are typically analysed. Features may be weighted on a linear or non-linear scale, or in a binary fashion by removing those features which do not meet a certain level of significance. - The range of t values that will be generated from the many thousands of hypothesis tests during a signature update can be used to give an overall indication to the user of how far separated the detected mental states are during that update. The t value is an indication of that particular mean separation for the two actions, and the range of t values across all coordinates provides a metric for how well, on average, all of the coordinates separate.
- The above-mentioned techniques are termed univariate approaches as the
processing system 109 performs the analysis for each individual coordinate at a time, and make feature selections decisions based on those individual t test or ANOVA test results. Corrections may be made atstep 406 to adjust for the increased chance of probability error due to the use of the mass univariate approach. Statistical techniques suitable for this purpose include the following multiplicity correction methods: Bonferroni, False Discovery Rate and Dunn Sidack. - An alterative approach is for the
processing system 109 to analyse all coordinates together in a mass multivariate hypothesis test, which would account for any potential covariation between coordinates. Theprocessing system 109 can therefore employ such techniques as Discriminant Function Analysis and Multivariate analysis of variance (MANOVA), which not only provides a means to select feature space in a multivariate manner, but also allows the use of eigenvalues created during the analysis to actually classify unknown signal representations in a real-time environment. - At
step 408, theprocessing system 109 prepares for classifying incoming real-time data by weighting the coordinates so that those with the greatest significance in detecting a particular mental state are given precedence. This can be carried out by applying adaptive weight preparation, neural network training or statistical weightings. - The signatures stored in the memory of the
processing system 109 are updated or calibrated atstep 410. The updating process involves taking data samples, which is added to the evolving database. This data is elicited for the detection of a particular mental state. For example, to update a willed effort mental state, a user is prompted to focus on that willed effort and signal data samples are added to the database and used by theprocessing system 109 to modify the signature for that detection. When a signature exists, detections can provide feedback for updating the signatures that define that detection. For example, if a user wants to improve their signature for willing an object to be pushed away, the existing detection can be used to provide feedback as the signature is updated. In that scenario, the user sees the detection improving, which provides reinforcement to the updating process. - At
step 412, a supervised learning algorithm dynamically takes the update data fromstep 410 and combines it with the evolving database of recorded data samples to improve the signatures for the mental state that has been updated. Signatures may initially be empty or be prepared using historical data from other users which may have been combined to form a reference or universal starting signature. - At
step 414, the signature for the mental state that has been updated is made available for mental state classification (at step 400) as well as signature feedback rating atstep 416. As a user develops a signature for a given mental state, a rating is available in real-time which reflects how the mental state detection is progressing. Theapparatus 100 can therefore provide feedback to a user to enable them to observe the evolution of a signature over time. - The discussion above has focused on determination of the presence or absence of a particular mental state. However, it is also possible to determine the intensity of that particular mental state. The intensity can be determined by measuring the “distance” of the transformed signal from the user to a signature. The greater the distance, the lower the intensity. To calibrate the distance to the subjective intensity experienced by the user to an intensity scale, the user can be queried regarding the strength of the mental state. The resulting feedback from the user is applied to adjust the weights to calibrate the distance to the intensity scale.
- It will be appreciated from the foregoing that the
apparatus 100 advantageously enables the online creation of signatures in near real-time. The detection of a user's mental state and creation of a signature can be achieved in a few minutes, and then refined over time as the user's signature for that mental state is updated. This can be very important in interactive applications, where a short term result is important as well as incremental improvement over time. - It will also be appreciated from the foregoing that the
apparatus 100 advantageously enables the detection of a mental state having a pregenerated signature (whether predefined or created for the particular user) in real-time. Thus, the detection of the presence or absence of a user's particular mental state, or the intensity of that particular mental state, can be achieved in real-time. - Moreover, signatures can be created for mental states that need not be predefined. The
apparatus 100 can classify mental states that are recorded for, not just mental states that are predefined and elicited via pre-defined stimuli. - Each and every human brain is subtly different. While macroscopic structures such as the main gyri (ridges) and sulci (depressions) are common, it is only at the largest scale of morphology at which these generalizations can be made. The intricately detailed folding of the cortex is as individual as fingerprints. This variation in folding causes different parts of the brain to be near the skull in different individuals.
- For this reason the electrical impulses, when measured in combination on the scalp, differ between individuals. This means that the EEG recorded on the scalp must be interpreted differently from person to person. Historically, systems that aim to provide an individual with a means of control via EEG measurement have required extensive training, often of the system used and always by the user.
- The mental state detection system described herein can utilize a huge number of feature dimensions which cover many spatial areas, frequency ranges and other dimensions. In creating and updating a signature, the system ranks features by their ability to distinguish a particular mental state, thus highlighting those features that are better able to capture the brain's activity in a given mental state. The features selected by the user reflect characteristics of the electrical signals measured on the scalp that are able to distinguish a particular mental state, and therefore reflect how the signals in their particular cortex are manifested on the scalp. In short, the user's individual electrical signals that indicate a particular mental state have been identified and stored in a signature. This permits real-time mental state detection or generation within minutes, through algorithms which compensate for the individuality of EEG.
- It is to be understood that various modifications and/or additions may be made to the method and system for detecting and classifying a mental state without departing from the spirit or ambit of the present invention as defined in the claims appended hereto.
Claims (31)
1. A method of detecting and classifying mental states, comprising the steps of:
receiving bio-signals from one or more bio-signal detectors;
generating multiple different representations of each bio-signal;
determining the value of one or more features of the each bio-signal representation; and
comparing the feature values to one or more than one mental state signature, each mental state signature defining reference feature values indicative of a predetermined mental state.
2. The method according to claim 1 , wherein the step of generating multiple different representations of each bio-signal comprises the step of dividing the bio-signals into different epochs.
3. The method according to claim 2 , wherein the step of generating multiple different representations of each bio-signal further comprises the step of generating representations of the bio-signal epochs into one or more different domains.
4. The method according to claim 3 , wherein each bio-signal epoch is divided into one or more than one of different frequency, temporal and spatial domain representations.
5. The method according to claim 4 , wherein the different frequency domain representations are obtained by dividing each bio-signal epoch into distinguishable frequency bands.
6. The method according to claim 4 , wherein the different temporal domain representations are obtained by dividing each bio-signal epoch into a plurality of time segments.
7. The method according to claim 6 , wherein the time segments in each epoch are temporally overlapping.
8. The method according to claim 6 , wherein the time segments in each epoch do not temporally overlap.
9. The method according to claim 4 , wherein the different spatial domain representations are obtained by dividing each bio-signal epoch into a plurality of spatially distinguishable channels.
10. The method according to claim 9 , wherein each channel is derived from a different bio-signal detector.
11. The method according to claim 1 , wherein the step of determining the value of one or more features of the each bio-signal representation comprises determining values of features of individual bio-signal representations.
12. The method according to claim 11 , wherein one or more than one feature comprises signal power of one or more than one bio-signal representations.
13 The method according to claim 11 , wherein one or more than one feature comprises signal power of one or more than one spatially distinguishable channels.
14. The method according to claim 11 , wherein one or more than one feature comprises a change in signal power of one or more than one bio-signal representations.
15. The method according to claim 11 , wherein one or more than one feature comprises a change in signal power of one or more than one spatially distinguishable channels.
16. The method according to claim 1 , wherein the step of determining the value of one or more features of the each bio-signal representation comprises determining values of features between different bio-signal representations.
17. The method according to claim 16 , wherein at least coherence or correlation are detected between different bio-signal representations.
18. The method according to claim 17 , wherein one or more than one feature comprises the correlation or coherence between signal power in different spatially distinguishable channels.
19. The method according to claim 17 , wherein one or more than one feature comprises correlation or coherence between changes in signal power in different frequency bands.
20. The method according to claim 1 , wherein the step of determining the value of one or more features of the each bio-signal representation comprises applying one or more transforms to the different bio-signal representations.
21. The method according to claim 20 , wherein the one or more transforms comprises any one or more of a Fourier Transform, wavelet transform or other linear or non-linear mathematical transform.
22. The method according to claim 1 , wherein the step of comparing the feature values to one or more than one mental state signature comprises:
using a neural network to classify whether the feature values are indicative of the presence of a predefined mental state.
23. The method according to claim 1 , wherein the step of comparing the feature values to one or more than one mental state signature comprises:
performing a distance measure to measure the similarity between the feature values and the reference features values to classify whether the feature values are indicative of the presence of a predefined mental state.
24. The method according to claim 1 , wherein the mental state is an emotional state.
25. The method according to claim 1 , wherein the mental state results from mental focus on a task, image or other willed experience.
26. A method of creating a signature for use in a method of detecting and classifying mental states according to claim 1 , comprising the steps of:
eliciting a desired mental state from a user;
determining the features of the bio-signal representations that most significantly indicate the presence of the desired mental state by the user; and
generating the signature from a combination of those features.
27. A method according to claim 26 , wherein the step of determining the features of the bio-signal representations that most significantly indicate the presence of the desired mental state by the user comprises the step of:
performing any one or more of an ANOVA test, a T test, a Discriminant Function analysis, a MANOVA test, a Bonferroni analysis, False Discovery Rate analysis and Dunn Sidack analysis on the bio-signal representation features.
28. A method according to claim 26 , wherein the desired mental state is not predefined.
29. A method according to claim 26 , and further comprising the step of:
using feature values determined when the desired mental state is elicited from one or more users to update the signature for that mental state.
30. An apparatus for detecting and classifying mental states, comprising:
a processor and associated memory device for carrying out a method according to claim 1 .
31. A computer program product, tangibly stored on machine readable medium, the product comprising instructions operable to cause a processor to carry out a method according to claim 1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/531,238 US20070066914A1 (en) | 2005-09-12 | 2006-09-12 | Method and System for Detecting and Classifying Mental States |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/225,835 US7865235B2 (en) | 2005-09-12 | 2005-09-12 | Method and system for detecting and classifying the mental state of a subject |
US11/531,238 US20070066914A1 (en) | 2005-09-12 | 2006-09-12 | Method and System for Detecting and Classifying Mental States |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/225,835 Continuation-In-Part US7865235B2 (en) | 2005-09-12 | 2005-09-12 | Method and system for detecting and classifying the mental state of a subject |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070066914A1 true US20070066914A1 (en) | 2007-03-22 |
Family
ID=37856225
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/225,835 Expired - Fee Related US7865235B2 (en) | 2005-09-12 | 2005-09-12 | Method and system for detecting and classifying the mental state of a subject |
US11/531,238 Abandoned US20070066914A1 (en) | 2005-09-12 | 2006-09-12 | Method and System for Detecting and Classifying Mental States |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/225,835 Expired - Fee Related US7865235B2 (en) | 2005-09-12 | 2005-09-12 | Method and system for detecting and classifying the mental state of a subject |
Country Status (3)
Country | Link |
---|---|
US (2) | US7865235B2 (en) |
TW (1) | TW200727866A (en) |
WO (1) | WO2007030869A1 (en) |
Cited By (61)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060257834A1 (en) * | 2005-05-10 | 2006-11-16 | Lee Linda M | Quantitative EEG as an identifier of learning modality |
US20070055169A1 (en) * | 2005-09-02 | 2007-03-08 | Lee Michael J | Device and method for sensing electrical activity in tissue |
US20080177197A1 (en) * | 2007-01-22 | 2008-07-24 | Lee Koohyoung | Method and apparatus for quantitatively evaluating mental states based on brain wave signal processing system |
US20080214902A1 (en) * | 2007-03-02 | 2008-09-04 | Lee Hans C | Apparatus and Method for Objectively Determining Human Response to Media |
US20080218472A1 (en) * | 2007-03-05 | 2008-09-11 | Emotiv Systems Pty., Ltd. | Interface to convert mental states and facial expressions to application input |
US20080222670A1 (en) * | 2007-03-07 | 2008-09-11 | Lee Hans C | Method and system for using coherence of biological responses as a measure of performance of a media |
US20080221400A1 (en) * | 2007-03-08 | 2008-09-11 | Lee Hans C | Method and system for measuring and ranking an "engagement" response to audiovisual or interactive media, products, or activities using physiological signals |
US20080221472A1 (en) * | 2007-03-07 | 2008-09-11 | Lee Hans C | Method and system for measuring and ranking a positive or negative response to audiovisual or interactive media, products or activities using physiological signals |
US20080222671A1 (en) * | 2007-03-08 | 2008-09-11 | Lee Hans C | Method and system for rating media and events in media based on physiological data |
US20080221969A1 (en) * | 2007-03-07 | 2008-09-11 | Emsense Corporation | Method And System For Measuring And Ranking A "Thought" Response To Audiovisual Or Interactive Media, Products Or Activities Using Physiological Signals |
US20090069652A1 (en) * | 2007-09-07 | 2009-03-12 | Lee Hans C | Method and Apparatus for Sensing Blood Oxygen |
US20090070798A1 (en) * | 2007-03-02 | 2009-03-12 | Lee Hans C | System and Method for Detecting Viewer Attention to Media Delivery Devices |
US20090094627A1 (en) * | 2007-10-02 | 2009-04-09 | Lee Hans C | Providing Remote Access to Media, and Reaction and Survey Data From Viewers of the Media |
US20090112077A1 (en) * | 2004-01-08 | 2009-04-30 | Neurosky, Inc. | Contoured electrode |
US20090133047A1 (en) * | 2007-10-31 | 2009-05-21 | Lee Hans C | Systems and Methods Providing Distributed Collection and Centralized Processing of Physiological Responses from Viewers |
US20090150919A1 (en) * | 2007-11-30 | 2009-06-11 | Lee Michael J | Correlating Media Instance Information With Physiological Responses From Participating Subjects |
US20090156925A1 (en) * | 2004-01-08 | 2009-06-18 | Kyung-Soo Jin | Active dry sensor module for measurement of bioelectricity |
US20090214060A1 (en) * | 2008-02-13 | 2009-08-27 | Neurosky, Inc. | Audio headset with bio-signal sensors |
US20090253996A1 (en) * | 2007-03-02 | 2009-10-08 | Lee Michael J | Integrated Sensor Headset |
US20090281408A1 (en) * | 2008-05-06 | 2009-11-12 | Neurosky, Inc. | Dry Electrode Device and Method of Assembly |
US20090318826A1 (en) * | 2008-06-18 | 2009-12-24 | Green George H | Method and apparatus of neurological feedback systems to control physical objects for therapeutic and other reasons |
US20100016753A1 (en) * | 2008-07-18 | 2010-01-21 | Firlik Katrina S | Systems and Methods for Portable Neurofeedback |
US20100042011A1 (en) * | 2005-05-16 | 2010-02-18 | Doidge Mark S | Three-dimensional localization, display, recording, and analysis of electrical activity in the cerebral cortex |
US20100234752A1 (en) * | 2009-03-16 | 2010-09-16 | Neurosky, Inc. | EEG control of devices using sensory evoked potentials |
US20110040202A1 (en) * | 2009-03-16 | 2011-02-17 | Neurosky, Inc. | Sensory-evoked potential (sep) classification/detection in the time domain |
US20120022392A1 (en) * | 2010-07-22 | 2012-01-26 | Washington University In St. Louis | Correlating Frequency Signatures To Cognitive Processes |
US8347326B2 (en) | 2007-12-18 | 2013-01-01 | The Nielsen Company (US) | Identifying key media events and modeling causal relationships between key events and reported feelings |
US8666484B2 (en) * | 2011-11-25 | 2014-03-04 | Persyst Development Corporation | Method and system for displaying EEG data |
US20140275879A1 (en) * | 2013-03-15 | 2014-09-18 | Paul Stanley Addison | Systems and methods for determining respiration information based on independent component analysis |
US20150080753A1 (en) * | 2012-04-12 | 2015-03-19 | Canon Kabushiki Kaisha | Evaluation method, evaluation device, program, and recording medium |
US8989835B2 (en) | 2012-08-17 | 2015-03-24 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US9013264B2 (en) | 2011-03-12 | 2015-04-21 | Perceptive Devices, Llc | Multipurpose controller for electronic devices, facial expressions management and drowsiness detection |
US9292858B2 (en) | 2012-02-27 | 2016-03-22 | The Nielsen Company (Us), Llc | Data collection system for aggregating biologically based measures in asynchronous geographically distributed public environments |
US9320450B2 (en) | 2013-03-14 | 2016-04-26 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US9402554B2 (en) | 2011-09-23 | 2016-08-02 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US9451303B2 (en) | 2012-02-27 | 2016-09-20 | The Nielsen Company (Us), Llc | Method and system for gathering and computing an audience's neurologically-based reactions in a distributed framework involving remote storage and computing |
US9622702B2 (en) | 2014-04-03 | 2017-04-18 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US9675274B2 (en) | 2011-09-23 | 2017-06-13 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US9693709B2 (en) | 2011-09-23 | 2017-07-04 | Nellcot Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US9693736B2 (en) | 2011-11-30 | 2017-07-04 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information using historical distribution |
US9737266B2 (en) | 2011-09-23 | 2017-08-22 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US9814426B2 (en) | 2012-06-14 | 2017-11-14 | Medibotics Llc | Mobile wearable electromagnetic brain activity monitor |
US10130277B2 (en) | 2014-01-28 | 2018-11-20 | Medibotics Llc | Willpower glasses (TM)—a wearable food consumption monitor |
US10213157B2 (en) * | 2017-06-09 | 2019-02-26 | Bose Corporation | Active unipolar dry electrode open ear wireless headset and brain computer interface |
WO2020053849A1 (en) * | 2018-09-13 | 2020-03-19 | Quantalx Neuroscience Ltd | A reliable tool for evaluating brain health |
US10664050B2 (en) | 2018-09-21 | 2020-05-26 | Neurable Inc. | Human-computer interface using high-speed and accurate tracking of user interactions |
CN111616680A (en) * | 2020-05-26 | 2020-09-04 | 北京航空航天大学 | Automatic mental load identification method and system |
US20210196140A1 (en) * | 2019-12-27 | 2021-07-01 | Fuji Xerox Co., Ltd. | Information processing apparatus and non-transitory computer readable medium |
US11269414B2 (en) | 2017-08-23 | 2022-03-08 | Neurable Inc. | Brain-computer interface with high-speed eye tracking features |
US11266342B2 (en) | 2014-05-30 | 2022-03-08 | The Regents Of The University Of Michigan | Brain-computer interface for facilitating direct selection of multiple-choice answers and the identification of state changes |
US11269891B2 (en) * | 2014-08-21 | 2022-03-08 | Affectomatics Ltd. | Crowd-based scores for experiences from measurements of affective response |
US11273283B2 (en) | 2017-12-31 | 2022-03-15 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to enhance emotional response |
US11344256B2 (en) | 2017-02-21 | 2022-05-31 | Bose Corporation | Collecting biologically-relevant information using an earpiece |
US11364361B2 (en) | 2018-04-20 | 2022-06-21 | Neuroenhancement Lab, LLC | System and method for inducing sleep by transplanting mental states |
US11452839B2 (en) | 2018-09-14 | 2022-09-27 | Neuroenhancement Lab, LLC | System and method of improving sleep |
US11717686B2 (en) | 2017-12-04 | 2023-08-08 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to facilitate learning and performance |
US11723579B2 (en) | 2017-09-19 | 2023-08-15 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement |
US11786694B2 (en) | 2019-05-24 | 2023-10-17 | NeuroLight, Inc. | Device, method, and app for facilitating sleep |
US12001602B2 (en) | 2017-11-13 | 2024-06-04 | Neurable Inc. | Brain-computer interface with adaptations for high-speed, accurate, and intuitive user interactions |
US12053308B2 (en) | 2018-01-18 | 2024-08-06 | Neurable Inc. | Brain-computer interface with adaptations for high-speed, accurate, and intuitive user interactions |
US12138065B2 (en) | 2022-12-13 | 2024-11-12 | Quantalx Neuroscience Ltd | Dementia detection |
Families Citing this family (60)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB0510728D0 (en) * | 2005-05-26 | 2005-06-29 | Bell Fruit Group Ltd | Amusement apparatus |
US7902867B2 (en) * | 2006-04-03 | 2011-03-08 | Blaise Laurent Mouttet | Memristor crossbar neural interface |
CN101711388B (en) | 2007-03-29 | 2016-04-27 | 神经焦点公司 | The effect analysis of marketing and amusement |
US9886981B2 (en) * | 2007-05-01 | 2018-02-06 | The Nielsen Company (Us), Llc | Neuro-feedback based stimulus compression device |
JP5361868B2 (en) * | 2007-05-01 | 2013-12-04 | ニューロフォーカス・インコーポレーテッド | Neural information storage system |
US8392253B2 (en) * | 2007-05-16 | 2013-03-05 | The Nielsen Company (Us), Llc | Neuro-physiology and neuro-behavioral based stimulus targeting system |
JP2008301971A (en) * | 2007-06-06 | 2008-12-18 | Fujitsu Component Ltd | Brain wave detector |
US8494905B2 (en) * | 2007-06-06 | 2013-07-23 | The Nielsen Company (Us), Llc | Audience response analysis using simultaneous electroencephalography (EEG) and functional magnetic resonance imaging (fMRI) |
EP2170161B1 (en) | 2007-07-30 | 2018-12-05 | The Nielsen Company (US), LLC. | Neuro-response stimulus and stimulus attribute resonance estimator |
US8386313B2 (en) * | 2007-08-28 | 2013-02-26 | The Nielsen Company (Us), Llc | Stimulus placement system using subject neuro-response measurements |
US8635105B2 (en) * | 2007-08-28 | 2014-01-21 | The Nielsen Company (Us), Llc | Consumer experience portrayal effectiveness assessment system |
WO2009032691A1 (en) * | 2007-08-28 | 2009-03-12 | Neurofocus, Inc. | Consumer experience assessment system |
US8392255B2 (en) * | 2007-08-29 | 2013-03-05 | The Nielsen Company (Us), Llc | Content based selection and meta tagging of advertisement breaks |
US20090083129A1 (en) | 2007-09-20 | 2009-03-26 | Neurofocus, Inc. | Personalized content delivery using neuro-response priming data |
US8494610B2 (en) * | 2007-09-20 | 2013-07-23 | The Nielsen Company (Us), Llc | Analysis of marketing and entertainment effectiveness using magnetoencephalography |
US20090143695A1 (en) * | 2007-11-30 | 2009-06-04 | Palo Alto Research Center Incorporated | Brainwave-facilitated presenter feedback mechanism |
US20100250325A1 (en) | 2009-03-24 | 2010-09-30 | Neurofocus, Inc. | Neurological profiles for market matching and stimulus presentation |
US20110046502A1 (en) * | 2009-08-20 | 2011-02-24 | Neurofocus, Inc. | Distributed neuro-response data collection and analysis |
US8655437B2 (en) | 2009-08-21 | 2014-02-18 | The Nielsen Company (Us), Llc | Analysis of the mirror neuron system for evaluation of stimulus |
US10987015B2 (en) * | 2009-08-24 | 2021-04-27 | Nielsen Consumer Llc | Dry electrodes for electroencephalography |
US20110106750A1 (en) | 2009-10-29 | 2011-05-05 | Neurofocus, Inc. | Generating ratings predictions using neuro-response data |
US9560984B2 (en) | 2009-10-29 | 2017-02-07 | The Nielsen Company (Us), Llc | Analysis of controlled and automatic attention for introduction of stimulus material |
US8209224B2 (en) | 2009-10-29 | 2012-06-26 | The Nielsen Company (Us), Llc | Intracluster content management using neuro-response priming data |
US20130096408A1 (en) * | 2010-01-13 | 2013-04-18 | Regent Of The University Of Minnesota | Imaging epilepsy sources from electrophysiological measurements |
WO2011133548A2 (en) | 2010-04-19 | 2011-10-27 | Innerscope Research, Inc. | Short imagery task (sit) research method |
US8655428B2 (en) | 2010-05-12 | 2014-02-18 | The Nielsen Company (Us), Llc | Neuro-response data synchronization |
US11471091B2 (en) * | 2010-07-29 | 2022-10-18 | Kulangara Sivadas | Mind strength trainer |
US8392251B2 (en) | 2010-08-09 | 2013-03-05 | The Nielsen Company (Us), Llc | Location aware presentation of stimulus material |
US8392250B2 (en) | 2010-08-09 | 2013-03-05 | The Nielsen Company (Us), Llc | Neuro-response evaluated stimulus in virtual reality environments |
US8396744B2 (en) | 2010-08-25 | 2013-03-12 | The Nielsen Company (Us), Llc | Effective virtual reality environments for presentation of marketing materials |
US10244988B2 (en) * | 2010-12-16 | 2019-04-02 | Nokia Technologies Oy | Method, apparatus and computer program of using a bio-signal profile |
KR101208719B1 (en) * | 2011-01-07 | 2012-12-06 | 동명대학교산학협력단 | System for processing biological signal and portable instrumnet for processing biological signal |
EP2670304B1 (en) * | 2011-01-31 | 2020-04-01 | Koninklijke Philips N.V. | Automated spirogram analysis and interpretation |
US10893822B2 (en) | 2011-02-03 | 2021-01-19 | The Medical Research, Infrastructure and Health Services Fund of the Tel Aviv Medical Center | Method and system for use in monitoring neural activity in a subject's brain |
JP6109155B2 (en) * | 2011-05-06 | 2017-04-05 | ザ ジェネラル ホスピタル コーポレイション | System and method for tracking the state of the brain during anesthesia administration |
US11786132B2 (en) | 2011-05-06 | 2023-10-17 | The General Hospital Corporation | Systems and methods for predicting arousal to consciousness during general anesthesia and sedation |
WO2013019997A1 (en) | 2011-08-02 | 2013-02-07 | Emotiv Lifesciences Inc. | Methods for modeling neurological development and diagnosing a neurological impairment of a patient |
US9569986B2 (en) | 2012-02-27 | 2017-02-14 | The Nielsen Company (Us), Llc | System and method for gathering and analyzing biometric user feedback for use in social media and advertising applications |
JP6123167B2 (en) * | 2012-04-05 | 2017-05-10 | ソニー株式会社 | EEG analysis apparatus and EEG analysis program |
US20140024913A1 (en) * | 2012-07-18 | 2014-01-23 | Neurotopia, Inc. | Neurophysiologic performance measurement and training system |
US9015087B2 (en) * | 2012-10-09 | 2015-04-21 | At&T Intellectual Property I, L.P. | Methods, systems, and products for interfacing with neurological and biological networks |
US20150025352A1 (en) * | 2013-07-22 | 2015-01-22 | NorDocs Technologies Inc. | Method and device for determining brain and scalp state |
US10028703B2 (en) | 2013-07-30 | 2018-07-24 | Emotiv, Inc. | Wearable system for detecting and measuring biosignals |
WO2015057709A1 (en) * | 2013-10-14 | 2015-04-23 | Neurovigil, Inc. | Localized collection of biological signals, cursor control in speech-assistance interface based on biological electrical signals and arousal detection based on biological electrical signals |
US9993386B2 (en) | 2013-11-29 | 2018-06-12 | Louis G. RACETTE | Instrumentation absolute value differential amplifier circuit and applications |
CN103750844B (en) * | 2014-01-15 | 2015-07-29 | 杭州电子科技大学 | A kind of based on the phase locked personal identification method of brain electricity |
US9531998B1 (en) | 2015-07-02 | 2016-12-27 | Krush Technologies, Llc | Facial gesture recognition and video analysis tool |
US20160023116A1 (en) * | 2014-07-03 | 2016-01-28 | Spitfire Technologies, Llc | Electronically mediated reaction game |
US10108264B2 (en) | 2015-03-02 | 2018-10-23 | Emotiv, Inc. | System and method for embedded cognitive state metric system |
US9936250B2 (en) | 2015-05-19 | 2018-04-03 | The Nielsen Company (Us), Llc | Methods and apparatus to adjust content presented to an individual |
US10291977B2 (en) * | 2015-08-05 | 2019-05-14 | Emotiv Inc. | Method and system for collecting and processing bioelectrical and audio signals |
WO2017177128A1 (en) | 2016-04-08 | 2017-10-12 | The Trustees Of Columbia University In The City Of New York | Systems and methods for deep reinforcement learning using a brain-artificial intelligence interface |
WO2018058253A1 (en) | 2016-09-29 | 2018-04-05 | 9617094 Canada Inc. | Biosignal headphones |
DE102017002700A1 (en) * | 2017-03-21 | 2018-09-27 | Imat-Uve Gmbh | Method for assessing a product by at least one person and arrangement for carrying out this method |
EP3651038A1 (en) | 2018-11-12 | 2020-05-13 | Mastercard International Incorporated | Brain activity-based authentication |
WO2020100095A1 (en) * | 2018-11-18 | 2020-05-22 | Asociația Transylvanian Institute Of Neuroscience | Method, human machine interface, machine computing unit and computer programs to control at least one actuator to carry out at least one task |
US11553871B2 (en) | 2019-06-04 | 2023-01-17 | Lab NINE, Inc. | System and apparatus for non-invasive measurement of transcranial electrical signals, and method of calibrating and/or using same for various applications |
CN110338820A (en) * | 2019-06-13 | 2019-10-18 | 四川大学 | A kind of depression and schizophrenia recognition methods |
CN111091074B (en) * | 2019-12-02 | 2023-04-25 | 杭州电子科技大学 | Motor imagery electroencephalogram signal classification method of optimal region co-space mode |
CN111012340A (en) * | 2020-01-07 | 2020-04-17 | 南京邮电大学 | Emotion classification method based on multilayer perceptron |
Family Cites Families (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5724987A (en) * | 1991-09-26 | 1998-03-10 | Sam Technology, Inc. | Neurocognitive adaptive computer-aided training method and system |
US5601090A (en) * | 1994-07-12 | 1997-02-11 | Brain Functions Laboratory, Inc. | Method and apparatus for automatically determining somatic state |
US6349231B1 (en) * | 1994-01-12 | 2002-02-19 | Brain Functions Laboratory, Inc. | Method and apparatus for will determination and bio-signal control |
JP3310498B2 (en) * | 1994-09-02 | 2002-08-05 | 独立行政法人産業技術総合研究所 | Biological information analyzer and biological information analysis method |
US5740812A (en) * | 1996-01-25 | 1998-04-21 | Mindwaves, Ltd. | Apparatus for and method of providing brainwave biofeedback |
US6292688B1 (en) * | 1996-02-28 | 2001-09-18 | Advanced Neurotechnologies, Inc. | Method and apparatus for analyzing neurological response to emotion-inducing stimuli |
US5676138A (en) | 1996-03-15 | 1997-10-14 | Zawilinski; Kenneth Michael | Emotional response analyzer system with multimedia display |
US6097981A (en) * | 1997-04-30 | 2000-08-01 | Unique Logic And Technology, Inc. | Electroencephalograph based biofeedback system and method |
KR100281650B1 (en) * | 1997-11-13 | 2001-02-15 | 정선종 | EEG analysis method for discrimination of positive / negative emotional state |
US6609017B1 (en) * | 1998-08-07 | 2003-08-19 | California Institute Of Technology | Processed neural signals and methods for generating and using them |
US6422999B1 (en) * | 1999-05-13 | 2002-07-23 | Daniel A. Hill | Method of measuring consumer reaction |
KR100306295B1 (en) | 1999-07-24 | 2001-09-24 | 박병운 | Game machine using brain waves and method therefor |
EP1139240A3 (en) | 2000-03-28 | 2003-11-05 | Kenji Mimura | Design method and design evaluation method, and equipment thereof |
JP2001356849A (en) | 2000-05-08 | 2001-12-26 | Ken Tamada | Business model for human interface and hardware |
LU90582B1 (en) * | 2000-05-16 | 2001-11-19 | Europ Economic Community | System for identifying brain activity |
CA2349021C (en) | 2000-06-16 | 2010-03-30 | Bayer Corporation | System, method and biosensor apparatus for data communications with a personal data assistant |
JP2004535231A (en) * | 2001-06-07 | 2004-11-25 | ローレンス フェアウェル | Method and apparatus for brain fingerprint identification, measurement, evaluation, and analysis in brain function |
US20030032890A1 (en) * | 2001-07-12 | 2003-02-13 | Hazlett Richard L. | Continuous emotional response analysis with facial EMG |
WO2004037086A1 (en) | 2002-10-23 | 2004-05-06 | Daimlerchrysler Ag | Method for optimising and recording product attractiveness or product acceptance by observing cerebral activity |
US7546158B2 (en) * | 2003-06-05 | 2009-06-09 | The Regents Of The University Of California | Communication methods based on brain computer interfaces |
US7120486B2 (en) * | 2003-12-12 | 2006-10-10 | Washington University | Brain computer interface |
WO2006009129A1 (en) | 2004-07-16 | 2006-01-26 | Semiconductor Energy Laboratory Co., Ltd. | Biological signal processing unit, wireless memory, biological signal processing system, and control system of device to be controlled |
-
2005
- 2005-09-12 US US11/225,835 patent/US7865235B2/en not_active Expired - Fee Related
-
2006
- 2006-09-12 US US11/531,238 patent/US20070066914A1/en not_active Abandoned
- 2006-09-12 TW TW095133792A patent/TW200727866A/en unknown
- 2006-09-12 WO PCT/AU2006/001332 patent/WO2007030869A1/en active Application Filing
Cited By (109)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8301218B2 (en) | 2004-01-08 | 2012-10-30 | Neurosky, Inc. | Contoured electrode |
US20090156925A1 (en) * | 2004-01-08 | 2009-06-18 | Kyung-Soo Jin | Active dry sensor module for measurement of bioelectricity |
US20090112077A1 (en) * | 2004-01-08 | 2009-04-30 | Neurosky, Inc. | Contoured electrode |
US8290563B2 (en) | 2004-01-08 | 2012-10-16 | Neurosky, Inc. | Active dry sensor module for measurement of bioelectricity |
US20060257834A1 (en) * | 2005-05-10 | 2006-11-16 | Lee Linda M | Quantitative EEG as an identifier of learning modality |
US20100042011A1 (en) * | 2005-05-16 | 2010-02-18 | Doidge Mark S | Three-dimensional localization, display, recording, and analysis of electrical activity in the cerebral cortex |
US9179854B2 (en) | 2005-05-16 | 2015-11-10 | Mark S. Doidge | Three-dimensional localization, display, recording, and analysis of electrical activity in the cerebral cortex |
US10506941B2 (en) | 2005-08-09 | 2019-12-17 | The Nielsen Company (Us), Llc | Device and method for sensing electrical activity in tissue |
US11638547B2 (en) | 2005-08-09 | 2023-05-02 | Nielsen Consumer Llc | Device and method for sensing electrical activity in tissue |
US20070055169A1 (en) * | 2005-09-02 | 2007-03-08 | Lee Michael J | Device and method for sensing electrical activity in tissue |
US9351658B2 (en) | 2005-09-02 | 2016-05-31 | The Nielsen Company (Us), Llc | Device and method for sensing electrical activity in tissue |
US20080177197A1 (en) * | 2007-01-22 | 2008-07-24 | Lee Koohyoung | Method and apparatus for quantitatively evaluating mental states based on brain wave signal processing system |
US20090253996A1 (en) * | 2007-03-02 | 2009-10-08 | Lee Michael J | Integrated Sensor Headset |
US9215996B2 (en) | 2007-03-02 | 2015-12-22 | The Nielsen Company (Us), Llc | Apparatus and method for objectively determining human response to media |
US20090070798A1 (en) * | 2007-03-02 | 2009-03-12 | Lee Hans C | System and Method for Detecting Viewer Attention to Media Delivery Devices |
US20080214902A1 (en) * | 2007-03-02 | 2008-09-04 | Lee Hans C | Apparatus and Method for Objectively Determining Human Response to Media |
US20080218472A1 (en) * | 2007-03-05 | 2008-09-11 | Emotiv Systems Pty., Ltd. | Interface to convert mental states and facial expressions to application input |
US8230457B2 (en) | 2007-03-07 | 2012-07-24 | The Nielsen Company (Us), Llc. | Method and system for using coherence of biological responses as a measure of performance of a media |
US20080221969A1 (en) * | 2007-03-07 | 2008-09-11 | Emsense Corporation | Method And System For Measuring And Ranking A "Thought" Response To Audiovisual Or Interactive Media, Products Or Activities Using Physiological Signals |
US20080221472A1 (en) * | 2007-03-07 | 2008-09-11 | Lee Hans C | Method and system for measuring and ranking a positive or negative response to audiovisual or interactive media, products or activities using physiological signals |
US8473044B2 (en) | 2007-03-07 | 2013-06-25 | The Nielsen Company (Us), Llc | Method and system for measuring and ranking a positive or negative response to audiovisual or interactive media, products or activities using physiological signals |
US20080222670A1 (en) * | 2007-03-07 | 2008-09-11 | Lee Hans C | Method and system for using coherence of biological responses as a measure of performance of a media |
US8973022B2 (en) | 2007-03-07 | 2015-03-03 | The Nielsen Company (Us), Llc | Method and system for using coherence of biological responses as a measure of performance of a media |
US20080222671A1 (en) * | 2007-03-08 | 2008-09-11 | Lee Hans C | Method and system for rating media and events in media based on physiological data |
US20080221400A1 (en) * | 2007-03-08 | 2008-09-11 | Lee Hans C | Method and system for measuring and ranking an "engagement" response to audiovisual or interactive media, products, or activities using physiological signals |
US8764652B2 (en) * | 2007-03-08 | 2014-07-01 | The Nielson Company (US), LLC. | Method and system for measuring and ranking an “engagement” response to audiovisual or interactive media, products, or activities using physiological signals |
US8782681B2 (en) | 2007-03-08 | 2014-07-15 | The Nielsen Company (Us), Llc | Method and system for rating media and events in media based on physiological data |
US8376952B2 (en) | 2007-09-07 | 2013-02-19 | The Nielsen Company (Us), Llc. | Method and apparatus for sensing blood oxygen |
US20090069652A1 (en) * | 2007-09-07 | 2009-03-12 | Lee Hans C | Method and Apparatus for Sensing Blood Oxygen |
US9571877B2 (en) | 2007-10-02 | 2017-02-14 | The Nielsen Company (Us), Llc | Systems and methods to determine media effectiveness |
US8151292B2 (en) | 2007-10-02 | 2012-04-03 | Emsense Corporation | System for remote access to media, and reaction and survey data from viewers of the media |
US9021515B2 (en) | 2007-10-02 | 2015-04-28 | The Nielsen Company (Us), Llc | Systems and methods to determine media effectiveness |
US9894399B2 (en) | 2007-10-02 | 2018-02-13 | The Nielsen Company (Us), Llc | Systems and methods to determine media effectiveness |
US20090094628A1 (en) * | 2007-10-02 | 2009-04-09 | Lee Hans C | System Providing Actionable Insights Based on Physiological Responses From Viewers of Media |
US20090094286A1 (en) * | 2007-10-02 | 2009-04-09 | Lee Hans C | System for Remote Access to Media, and Reaction and Survey Data From Viewers of the Media |
US20090094629A1 (en) * | 2007-10-02 | 2009-04-09 | Lee Hans C | Providing Actionable Insights Based on Physiological Responses From Viewers of Media |
US8327395B2 (en) | 2007-10-02 | 2012-12-04 | The Nielsen Company (Us), Llc | System providing actionable insights based on physiological responses from viewers of media |
US20090094627A1 (en) * | 2007-10-02 | 2009-04-09 | Lee Hans C | Providing Remote Access to Media, and Reaction and Survey Data From Viewers of the Media |
US8332883B2 (en) | 2007-10-02 | 2012-12-11 | The Nielsen Company (Us), Llc | Providing actionable insights based on physiological responses from viewers of media |
US20090133047A1 (en) * | 2007-10-31 | 2009-05-21 | Lee Hans C | Systems and Methods Providing Distributed Collection and Centralized Processing of Physiological Responses from Viewers |
US11250447B2 (en) | 2007-10-31 | 2022-02-15 | Nielsen Consumer Llc | Systems and methods providing en mass collection and centralized processing of physiological responses from viewers |
US10580018B2 (en) | 2007-10-31 | 2020-03-03 | The Nielsen Company (Us), Llc | Systems and methods providing EN mass collection and centralized processing of physiological responses from viewers |
US9521960B2 (en) | 2007-10-31 | 2016-12-20 | The Nielsen Company (Us), Llc | Systems and methods providing en mass collection and centralized processing of physiological responses from viewers |
US20090150919A1 (en) * | 2007-11-30 | 2009-06-11 | Lee Michael J | Correlating Media Instance Information With Physiological Responses From Participating Subjects |
US8793715B1 (en) | 2007-12-18 | 2014-07-29 | The Nielsen Company (Us), Llc | Identifying key media events and modeling causal relationships between key events and reported feelings |
US8347326B2 (en) | 2007-12-18 | 2013-01-01 | The Nielsen Company (US) | Identifying key media events and modeling causal relationships between key events and reported feelings |
US20090214060A1 (en) * | 2008-02-13 | 2009-08-27 | Neurosky, Inc. | Audio headset with bio-signal sensors |
US8271075B2 (en) | 2008-02-13 | 2012-09-18 | Neurosky, Inc. | Audio headset with bio-signal sensors |
US8170637B2 (en) | 2008-05-06 | 2012-05-01 | Neurosky, Inc. | Dry electrode device and method of assembly |
US20090281408A1 (en) * | 2008-05-06 | 2009-11-12 | Neurosky, Inc. | Dry Electrode Device and Method of Assembly |
US20090318826A1 (en) * | 2008-06-18 | 2009-12-24 | Green George H | Method and apparatus of neurological feedback systems to control physical objects for therapeutic and other reasons |
US8326408B2 (en) | 2008-06-18 | 2012-12-04 | Green George H | Method and apparatus of neurological feedback systems to control physical objects for therapeutic and other reasons |
US20100016753A1 (en) * | 2008-07-18 | 2010-01-21 | Firlik Katrina S | Systems and Methods for Portable Neurofeedback |
US8155736B2 (en) | 2009-03-16 | 2012-04-10 | Neurosky, Inc. | EEG control of devices using sensory evoked potentials |
US20110040202A1 (en) * | 2009-03-16 | 2011-02-17 | Neurosky, Inc. | Sensory-evoked potential (sep) classification/detection in the time domain |
US8391966B2 (en) | 2009-03-16 | 2013-03-05 | Neurosky, Inc. | Sensory-evoked potential (SEP) classification/detection in the time domain |
US20100234752A1 (en) * | 2009-03-16 | 2010-09-16 | Neurosky, Inc. | EEG control of devices using sensory evoked potentials |
US20120022392A1 (en) * | 2010-07-22 | 2012-01-26 | Washington University In St. Louis | Correlating Frequency Signatures To Cognitive Processes |
US9013264B2 (en) | 2011-03-12 | 2015-04-21 | Perceptive Devices, Llc | Multipurpose controller for electronic devices, facial expressions management and drowsiness detection |
US9402554B2 (en) | 2011-09-23 | 2016-08-02 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US9737266B2 (en) | 2011-09-23 | 2017-08-22 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US9675274B2 (en) | 2011-09-23 | 2017-06-13 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US9693709B2 (en) | 2011-09-23 | 2017-07-04 | Nellcot Puritan Bennett Ireland | Systems and methods for determining respiration information from a photoplethysmograph |
US8666484B2 (en) * | 2011-11-25 | 2014-03-04 | Persyst Development Corporation | Method and system for displaying EEG data |
US9693736B2 (en) | 2011-11-30 | 2017-07-04 | Nellcor Puritan Bennett Ireland | Systems and methods for determining respiration information using historical distribution |
US9451303B2 (en) | 2012-02-27 | 2016-09-20 | The Nielsen Company (Us), Llc | Method and system for gathering and computing an audience's neurologically-based reactions in a distributed framework involving remote storage and computing |
US9292858B2 (en) | 2012-02-27 | 2016-03-22 | The Nielsen Company (Us), Llc | Data collection system for aggregating biologically based measures in asynchronous geographically distributed public environments |
US20150080753A1 (en) * | 2012-04-12 | 2015-03-19 | Canon Kabushiki Kaisha | Evaluation method, evaluation device, program, and recording medium |
US9814426B2 (en) | 2012-06-14 | 2017-11-14 | Medibotics Llc | Mobile wearable electromagnetic brain activity monitor |
US9060671B2 (en) | 2012-08-17 | 2015-06-23 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US9215978B2 (en) | 2012-08-17 | 2015-12-22 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US8989835B2 (en) | 2012-08-17 | 2015-03-24 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US10842403B2 (en) | 2012-08-17 | 2020-11-24 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US9907482B2 (en) | 2012-08-17 | 2018-03-06 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US10779745B2 (en) | 2012-08-17 | 2020-09-22 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US11980469B2 (en) | 2012-08-17 | 2024-05-14 | Nielsen Company | Systems and methods to gather and analyze electroencephalographic data |
US9320450B2 (en) | 2013-03-14 | 2016-04-26 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US9668694B2 (en) | 2013-03-14 | 2017-06-06 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US11076807B2 (en) | 2013-03-14 | 2021-08-03 | Nielsen Consumer Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US20140275879A1 (en) * | 2013-03-15 | 2014-09-18 | Paul Stanley Addison | Systems and methods for determining respiration information based on independent component analysis |
US10130277B2 (en) | 2014-01-28 | 2018-11-20 | Medibotics Llc | Willpower glasses (TM)—a wearable food consumption monitor |
US11141108B2 (en) | 2014-04-03 | 2021-10-12 | Nielsen Consumer Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US9622702B2 (en) | 2014-04-03 | 2017-04-18 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US9622703B2 (en) | 2014-04-03 | 2017-04-18 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
US12121360B2 (en) | 2014-05-30 | 2024-10-22 | The Regents Of The University Of Michigan | Brain-computer interface for facilitating direct selection of multiple-choice answers and the identification of state changes |
US11266342B2 (en) | 2014-05-30 | 2022-03-08 | The Regents Of The University Of Michigan | Brain-computer interface for facilitating direct selection of multiple-choice answers and the identification of state changes |
US11269891B2 (en) * | 2014-08-21 | 2022-03-08 | Affectomatics Ltd. | Crowd-based scores for experiences from measurements of affective response |
US11344256B2 (en) | 2017-02-21 | 2022-05-31 | Bose Corporation | Collecting biologically-relevant information using an earpiece |
US10213157B2 (en) * | 2017-06-09 | 2019-02-26 | Bose Corporation | Active unipolar dry electrode open ear wireless headset and brain computer interface |
US11269414B2 (en) | 2017-08-23 | 2022-03-08 | Neurable Inc. | Brain-computer interface with high-speed eye tracking features |
US11972049B2 (en) | 2017-08-23 | 2024-04-30 | Neurable Inc. | Brain-computer interface with high-speed eye tracking features |
US11723579B2 (en) | 2017-09-19 | 2023-08-15 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement |
US12001602B2 (en) | 2017-11-13 | 2024-06-04 | Neurable Inc. | Brain-computer interface with adaptations for high-speed, accurate, and intuitive user interactions |
US11717686B2 (en) | 2017-12-04 | 2023-08-08 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to facilitate learning and performance |
US11478603B2 (en) | 2017-12-31 | 2022-10-25 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to enhance emotional response |
US11318277B2 (en) | 2017-12-31 | 2022-05-03 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to enhance emotional response |
US11273283B2 (en) | 2017-12-31 | 2022-03-15 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to enhance emotional response |
US12053308B2 (en) | 2018-01-18 | 2024-08-06 | Neurable Inc. | Brain-computer interface with adaptations for high-speed, accurate, and intuitive user interactions |
US11364361B2 (en) | 2018-04-20 | 2022-06-21 | Neuroenhancement Lab, LLC | System and method for inducing sleep by transplanting mental states |
WO2020053849A1 (en) * | 2018-09-13 | 2020-03-19 | Quantalx Neuroscience Ltd | A reliable tool for evaluating brain health |
US11452839B2 (en) | 2018-09-14 | 2022-09-27 | Neuroenhancement Lab, LLC | System and method of improving sleep |
US10664050B2 (en) | 2018-09-21 | 2020-05-26 | Neurable Inc. | Human-computer interface using high-speed and accurate tracking of user interactions |
US11366517B2 (en) | 2018-09-21 | 2022-06-21 | Neurable Inc. | Human-computer interface using high-speed and accurate tracking of user interactions |
US11786694B2 (en) | 2019-05-24 | 2023-10-17 | NeuroLight, Inc. | Device, method, and app for facilitating sleep |
JP2021106669A (en) * | 2019-12-27 | 2021-07-29 | 富士フイルムビジネスイノベーション株式会社 | Information processing device, and information processing program |
US20210196140A1 (en) * | 2019-12-27 | 2021-07-01 | Fuji Xerox Co., Ltd. | Information processing apparatus and non-transitory computer readable medium |
JP7296121B2 (en) | 2019-12-27 | 2023-06-22 | 株式会社Agama-X | Information processing device and information processing program |
CN111616680A (en) * | 2020-05-26 | 2020-09-04 | 北京航空航天大学 | Automatic mental load identification method and system |
US12138065B2 (en) | 2022-12-13 | 2024-11-12 | Quantalx Neuroscience Ltd | Dementia detection |
Also Published As
Publication number | Publication date |
---|---|
US20070060831A1 (en) | 2007-03-15 |
US7865235B2 (en) | 2011-01-04 |
WO2007030869A1 (en) | 2007-03-22 |
TW200727866A (en) | 2007-08-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20070066914A1 (en) | Method and System for Detecting and Classifying Mental States | |
US20070173733A1 (en) | Detection of and Interaction Using Mental States | |
Hsu et al. | EEG-based motor imagery analysis using weighted wavelet transform features | |
US20070179396A1 (en) | Method and System for Detecting and Classifying Facial Muscle Movements | |
JP7373555B2 (en) | Quantification of motor function using EEG signals | |
JP2009542351A (en) | Analysis of brain patterns using temporal scales | |
JP2008517636A (en) | Biopotential waveform data combination analysis and classification device | |
Chiang | Ecg-based mental stress assessment using fuzzy computing and associative petri net | |
CN109715049A (en) | For the multi-modal physiological stimulation of traumatic brain injury and the agreement and signature of assessment | |
Koçer | Classification of EMG signals using neuro-fuzzy system and diagnosis of neuromuscular diseases | |
Sharma et al. | Recent trends in EEG based Motor Imagery Signal Analysis and Recognition: A comprehensive review. | |
Ahamad | System architecture for brain-computer interface based on machine learning and internet of things | |
Qaisar et al. | Advances in Non-Invasive Biomedical Signal Sensing and Processing with Machine Learning | |
Chakraborty et al. | A survey on Internet-of-Thing applications using electroencephalogram | |
Attallah | Multi-tasks biometric system for personal identification | |
Saini et al. | Discriminatory features based on wavelet energy for effective analysis of electroencephalogram during mental tasks | |
Lima et al. | Heart rate variability and electrodermal activity biosignal processing: predicting the autonomous nervous system response in mental stress | |
Nouri et al. | A new approach to feature extraction in MI-based BCI systems | |
Heaton et al. | Systems design for EEG signal classification of sensorimotor activity using machine learning | |
Apicella et al. | Metrological foundations of emotional valence measurement through an EEG-based system | |
Fouad et al. | Attempts towards the first brain-computer interface system in INAYA Medical College | |
KR100327117B1 (en) | Apparatus and method for making a judgement on positive and negative intention by using the EEG | |
Kang et al. | A sleep stage classification method using deep learning by extracting the characteristics of frequency domain from a single EEG channel | |
KR102452201B1 (en) | Automatic noise signal section detection method and apparatus | |
Mehla | Representation and Analysis of Biomedical Signals using Fourier Decomposition Method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: EMOTIV SYSTEMS PTY LTD., AUSTRALIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LE, TAN THI THAI;DO, NAM HOAI;DELLA TORRE, MARCO KENNETH;AND OTHERS;REEL/FRAME:018594/0014;SIGNING DATES FROM 20061129 TO 20061201 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |