[Eeglablist] How to recognize successful ICA (number of data points issue)

Stefan Debener s.debener at uke.uni-hamburg.de
Thu Oct 26 03:35:49 PDT 2006


Hi Jim,

in my experience there is no perfect rule of thumb, since the size of 
the data matrix you feed into ICA is only one of several factors 
influencing the quality of a solution. Here are my thoughts about what 
matters as well:

1. Number of training data 1: Consider for instance the case of a short 
and a long recording of, let's say, 128 channels. The long recording 
will return a more reliable and cleaner (more dipolar IC maps) solution 
only if the quality of the data remains roughly the same, and only if 
there is no 'new' stuff coming into the data. So more is only better if 
it's more of the same...see below for an idea how to test this.

2. Number of training data 2: In my personal opinion, the pure amount of 
training data, that is, size(data,2) of the data you feed into ICA, 
needs to be corrected for data points carrying information. Let's assume 
that EEG picks up activity up to 100 Hz, and you recorded your data with 
a 100 Hz analog low pass filter and a 500 Hz sampling rate. In this case 
you should do the rule-of-thumb calculation for a sampling rate of 200 
Hz, because, in the frequency domain, the higher frequencies contain 
zeros only. So it all depends on your filter settings (and your personal 
opinion about the frequency range being relevant for EEG ). Note also 
that different EEG recording systems on the market enfore a different 
oversampling factor, possibly due to a different quality/slope of analog 
low pass filters, to avoid aliasing....

3. Stationary data: It seems to me that it's sometimes much easier to 
get a decent decomposition with a higher high pass filter. So, if you 
consider the identical data with the identical number of data points, 
you may get a much better solution for high pass filtered data. The 
reason is that ICA really likes stationarity, and, the more dimensions 
you have to spend, the more relevant this may be.

4. Bad channels. I don't know for sure, but a few bad channels may 
easily blow up the whole decomposition, and prior bad channel 
replacement always means information loss. That's why we try to collect 
our data without any bad channels...

5. Adequate spatial sampling. It's really hard to tell how many channels 
are necessary to obtain the best-possible decomposition. My experience 
with 32, 64, 68 and 128 channels is that 128 not necessarily returns 
more reliable ICs. I am not aware of a single paper investigating the 
spatial sampling question for ICA. I used to be happy with 128 channels, 
but I currently get even better results with equidistant 68 channel 
recordings, covering a much larger part of the head sphere than the 
usual 10-20 system (the system is called intra-cerebral cap, extremely 
useful, and available from www.easycap.de). So, the pure number of 
channels seems to be one thing, and the covered head space another one. 
A larger inter-electrode distance may be just fine if you cover a larger 
part of the sphere as well. The underlying problem here of course is 
that we have no clue how many different sources contribute to the EEG. 
Modelling your data with less dimensions than 'true' sources is 
certainly no good (underdetermined case), but modelling more dimensions 
may not be good either. If one considers every single channel to have 
it's own noise term, the summed channel noise activity from spatially 
oversampled data may actually be responsible for reducing the quality of 
the decompositions. Just a guess...

6. Modeling the neurocognitive process of interest. Another issue is 
whether you should run ICA for continuous or for (concatenated) epoched 
data to obtain components reflecting the process(es) you are interested 
in. If you believe in the standard ERP data model (brain response is 
summed to, and independent from, ongoing activity), you would certainly 
like to tailor your data towards your ERP interval. If you on the other 
hand believe that ongoing EEG activity (or 99% of your recorded data) is 
not just noise, you may like the idea of decomposing continuous data (or 
longer epochs, allowing preparatory brain activity to be modelled!). 
Even if the ERP model was 'true', the decomposition would be less 
optimal for artifact ICs reflecting processes being continuoulsy on. ICs 
reflecting ECG for instance, an ongoing artefact, are much easier to 
obtain from continuous data. So if your focus is on regional EEG 
asymmetries in the low frequency domain (<20 Hz), you would certainly 
like to remove this asymmetric artifactual activity before analysing 
your (short epoched) data. So it all depends on what you are looking 
for....I started with decomposing short epoched data, but there may be 
neurocognitive processes being more or less continuously 'on', like 
performance monitoring, which I found easier to be separated from other 
stuff by decomposing continuous data (Debener et al., 2005, J Neurosci). 
Assessing the spatio-temporal overlap of ICs reflecting major portions 
of novelty-P3 and P3b ERP components on the other hand was better with 
decomposing epoched data (Debener et al., 2005, Cog Brain Res). It would 
be actually nice to systematically investigate this issue....

7. Moving sources. ICA likes your electrodes not to move relative to the 
signal sources. That's why ICA of MEG data may sometimes be a bit of a 
problem (head movement). Same holds true for slipping EEG caps, or when 
subjects scratched at electrodes during the recording and, as a result, 
the cap is a bit shifted. Or assume you have a long recording and the 
subjects walks to the loo in a break. More likely than not that the cap 
is shifted a bit. Flipping sensors of geodesic sensor nets is no good 
either, of course, or moving a head with the cap resting on a pillow. 
Luckily its rather unlikely that brain sources move...but there can be 
artefacts that can be modelled as a moving source! This holds true in 
particular for the BCG artifact, present in inside MRI scanner EEG 
recordings. This artifact contributes a moving, rotating and polarity 
reversing signal to the EEG (see a movie on my homepage, www.debener.de, 
illustrating this). If ICA is applied to remove the BCG, much of the EEG 
signal is removed as well (Debener et al., in press, Neuroimage), which 
is a nice example that ICA is not a perfect wizard dealing with any 
problem. But if applied correctly, ICA can help a lot in analysing 
inside scanner recorded EEG data (Debener et al., 2006, December issue 
of Trends in Cognitive Sciences).


So what is a good way to determine the quality of a solution, taking 
into account issues 1-7? Our lab routine is to look at dipolar 
components only, with dipolarity being determined by the residual 
variance of the dipole source modeling tool in EEGLAB, dipfit. With 
reasonable spatial (issues 4 and 5) and temporal (other issues) 
sampling, more dipolar components seem to implicate a better 
decomposition. 'Better' here refers primarily to the reproducability of 
the decomposition. If you have enough training data AND the quality of 
your data is OK, the resulting dipolar components will be very robust. 
That is, if you repeat your decomposition, the weights of your dipolar 
components will be correlated by r>.98. When analysing a new experiment 
we always run ICA repeatedly for about 1-3 randomly selected subjects 
and determine the reproducability of the model, before applying the same 
model to all subjects. It takes some time, but we found that it's worth 
the effort.

Sorry that there is no quick answer, but the issue is rather 
complicated, and I am not at all claiming to have it fully understood by 
myself. Please note also that while my comments are based on our 
experience, I have not systematically tested all of them. Where 
appropriate I have cited my own papers, but others have published very 
informative papers as well, and helpful Matlab code! There are promising 
open source tools available which may help to identify good, stable 
components (ICASSO), or aim to estimate the number of sources in the 
data before running ICA (GIFT). And finally, the new STUDY functionality 
in EEGLAB could also be used to determine the quality of a decomposition....

Best,
Stefan








Jim Kroger wrote:
> In the tutorial, as well as in discussion here, it's been pointed out that 
> successful ICA requires sufficient data. The tutorial discusses that the 
> factor: (number of data points)/(the number of channels)^2 should be >= 30 
> for 32 channels, and greater for more channels. We often find that 15 
> minute segments that fit well into EEGLAB/Matlab (given our RAM constraints 
> and 128 channels, 1024 sampling) provide us with a factor of around 80 
> before epoching, less after.  I suspect this is borderline, but don't know 
> how to judge whether the resulting components are "good." What would "good" 
> and "bad" components be like? Any commentary would be appreciated.
>
> Jim
>
> _______________________________________________
> eeglablist mailing list eeglablist at sccn.ucsd.edu
> Eeglablist page: http://sccn.ucsd.edu/eeglab/eeglabmail.html
> To unsubscribe, send an empty email to eeglablist-unsubscribe at sccn.ucsd.edu
>
>
>   






More information about the eeglablist mailing list