Introduction

Neurofeedback is a well-investigated treatment for ADHD and epilepsy, especially when restricted to standard protocols such as theta/beta (TBR), slow cortical potentials (SCP) and sensori-motor rhythm (SMR) neurofeedback (Arns et al. 2014). This has become evident from several meta-analyses (Arns et al. 2009; Micoulaud-Franchi et al. 2014), including a critical meta-analysis from the European ADHD Guidelines Group (EAGG) that also conducted a sensitivity analysis focused on so called ‘blinded’ ratings and hence focused on teacher reports only (Cortese et al. 2016). This latter meta-analysis did not find an overall effect of neurofeedback on teacher rated ADHD symptoms, but only when restricting the analysis to the above mentioned ‘standard protocols’ (Cortese et al. 2016). In addition to the meta-analytical support above and only focusing on randomized controlled trials (RCT’s) for the treatment of ADHD there is evidence for clinical efficacy of standard neurofeedback protocols from three large multicenter RCT’s (Gevensleben et al. 2009; Steiner et al. 2014; Strehl et al. 2017); non-superiority of methylphenidate relative to neurofeedback (Duric et al. 2012; Meisel et al. 2013) and sustained effects of neurofeedback from 3 to 9 months follow-up (Van Doren et al. 2018; Heinrich et al. 2004; Strehl et al. 2006; Leins et al. 2007; Gevensleben et al. 2010; Duric et al. 2012; Li et al. 2013; Meisel et al. 2013; Christiansen et al. 2014; Steiner et al. 2014; Bink et al. 2016) to 2 years follow-up (Gani et al. 2008). In line with the guidelines for rating evidence developed by the APA, ‘standard’ neurofeedback protocols have been considered to be ‘Efficacious and Specific, Level V’ in the treatment of ADHD (AAPB guidelines: Arns et al. 2016).

There is also a lineage of research and support for the use of SMR training, and to some degree SCP, in the treatment of seizure disorders (Sterman and Egner 2006; Tan et al. 2009). In addition there is also growing evidence to support standard protocols (1–2 channel) applied to autism, sleep, head injury and learning disorders (For review see: Tan et al. 2016), albeit to date the strongest evidence for efficacy exists for ADHD. Controlled studies have provided some additional empirical support for the use of two channel coherence training for head injury, autism and learning disabilities (Coben and Myers 2010; Coben et al. 2014, 2015; Thornton and Carmody 2005).

This notion of various types of neurofeedback protocols as well as various neurofeedback implementations is an important topic causing a lot of controversy in the field of psychiatry and neuromodulation. To most people, neurofeedback is regarded as the unitary phenomenon of neurofeedback, implying that all neurofeedback implementations are ‘thought’ to be the same, and hence combined into one meta-analysis or review. However, neurofeedback should be regarded as an umbrella term, incorporating an almost infinite number of possibilities for protocols and implementations, much similar to the umbrella term ‘medication’ which includes many different classes of medications (i.e. antidepressants, antipsychotics, antibiotics, pain-killers etc.) as well as implementations (e.g. oral dispension, intra-venously, intra-muscular etc.). The fact that it is now known that some specific neurofeedback protocols e.g. posterior alpha training in ADHD, (Nall 1973) and implementations e.g. entertainment-like playstation feedback (Arnold et al. 2013; DeBeus and Kaiser 2011) do not work in the treatment of ADHD, actually implies ‘specificity’ for some neurofeedback protocols and further supports the notion to look in more detail into the specifics of different neurofeedback protocols and implementations. For further discussions on appropriate neurofeedback implementations e.g. learning theory and conditioning, the interested reader is also referred to Sherlin et al. (2011), Arns et al. (2014) and Strehl (2014).

Medical devices in the US are regulated by the Federal Drug Administration (FDA), who can grant FDA marketing approval to specific devices. This procedure consists of companies providing their own data to the FDA for inspection and when FDA marketing approval is granted, such a company can use that claim in their marketing and selling of the device. Unfortunately, for biofeedback and neurofeedback there are no FDA marketing approvals other than the ‘exempt’ category of biofeedback for relaxation (under neurological devices, 882.5050). On the other hand, FDA marketing approval does not reflect the actual level of evidence for a given approach as accepted and endorsed in fields such as psychology or psychiatry (i.e. EMDR, mindfulness etc. are not FDA approved, yet widely used) and some FDA marketing approvals are not in line with actual evidence levels since only data provided by the manufacturer are inspected (e.g. the example of the theta/beta ratio as a ‘diagnostic test’ in ADHD, also see Arns et al. 2016). Therefore, officially, no clinical claims can be made associated with neurofeedback devices, although the scientific evidence for ‘standard neurofeedback protocols’ are more widely accepted these days as explained above. Yet, many manufacturers and clinicians are marketing ‘superior’ neurofeedback approaches. Several examples of this are Z-score 19 channel neurofeedback, 3-D LORETA neurofeedback (with or without Z-scores), neuroptimal and infra-slow neurofeedback. Proponents of these latter approaches often claim that no research is needed to support the efficacy of their equipment, since it is only intended for non-clinical use (neuroptimal) and thereby fall outside the scope of the FDA. However, the former approaches take a different approach, where they imply that the approach is backed by a wealth of scientific studies. Proponents of these approaches (19 channel Z-score and LORETA training) have suggested advantages over more traditional forms of neurofeedback. It is all too common for them to suggest that such full cap training reduces the number of sessions required for effective intervention (Krigbaum and Wigton 2015; Simkin et al. 2014). In the extreme, some allude to it’s superiority and suggest that it is the best and most “scientific” neurofeedback option available. For example, “…Deep brain (LORETA) neurofeedback is the most flexible and precise brain training tool available. Better imaging means the training is more specific to your goals, and fewer sessions are required to see results. That is why the vast majority of experienced clinicians (10 years plus) use this method of neurofeedback in their clinics… LORETA Z-Score neurofeedback is FDA approved, and meets the highest criteria for evidence-based neurofeedback practice. It has been cross-validated by both PET and MRI, is used by Universities, Medical Centers, the US military, Research Institutes, and in over 650 neurofeedback studies...” (http://www.brainworksneurotherapy.com/deep-brain-neurofeedback). Such claims sound really promising and exciting and thus were the primary aim to undertake this systematic review, to investigate the scientific evidence base behind these claims.

In line with what was stated earlier, it should be noted that LORETA Z-score neurofeedback is not “FDA approved”, other then maybe registered at the FDA under the exempt category.

Below we will summarize these approaches in a bit more detail to familiarize the reader with these methods.

Z-Score Neurofeedback (ZNFB)

One of the initial descriptions of 19 channel ZNFB training can be found in Collura et al. (2009). The approach includes applying live z-score training across all 19 surface channels using joint time frequency analysis using a QEEG database (i.e., Neuroguide, BrainDx). One may use any number of targeting approaches such as training amplitude, power ratios, coherence, or a combination of targets. The goal of training is for the client to move towards a z = 0 or ‘normalcy’ compared to a QEEG database. It has been claimed that using this approach leads to effective clinical outcomes in fewer sessions than traditional neurofeedback (Krigbaum and Wigton 2015).

LORETA Neurofeedback (LNFB)

LORETA is a popular inverse solution technique that estimates the three-dimensional origination of electrical signals based on a grid of electrodes placed all over the scalp (Pascual-Marqui et al. 1994). Congedo et al. (2004) initially presented using this current source density calculation as the basis for an approach to neurofeedback. Several well-conducted proof-of-concept studies on this method have been pioneered by Marco Congedo and Rex Cannon (See Cannon et al. 2007; Congedo et al. 2004), however in this systematic review we mainly intend to summarize the evidence of the clinical applications of these methods. Other applications of this approach have combined it with the use of live z scores for training into what is now known as LORETA ZNFB (Thatcher 2013). In this approach it is theorized that the neurofeedback is training deeper sources of the EEG than can be accomplished by training at the surface with a small number of electrodes.

These approaches do represent innovations compared to other more traditional forms of neurofeedback and aim to get similar improved signal:noise ratio’s and spatial resolution as is seen with fMRI-neurofeedback.

Methods

We conducted a systematic literature review using the search terms “LORETA neurofeedback” and “19 channel ZNFB” in both PubMed and google scholar as of the beginning of June 2017. PubMed automatically searches the Medline database of the US National Library of Medicine with references dating back to 1996. Google scholar (https://scholar.google.com/) is a freely accessible web search engine that indexes metadata of scholarly literature across an array of publishing formats and disciplines (not just medical fields). This includes peer reviewed journals, books, conference papers abstracts, etc. As such, it provides a greater breadth of references that go beyond PubMed and includes publications in neurofeedback and neuromodulation journals and books. Google scholar began in 2004 and indexes references from this year forward. We also included in this review a list of ZNFB publications made available at the applied neuroscience website (http://appliedneuroscience.com/Z_Score_publications.pdf).

These sources were reviewed for empirical research designs that were published in peer reviewed journals and directly measured the efficacy of one of these approaches in a clinical population. Due to the scarcity of literature, the requirement for randomized controlled studies (RCT’s) was dropped rather early. Papers were only included in this review when they met the following criteria:

  1. 1.

    Group-based studies: case studies and case series are not included

  2. 2.

    Studies investigating the clinical use of neurofeedback in a specific disorder with a specifically defined primary endpoint of clinical improvement [i.e. symptom improvement defined by a valid rating scale e.g. BDI, ADHD-RS etc; proof of concept studies demonstrating a specific region of interest (ROI) can be trained were beyond the scope of this review]

  3. 3.

    Peer-reviewed articles in scientific journals: book chapters, magazine articles or conference presentations are not included since these are not peer reviewed. Furthermore, articles should have been published in journals that engage in adequate peer-review and thus journals that were listed on ‘Beall’s list of predatory journals and publishers’ (https://beallslist.weebly.com) were excluded (for review of predatory journals see Butler 2013)

Results

The PubMed search under the term LNFB resulted in five studies, and under 19 channel ZNFB four studies. The additional Google scholar search attracted six additional LNFB studies and no new 19 channel ZNFB papers. Searching the applied neuroscience list of ZNFB publications resulted in no additions to our list. There are a total of 15 research studies that were reviewed in detail and form the basis of our empirical review. Of these 15, only six met the criteria listed above. Of the nine that were excluded, five included only nonclinical subjects and were considered proof of concept studies and did not test clinical efficacy. The remaining four were either case studies (YuLeung To et al. 2016), case series (Koberda et al. 2012), or not a study at all but a discussion of related topics (Cannon 2012; Simkin et al. 2014). The overall quality of these six studies were judged based on the inclusion of comparison or control groups in which subjects received an experimental manipulation that was not z-score or LNFB training. Only three such studies were found. Random assignment of subjects (another indicator of a higher quality research design) was used in only one of these.

figure a

Six studies were included that fulfilled our inclusion criteria. These are listed in Table 1 above. Three of these focused on surface ZNFB. Krigbaum and Wigton (2015) sought to demonstrate a method of monitoring progress in ten cases receiving 19 channel ZNFB. They showed an improvement towards the mean in 90% of the cases in 15 or fewer sessions. Wigton and Krigbaum (2015) conducted a pilot study of 21 subjects where they assessed neuropsychological and behavioral rating measures on about half of those subjects. Most subjects had ADHD/ADD, albeit no diagnostic information is provided, nor supportive details of diagnosis. There were significant pre- to post-training changes on all measures, but no comparison group was included for comparison and none of the outcome metrics assessed, allowed a comparison to prior meta-analysis (e.g. no DSM based ADHD-RS data). Hammer et al. (2011) conducted a study to investigate if ZNFB could help a group with insomnia. Twelve adults with insomnia met their inclusion criterion, but only eight completed the training. These subjects were divided into two groups, including a four channel ZNFB surface group and a two channel (Cz, C4) ZNFB SMR training group. Sleep measures improved in both groups and there was no difference between the two groups in terms of outcome. There have been no studies using 19 channel ZNFB published in peer reviewed journals that have included any comparison group to understand how this form of neurofeedback might compare to control conditions (not even a wait list control group) or (semi-)active treatments.

Table 1 Review of empirical studies

Only three research studies have evaluated if LNFB has a significant impact on clinical populations compared to appropriate comparison groups. Liechti et al. (2012) have conducted the only controlled trial comparing the effects of LNFB in children with ADHD. Subjects were randomly assigned to a LNFB, standard neurofeedback (Cz) or EMG biofeedback condition, with 14 total participants in the LNFB group. In this latter group they conducted theta-beta and SCP training in counterbalanced order. Symptom ratings decreased significantly in the LNFB group over a span of 36 sessions with effect sizes (ES) within the range of ES published before for ‘standard neurofeedback protocols’ in ADHD (Parent rated inattention ES = 1.26; Hyperactivity/Impulsivity ES = 0.65). However, learning to control the targeted region (anterior cingulate) in the LNFB group was limited. Rather, they demonstrated good control of artifact, eye movement and muscle activity. While there were control conditions in this research, they were not directly compared in this study. The authors concluded that the effects could be from artifact reduction and not the direct effects of the brain regions targeted. In addition, the number of training sessions was approximately the number of sessions commonly seen in traditional neurofeedback research with ADD/ADHD and ES comparable to studies using ‘standard protocols’.

More recently, Cannon et al. (2014) studied the effects of LNFB training in the precuneus. Thirteen participants received 10–20 sessions of training over the left precuneus, including 5 healthy students and 8 subjects with a variety of psychiatric diagnoses. Outcome was further assessed with the Personality Assessment Inventory and Delis-Kaplan Executive Functioning System. Their findings showed significant decreases on PAI scales of psychological distress, but not as much on the cognitive processing measures. Also, they were able to show ROI EEG changes in the precuneus of the controls but not in the clinical subjects.

In another higher quality study, Vanneste et al. (2016) researched the effects of LNFB on tinnitus related distress. A total of 58 subjects were included, 23 received LNFB of the posterior cingulate (alpha up and beta down), 17 LNFB of the lingual gyrus and 18 were part of a wait list control group. The findings showed no significant change on tinnitus loudness, but distress was diminished in the posterior cingulate trained group with a large ES (ES = 0.70) but not for the other groups. This effect was explained by a select number in that group; close to half of those with grade iv distress moved to a lower level of distress. Analysis of the brain change data showed no change of the targeted activity at specific regions, but rather a decrease in connectivity between the posterior cingulate and dorsal anterior cingulate. These findings are fascinating, also due to the protocol specific effects and did include appropriate control groups.

Discussion

Progress in the field of neurofeedback has been stunted by premature claims of efficacy that are not accepted by those outside of our field. For this reason, it is crucial to be clear about what the empirical evidence says about our approaches and techniques (see Coben and Evans 2010).

From a theoretical perspective, these methods that include a greater number electrodes, train towards normality and involve source localized activity as their target would represent methodological and conceptual enhancements. Congedo et al. (2004) were able to design and demonstrate the first application of LNFB which they used in a non-clinical population. These subjects were able to learn volitional control over this feedback signal. Similarly, Cannon and his colleagues have used LNFB in samples of non-clinical subjects to demonstrate its utility. Cannon et al. (2006) were able to target the cognitive division of the anterior cingulate gyrus leading to enhancements in attention and working memory in college students. In a subsequent project, they showed that training led to changes in the anterior cingulate (Cannon et al. 2007), which was targeted, but also to other frontal and even parietal regions as well. Interestingly, Getter et al. (2015) was able to show in only two cases that training over parietal regions altered a mental rotation task (related to parietal functioning). So, while these experimental manipulations and demonstrations are interesting, they are not a demonstration of clinical efficacy which requires specific research findings in clinical subjects with valid clinical outcome measures and control groups.

This review of empirical studies shows that such information is sorely lacking. For ZNFB there is virtually no evidence that it is effective nor that it is “superior” in any way to other forms of neurofeedback. For ZNFB none of the studies included an empirical comparison to any other type of control or treatment group in a clinical population and also none of the outcome measures allowed comparison to prior meta-analysis. As a result, it must be considered baseless in it’s demonstration of clinical efficacy. The quality of studies for LNFB was better compared to ZNFB and some suggestion for efficacy was demonstrated for ADHD (Liechti et al. 2012) and Tinnitus (Vanneste et al. 2016), albeit none of the findings have been replicated and the ES for ADHD was similar to those reported before for ‘traditional neurofeedback protocols’ (Arns et al. 2009). In fact, these effects were hypothesized to be related to control external artifact. For tinnitus, no effects were found for tinnitus loudness, whereas a protocol specific effect was found for distress with a large effect size. Replication and expansion of these findings would clearly be important. Furthermore, LNFB did not show ROI EEG changes in clinical subjects in Cannon et al. (2014) analyses. In light of these findings ZNFB and LNFB protocols cannot be recommended as a first line treatment option and these protocols remain experimental. When standard neurofeedback protocols and other well-investigated treatments have not yielded the desired effects, ZNFB and LNFB could be considered as an experimental off-label treatment, if patients are sufficiently informed about the experimental nature of the protocols and informed consent is used. However, the evidence reviewed in this manuscript does not justify widely advertising ZNFB and LNFB as better and faster alternatives until more solid studies have been conducted.

Our conclusions continue to emphasize the pervasive lack of evidence supporting these approaches to neurofeedback. It is interesting in this light that Koberda et al. (2012) show very little change in efficacy moving from one channel to LNFB. Also, Cannon (2012) discussed possible adverse effects for specific LNFB protocols. We would recommend considering these approaches as potentially promising with a great need for more research prior to making claims of efficacy and certainly could not endorse the notion of their superiority. We strongly encourage more research be devoted to these approaches across multiple clinical conditions and would be supportive with further evidence supporting efficacy and safety. Until then, these should be considered experimental.