Understanding the Effects of General Anesthetics on Cortical Network Activity Using Ex Vivo Preparations

General anesthetics have been used to ablate consciousness during surgery for more than 150 years. Despite significant advances in our understanding of their molecular-level pharmacological effects, comparatively little is known about how anesthetics alter brain dynamics to cause unconsciousness. Consequently, while anesthesia practice is now routine and safe, there are many vagaries that remain unexplained. In this paper we review the evidence that cortical network activity is particularly sensitive to general anesthetics ― and suggest that disruption to communication in and/or among cortical brain regions is a common mechanism of anesthesia that ultimately produces loss of consciousness. We review data from acute brain slices and organotypic cultures showing that anesthetics with differing molecular mechanisms of action share in common the ability to impair neurophysiologic communication. While many questions remain, ex vivo, and in vivo investigations together suggest that a unified understanding of both clinical anesthesia and the neural basis of consciousness is attainable. Introduction and background Since the inception of clinical anesthesia in the mid-19th century, scientists have greatly advanced our fundamental understanding of modern anesthetics. In particular, there is now consensus that their molecular targets are proteins that control neuronal excitability and connectivity, including ligandand voltage-gated channels 1,2. Anesthetic interactions at the GABAA receptor are best understood ― where they bind to receptor subunit interfaces to potentiate the effect of GABA binding and prolong the resulting increase in chloride conductance 3. This explains why many anesthetics broadly depress brain activity. However, understanding how molecular-level effects translate into macro network-level phenomena underlying unconsciousness has proven problematic. This is where our understanding of anesthetic action falters ― and bridging this gap is an important focus for ongoing research efforts. The aim of this review is to highlight the insights into network-level mechanisms of anesthesia derived from experiments utilizing ex vivo brain slice preparations. We begin HHS Public Access Author manuscript Anesthesiology. Author manuscript; available in PMC 2019 June 01. Published in final edited form as: Anesthesiology. 2019 June ; 130(6): 1049–1063. doi:10.1097/ALN.0000000000002554. A uhor M anscript


Introduction and background
Since the inception of clinical anesthesia in the mid-19 th century, scientists have greatly advanced our fundamental understanding of modern anesthetics. In particular, there is now consensus that their molecular targets are proteins that control neuronal excitability and connectivity, including ligand-and voltage-gated channels 1,2 . Anesthetic interactions at the GABA A receptor are best understoodwhere they bind to receptor subunit interfaces to potentiate the effect of GABA binding and prolong the resulting increase in chloride conductance 3 . This explains why many anesthetics broadly depress brain activity. However, understanding how molecular-level effects translate into macro network-level phenomena underlying unconsciousness has proven problematic. This is where our understanding of anesthetic action faltersand bridging this gap is an important focus for ongoing research efforts.
The aim of this review is to highlight the insights into network-level mechanisms of anesthesia derived from experiments utilizing ex vivo brain slice preparations. We begin An alternative to acute ex vivo preparations of the brain are organotypic slice cultures, essentially slices from neonatal or embryonic animals, cultivated for days to months. Although they have been in use for electrophysiological experiments for an equally long time 14 they are not nearly as widely used as acute slices. As has been detailed elsewhere 15,16 there is a tradeoff: high levels of spontaneous activity, the possibility to design a wide range of brain sub-networks not feasible in acute preparations, and fast diffusion of drugs into the thin tissue have to be weighed against a less faithful neuronal morphology and synaptic architecture.
In order to illustrate the important contribution ex vivo models have made to our understanding of anesthesia mechanisms, we next present the context of related and equally important in vivo experimental and clinical investigations. With a spotlight on ex vivo models, we have endeavored to integrate a broad knowledge base into a coherent picture of cortical anesthetic effects.

Cortical versus subcortical anesthetic action
While cerebrocortical networks are central to consciousness, there remains contention as to whether the behavioral effects of anesthetics are better explained by direct action on the cortexor via subcortical effects on the thalamus or on natural sleep and arousal pathways. The potential influence of these neurophysiologic details on the quality of anesthetic induction, maintenance and emergence highlights the importance of this research avenue.
The subcortical and cortical areas involved in anesthetic action are not easily functionally separated when considering the patient as a whole. It can be helpful to conceptually divide general anesthesia appropriate for surgery as targeting two aspects of consciousness: level of consciousness, including overall arousal and responsiveness to external stimuli, and content of consciousness, including interpretation of the saliency of these signals as sensations that could be harmful or warrant decisive action, as well as mentation that may be unrelated or only tangentially related to the sensory environment 17 . This functional separation can be used as a framework for understanding the neuroanatomy of clinical anesthesia at the systems neuroscience level. Accumulating evidence suggests that general anesthetics produce some aspects of hypnosis by acting on sleep and arousal centers in the brain. For example, anesthetics promote sleep-like patterns of activity in the subcortical nuclei in hypothalamus and brainstem that control sleep, for example activating sleep-active neurons in the ventrolateral preoptic nucleus [18][19][20] . These agents also depress activity in subcortical arousal centers, e.g. noradrengergic, cholinergic and dopaminergic neurons in the locus ceruleus, basal forebrain and ventral tegmental area, respectively 21,22 . There is also extensive evidence for anesthetic actions on the thalamus. Thalamus is a critical hub in the ascending arousal system, in relaying sensory information to cortex 23 , and in corticocortical communication 24 . Activity in thalamus is suppressed by nearly all anesthetics, with the exception of ketamine 25 . The functional roles of each of these anesthetic targets in producing and maintaining hypnosis and in emergence from anesthesia, can be defined within the dichotomy of level versus content of consciousness 17 . Actions on sleep and arousal centers (including the arousal-promoting functions of thalamus) would control arousal, i.e. level of consciousness, whereas the content of consciousness would be due to direct actions on the cortico-thalamic network, and also secondary to actions on sleep and arousal centers. Understanding the relative contributions of each to the functional endpoint of clinical anesthesia is a formidable challengeand may ultimately prove to be dependent on clinical context and individual-specific differences.

EEG changes under general anesthesia
The importance of subcortical targets for anesthesia is supported by similarities in the electroencephalogram (EEG) during general anesthesia and natural sleep. For agents that predominantly act by enhancing GABA A receptor inhibition (for example, propofol, volatile agents etc), surgical anesthesia is characterised by heightened power in the alpha and delta bands, resembling non-REM Stage 2 and non-REM Stage 3 sleep, respectively [26][27][28][29] . Recent in vivo work in rats has demonstrated that the central medial thalamus may act as an initiation hub for natural sleep and propofol anesthesia, with changes in dynamics of high frequency bands occurring prior to similar changes in the neocortex 27 . Neuroimaging studies also support functional thalamic disconnection during anesthesia 30,31 . However, as reviewed by Antkowiak, most of the EEG field potential patterns indicative of general anesthesia (thalamocortical oscillations being the obvious exception) can be reproduced in isolated cortical networks in the absence of subcortical connections 32 . This suggests that an EEG pattern resembling that of NREM sleep does not prove subcortical causation. Furthermore, not all anesthetics depress the EEG, ketamine being an obvious case in point 33,34 .

Subcortical microinjection studies
Subcortical action is further supported by microinjection experiments showing that highly localised drug application to some mesopontine/midbrain regions can, on their own and without direct cortical effects, cause an anesthesia-like state 18,35,36 , as well as studies showing that stimulation of dopaminergic VTA can hasten arousal from anesthesia in rodents 37 . Devor and colleagues have shown that anesthetic microinjection into a highly localised region of the upper brainstem called the mesopontine tegmental anesthetic area (MPTA) induces a reversible state of unconsciousness 38 , along with associated anesthesialike changes in the electroencephalogram 35 . Investigation of fos expression suggests that general anesthesic cortical action may be attributable to effects at subcortical neuromodulatory sites 39 . It should be noted, however that these results have not been fully replicated elsewhere 40 . Also, because the extent of drug diffusion is difficult to estimate, the choice of drug concentration is not straightforward in these types of microinjection studies and may result in a tissue concentration as much as 50x the cerebral tissue concentration required for thiopental anesthesia in rats 35,41,42 . Under the reasonable assumption that such concentrations incapacitate neurons in any brain region, these studies provide valuable information on the role of the subcortical site in question in consciousness. However, this should not be considered evidence of the sites' pivotal role in bringing about unconsciousness by clinically relevant concentrations of anesthetics. It is also worth noting that the microinjection studies that document dramatic anesthesia-like effects form a small minority of a large number of similar investigations. As summarised recently by Leung et al 43 , these studies generally report an increase in anesthetic sensitivity, but not general anesthesia outright. At the very least, no one subcortical site seems to form a unifying "anesthesia -sensitive" area.

Pharmacologic investigations of a hyperpolarized thalamus
An idea that has remained prominent since its introduction by Angel in 1991 44 , is that thalamic hyperpolarisation causes a transmission block to the passage of sensory information from periphery to cortex. The premise is that consciousness is unsustainable in the absence of a sensory substrate. While also appealing in its intuitive simplicity, there is now compelling experimental data refuting this hypothesis. Firstly, in keeping with its tendency to maintain cortical activity, ketamine at anesthetic concentrations neither suppresses the thalamus 45 , nor blocks thalamo-cortical sensory transmission 46 . Cortical sensory evoked potential studies also show that primary sensory cortical responsiveness is preserved, and can even be enhanced during anesthesia [47][48][49][50] . It thus seems likely that unconsciousness occurs at the level of disruption to cortical information processing, not cortical sensory reception. Importantly, this does not preclude the thalamus as an essential anesthetic target, particularly in light of the role of thalamo-cortical relays in mediating cortico-cortical communication 51 . Dexmedetomidine is reported to block resting state thalamo-cortical connectivity to a greater extent than cortico-cortical connectivity 52 . Recent data also show that xenon anesthesia may be mediated by an HCN2-blocking action with effects on thalamo-cortical primary sensory relay pathways 53 . The role of HCN channels in mediating anesthesia effects is not clear however. For example, targeted blockade of HCN channels with ZD-7288 is without effect on highly synchronised "seizure-like event" (SLE) activity in cortical slices 54 while anesthetics such as propofol, etomidate and ketamine consistently reduce SLE activity in cortical slices 55 .

Cortical versus subcortical anesthetic sensitivity
The cortical /subcortical debate hinges not so much on whether anesthesia can be induced by action at one or the other siteit seems that both are possible 17 . Instead, the relative importance of one or the other depends on their relative anesthetic-sensitivity at clinically relevant concentrations. This distinction is not only of theoretical interest but is vital for understanding clinically relevant behavioral phenomena associated with anesthesia. For example, even though emergence from anesthesia may not be a simple passive reversal of the induction process 56 , knowing the relative influence of dwindling anesthetic concentrations on the spiking activity at cortical vs subcortical sites is pivotal for understanding the nature of these transitions. If the subcortex is more sensitive than the cortex, emergence from anesthesia may be dominated by lingering anesthetic effects on subcortical structures (breathing, detection of sensory stimuli). Conversely, lingering cortical effects during emergence could result in a mismatch between subcortical sensory input mediating arousal and readiness to integrate complex information from different cortical areas. The arousal disturbance in children resulting in so-called "night terrors" may be a manifestation of this in natural sleep 57 .
The balance of evidence seems to favour cortical over subcortical sensitivity, with the proviso that this remains an area of active researchand is perhaps even individualspecific. High frequency EEG power (> 14 Hz, beta and gamma waves) represent the transfer of information among cortical regions during wakefulness 58 and when present during clinical anesthesia for surgery, these features are associated with patient movement and heightened arousal 59 . These "cortical arousals" can be pharmacologically induced and appear to hasten emergence from anesthesia in vivo 60 . Furthermore, anesthetic reduction in cortical firing rates have been shown to be similar with or without subcortical connectivity 61 .

Interactions among the thalamus and cortex
Thalamo-cortical slice studies add further support to the corticocentric view of anesthetic action. Although early reports provided evidence that anesthetics might suppress cortical activity via actions in the thalamus 62 , more recent reports have demonstrated greater sensitivity of intracortical compared to thalamocortical signal pathways. By selectively activating cortico-cortical "top-down" and thalamo-cortical "bottom-up" pathways, Raz and colleagues showed that evoked cortico-cortical responses exhibited greater sensitivity to isoflurane compared to thalamo-cortical responses (Figure 2) 49 . In vivo, visual responses in A1, mediated at least in part by projections from V2 63 , are blocked by anesthetics at doses that do not suppress auditory responses to pure tones 49 . Consistent with these data, earlyand mid-latency components of auditory evoked potentials are resistant to the effects of anesthesia, at least at doses causing loss of consciousness (LOC) 64,65 . These components most likely correspond to subcortical and thalamo-cortical synaptic potentials [66][67][68] . At higher doses, corresponding to surgical levels of anesthesia, even these shorter components are suppressed by general anesthetics 69,70 . By contrast, longer latency components, likely reflecting cortico-cortical signaling, are sensitive to even low doses of anesthetics 71,72 . These data suggest that many of the effects on sensory evoked potentials during clinical anesthesia may be due to direct cortical effects rather than effects on thalamo-cortical afferents. In support of this, changes in cortical evoked potentials in isolated cortical slices closely resemble those observed in vivo 73 for a range of anesthetic classes 50 .
Imaging and EEG studies in primates and human subjects also point to cortico-cortical synaptic signaling as a critical locus for effects of anesthetics on consciousness 74,75 . During midazolam-induced LOC and during slow wave sleep, cortical responses to TMS are enhanced locally but the spread of activity due to cortico-cortical interactions is reduced 47,76 ; similar effects are observed with the dissociative anesthetic ketamine [77][78][79] . 'Feedback' cortical connections are particularly sensitive to hypnotic doses of anesthetics 30,80-86 , as they are in non-REM sleep and in disorders of consciousness 87,88 . These data, which are conserved across species 89 , suggest that LOC is accompanied by reduced cortical connectivity in the presence of maintained responsiveness to thalamic inputs in primary sensory cortex 90,91 (Figure 3).
In the remaining sections we will explore the contribution that acute and organotypic slices can make to understanding and substantiating cortical anesthetic action. In particular, evidence suggests that cortical network activity may underlie bidirectional communication within the cortical hierarchy 92 , a breakdown in which may disrupt the integrative processes central to conscious experience 93,94 .While the cortex is central to the information that mediates our conscious experience it should once again be pointed out that this does not preclude the involvement of thalamic and other subcortical nuclei 95,96 .

Anesthetic effects on physiological cortical network activity
Are effects on cortical networks consistent with a corticocentric view of anesthetic action? At the very least, there is strong data to indicate that the cerebral cortex is an important direct target of anesthetic drugs. Accordingly, models utilising isolated cortical networks provide a valuable tool for investigating anesthetic mechanisms of action. Many have approached this by examining signature electrical patterns in cortical networks that are relevant to anesthetic mechanisms causing unconsciousness.

Network activity in organotypic cultures inform anesthesia research
A commonly reported effect of anesthetics consists of a reduction of network activity. Evidence that this effect is mediated at least in part by direct actions on cortical circuits is derived from experiments on neocortical organotypic cultures, where spontaneous activity rates are very sensitive to virtually all classes of anesthetics and sedatives 97 . Diazepam was shown to depress spontaneous firing rates in a biphasic concentration-dependent manner, highlighting the different roles of classical and non-classical benzodiazepine binding sites of GABA(A) receptors 98 . Analysis of activity rates and patterns in the same model system allowed disentangling the effects of midazolam from its main metabolite, suspected to have potent depressive actions on neuronal activity as well 99 . In vivo, however, the picture is more complex. While a reduction in cortical neuronal firing (with accompanying changes in the EEG) does reliably accompany slow-wave sleep 100 and anesthesia 61,101-104 , dissociative anesthetics such as ketamine maintain or even increase cortical activity 105 . In addition, depression of cortical activity by anesthetics can be unrelated to LOC per se [106][107][108] . Therefore, we posit that although changes of activity rates in cortical networks ex vivo are a very useful approximate marker of the potency/efficacy of anesthetics, there is no straightforward causal relation between their change and the state of consciousness.

Anesthetic effects on cortical network activity and sensory information processing in acute slices
Emerging evidence suggests that information transfer within the cortical network may occur via all-or-none responses in cortical ensembles rather than stochastic firing of individual cells 92,[109][110][111] . In auditory cortex, network bursts are triggered by sensory stimulation and contain specific spatio-temporal spike sequences ('packets') whose organization is stimulusspecific and thus may underlie a population-based encoding process 112,113 . Importantly, these network bursts and the network bistability that underlies them are observed in waking conditions [114][115][116] , suggesting an important role in sensory awareness. Spontaneous and induced network activity similar to that recorded in vivo during sensory processing can readily be observed in acute slices of rodent neocortex [117][118][119][120] . The occurrence and propagation of this activity in cortex is modulated by volatile anesthetics, which can both promote its occurrence by synchronizing network activity and disrupt its propagation by interfering with cortico-cortical communication. Indeed, we have shown that in murine auditory thalamo-cortical brain slices, nearly all spiking activity in response to stimulation of thalamo-cortical afferents occurs in the context of network bursts 119 . In acute slices, which dwell primarily in quiescent 'Down' states, spontaneous and induced network bursts are suppressed by the volatile anesthetic isoflurane at moderate doses ( Figure 4A) 49,121 . Importantly, spiking that is monosynaptically-driven by thalamo-cortical afferents is significantly more difficult to block by isoflurane than reverberant activity generated within the burst by local cortico-cortical connections ( Figure 4B,C). These results are consistent with a model in which anesthetics, at doses causing LOC and suppression of sensory awareness, act directly on circuits intrinsic to cortex rather than on pathways carrying information to cortex from the periphery.
Sharing of information between local cortical networks is central to supporting consciousness 122 . A major pathway for this integration process is via cortico-cortical connections between cortical columns, which form functional units for processing sensory information. Disruption of these connections by anesthetic agents would likely contribute to cortical disintegration during loss of consciousness. Experiments in brain slices support this model. Network activity propagates 'horizontally' in cortex from column to column via local connectivity 112,117,123,124 , especially among infragranular pyramidal cells 124 . In thalamocortical brain slices, isoflurane suppresses inter-columnar propagation of network activity, either spontaneous or induced by thalamo-cortical or cortico-cortical afferent stimulation ( Figure 5A,B) 121 . These effects are consistent with a model in which isoflurane suppresses the activation of local networks by the propagating wave. Anesthetic interruption of propagating network activity is also seen in cortical slices activated by removal of artificial cerebrospinal fluid magnesium ( Figure 6A,B). In this model spontaneous SLE activity spreads freely across the full extent of the cortex, even across hemispheres. While strictly an epileptiform model, correlated network activity in neocortex has also been described in the context of synchronized brain states that occur under surgical levels of anesthesia and during slow wave sleep 125 . Furthermore, anesthetic effects on SLE activity directly correlates with in vivo anesthetic hypnotic potency 55,[126][127][128] . Two main conclusions can be drawn from recent experiments. Firstly, agents of different classes, including ketamine, etomidate and propofol, constrain the spatial extent of SLE spread from the event initiation source 128 . When multiple initiation locations are evident in the same slice, the effect is identical to isolating one source from the other by physically sectioning the tissue. Thus, the anesthetic effect seems to equate to a functional disconnection. Secondly, the population events recorded near the source of SLE generation have shorter rise times and higher amplitudes during anesthetic exposure 129 , suggestive of enhanced local network synchrony. The dual effects of enhanced local response combined with reduced activity spread bears striking resemblance to the effect of midazolam on TMS responses in anesthetised human subjects 47 .

Hippocampal slice preparations and their role in anesthesia research
Much of the original work with brain slice techniques focused on the hippocampus, a specialized archaecortical structure in the temporal lobe that is conserved in form and function across mammals 4 . The hippocampal slice has led to many great advances in understanding learning and memory 130 and anesthesiology researchers have used this system extensively to better understand the amnesic effect of anesthetic drugs [131][132][133]  However, even with its classic and well described role in learning a memory consolidation, it's an over simplification to exclude the hippocampus from a discussion on the hypnotic effect of anesthetics. The hippocampus is exquisitely sensitive to anesthetics 134 and a case can be built for it subserving a critical role in binding episodic memory formation to our experience of the external world 135 . It has been argued that the neural processes underpinning episodic memory could actually be regarded as indistinguishable from those underpinning consciousness 135 . The neuroanatomical connectivity of the hippocampus to some extent supports this. Although not spatially arranged exactly like those from the neocortex (i.e., motor and sensory cortices) 23 , the hippocampus does have many reciprocal connections with corticothalamic networks.
Outright seizures are a rare but important complication of anesthesia 136,137 . The hippocampal slice has also been used in combination with pharmacological manipulation of the GABA(A) receptor for mechanistic investigations of seizure. For example, propofol 134 and isoflurane 138 have differential excitatory effects on cortical and hippocampal networks. Propofol, in particular, has a propensity to induce ictal-like events in hippocampal slices 134 . Interestingly, recent evidence suggests that inhibition of GABA signalling by the bath application of the macrolide antibiotic clarithromycin does not cause seizure-like events when administered alone to slices, but increases the frequency of these events when given in combination with a solution (high potassium) known to hyperexcite 139 . Human studies link clarithromycin with improved vigilance in patients with hypersomnia 140 , highlighting a potential avenue for hastening recovery from general anesthesia without hypoactive delirium in the recovery room 141 .

Functional implications and future directions
While we present for actions of anesthetics at multiple locations in the brain, each contributing to specific anesthetic endpoints, our emphasis has been on cortical actions that likely impact content of consciousness. That is, while actions on brainstem and midbrain nuclei play important roles in arousal and in the transitions into and out of awareness, there is substantial evidence that consciousness itself is a cortico-thalamic phenomenon. Within such a highly recurrent network, there may be multiple hubs and pathways that play critical roles, any one of which may represent the primary or 'first' site of anesthetic action depending on the agent, context and organism of interest. The greatest current gap in our understanding of loss and recovery of consciousness under anesthesia lies at this network level. Because monitors of brain activity in clinical settings invariably sample activity averaged over large numbers of cortical cells, improvements in assays of depth of anesthesia will derive from studies at this spatial scale. Herein, we have reviewed some of the evidence that cortical network activity is particularly sensitive to general anesthetics, but many questions remain. Primary on this list is the mechanism underlying this sensitivity. Is the disruption of cortical network activity a function of the structure of the network, i.e. sparsely-firing cells densely interconnected via synaptic connections with low release probability, balanced excitation and inhibition, and columnar stratification of external connectivity 142 , or do cortico-cortical synapses or cortical pyramidal cells possess intrinsic elements that are specifically sensitive to anesthetics, properties not found at thalamocortical synapses or in membranes of subcortical neurons? A second critical question is why disruption of network activity is so deleterious, given that sensory stimulation still drives activity in sensory cortex under anesthesia and gross features of tuning are preserved [143][144][145] . Evidence suggests that cortical network activity may underlie bidirectional communication within the cortical hierarchy 92 that is critical for information integration and predictive coding 93,94 , processes central to construction of self, mind and experience. Investigations in vivo and ex vivo aimed at understanding how the transfer of information packets 109 between cortical regions occurs and how this process is disrupted by general anesthetics will contribute to our understanding of both anesthesia in a clinical setting and to the neural basis of consciousness itself. (A) schematic illustrating preparation of brain slices. Brains are sliced in any desirable plane and orientation (shown is a coronal slicing plane) and, depending on subsequent use, the slices may be trimmed to include just the region of interest. The resulting slices are either allowed to recover 1-2 h and are then used for experimentation on the same day ('acute slice') or are placed on a substrate and cultivated in nutritional medium, resulting in an organotypic slice culture. Acute slices can be prepared from animals of any age (commonly juveniles) whereas for cultures neonatal animals or even embryos are often required. (B) sketch of a partial coronal brain slice including neocortex, thalamus and hippocampus. Blue dots illustrate commonly used orientations of multi-channel recording sites within neocortex. In the 'horizontal' orientation the sites are situated within one layer or along layer boundaries, allowing the recording of inter-areal propagation of neuronal activity. The 'vertical', cross-layer orientation, running parallel to pyramidal cells' apical dendrites, is usually chosen if the spread of activity within a cortical column/across layers is of interest, e.g. upon thalamic stimulation. Two pyramidal cells (magenta) are shown schematically to illustrate the orientation of the long apical dendrites.     Schematic (A) and recorded data (B) showing the effect of etomidate on the pattern of zeromagnesium seizure-like event (SLE) activity in the cortical slice. Shown is one hemisphere of a coronally cut slice with 2 recording electrodes (R1 and R2), with a hypothetical (but realistic 129 ) scenario of 2 independent sources of SLE activity (S1 and S2) -each of which initiate repeating waves of excitation that spread across the full extent of the cortex in opposite directions (A, left). Under this baseline (drug-free) condition, each event will be recorded by both electrodes, with small inter-electrode time-lags reflecting the speed of wave propagation. As such, each event will appear "synchronised" across both channels (B, left). A proposed explanation for the effect of etomidate is shown schematically (A, right). Propagation of some of the SLE wavefronts is curtailed such that some of the events initiated at S1 will not reach R1 and vice versa. Consequently, the recordings will take on a "desynchronised" appearance (B, right). Variations of this theme will be apparent from slice to slice, according to the number of SLE initiation sources present and where those sources are located relative to the recording electrode positions. Recorded data is from 129 .