<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Syst. Neurosci.</journal-id>
<journal-title>Frontiers in Systems Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Syst. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5137</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnsys.2022.869705</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Estimating How Sounds Modulate Orientation Representation in the Primary Visual Cortex Using Shallow Neural Networks</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>McClure</surname> <given-names>John P.</given-names><suffix>Jr.</suffix></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/212083/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Erkat</surname> <given-names>O. Batuhan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Corbo</surname> <given-names>Julien</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1665947/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Polack</surname> <given-names>Pierre-Olivier</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/396121/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Center for Molecular and Behavioral Neuroscience, Rutgers University&#x2013;Newark</institution>, <addr-line>Newark, NJ</addr-line>, <country>United States</country></aff>
<aff id="aff2"><sup>2</sup><institution>Behavioral and Neural Sciences Graduate Program, Rutgers University&#x2013;Newark</institution>, <addr-line>Newark, NJ</addr-line>, <country>United States</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Gyorgy Lur, University of California, Irvine, United States</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Guido Thomas Meijer, Champalimaud Centre for the Unknown, Champalimaud Foundation, Portugal; Xiaoyu Tang, Liaoning Normal University, China</p></fn>
<corresp id="c001">&#x002A;Correspondence: Pierre-Olivier Polack, <email>polack.po@rutgers.edu</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>09</day>
<month>05</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>16</volume>
<elocation-id>869705</elocation-id>
<history>
<date date-type="received">
<day>04</day>
<month>02</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>07</day>
<month>04</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2022 McClure, Erkat, Corbo and Polack.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>McClure, Erkat, Corbo and Polack</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Audiovisual perception results from the interaction between visual and auditory processing. Hence, presenting auditory and visual inputs simultaneously usually improves the accuracy of the unimodal percepts, but can also lead to audiovisual illusions. Cross-talks between visual and auditory inputs during sensory processing were recently shown to occur as early as in the primary visual cortex (V1). In a previous study, we demonstrated that sounds improve the representation of the orientation of visual stimuli in the na&#x00EF;ve mouse V1 by promoting the recruitment of neurons better tuned to the orientation and direction of the visual stimulus. However, we did not test if this type of modulation was still present when the auditory and visual stimuli were both behaviorally relevant. To determine the effect of sounds on active visual processing, we performed calcium imaging in V1 while mice were performing an audiovisual task. We then compared the representations of the task stimuli orientations in the unimodal visual and audiovisual context using shallow neural networks (SNNs). SNNs were chosen because of the biological plausibility of their computational structure and the possibility of identifying <italic>post hoc</italic> the biological neurons having the strongest influence on the classification decision. We first showed that SNNs can categorize the activity of V1 neurons evoked by drifting gratings of 12 different orientations. Then, we demonstrated using the connection weight approach that SNN training assigns the largest computational weight to the V1 neurons having the best orientation and direction selectivity. Finally, we showed that it is possible to use SNNs to determine how V1 neurons represent the orientations of stimuli that do not belong to the set of orientations used for SNN training. Once the SNN approach was established, we replicated the previous finding that sounds improve orientation representation in the V1 of na&#x00EF;ve mice. Then, we showed that, in mice performing an audiovisual detection task, task tones improve the representation of the visual cues associated with the reward while deteriorating the representation of non-rewarded cues. Altogether, our results suggest that the direction of sound modulation in V1 depends on the behavioral relevance of the visual cue.</p>
</abstract>
<kwd-group>
<kwd>neuronal representations</kwd>
<kwd>shallow neural network</kwd>
<kwd>primary visual cortex (V1)</kwd>
<kwd>sound modulation</kwd>
<kwd>orientation representation</kwd>
<kwd>audiovisual detection task</kwd>
<kwd>sensory processing</kwd>
</kwd-group>
<contract-num rid="cn001">R01 EY030860</contract-num>
<contract-sponsor id="cn001">National Eye Institute<named-content content-type="fundref-id">10.13039/100000053</named-content></contract-sponsor>
<contract-sponsor id="cn002">Whitehall Foundation<named-content content-type="fundref-id">10.13039/100001391</named-content></contract-sponsor>
<contract-sponsor id="cn003">Fondation Fyssen<named-content content-type="fundref-id">10.13039/501100003135</named-content></contract-sponsor>
<counts>
<fig-count count="6"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="43"/>
<page-count count="15"/>
<word-count count="11264"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="intro">
<title>Introduction</title>
<p>Multi-sensory integration leads to a multimodal unified percept. It was long thought that multimodal integration was performed in higher-order multisensory cortices such as the posterior parietal cortex (<xref ref-type="bibr" rid="B26">Molholm et al., 2006</xref>; <xref ref-type="bibr" rid="B37">Song et al., 2017</xref>) once the parallel unimodal processing of the different sensory modalities was completed. However, several recent studies have demonstrated the presence of direct mutual anatomical connections (<xref ref-type="bibr" rid="B10">Falchier et al., 2002</xref>; <xref ref-type="bibr" rid="B35">Rockland and Ojima, 2003</xref>; <xref ref-type="bibr" rid="B3">Cappe and Barone, 2005</xref>; <xref ref-type="bibr" rid="B20">Iurilli et al., 2012</xref>; <xref ref-type="bibr" rid="B19">Ibrahim et al., 2016</xref>; <xref ref-type="bibr" rid="B8">Deneux et al., 2019</xref>; <xref ref-type="bibr" rid="B12">Garner and Keller, 2021</xref>) and cross-modal sensory processing modulations (<xref ref-type="bibr" rid="B20">Iurilli et al., 2012</xref>; <xref ref-type="bibr" rid="B19">Ibrahim et al., 2016</xref>; <xref ref-type="bibr" rid="B25">Meijer et al., 2017</xref>; <xref ref-type="bibr" rid="B8">Deneux et al., 2019</xref>; <xref ref-type="bibr" rid="B21">Kn&#x00F6;pfel et al., 2019</xref>; <xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>; <xref ref-type="bibr" rid="B12">Garner and Keller, 2021</xref>) at the earliest stages of cortical sensory processing in primates and rodents. Hence, we have recently demonstrated that sounds modulate the visually evoked response of neurons of the primary visual cortex (V1) to the presentation of oriented stimuli (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). We showed that sounds potentiate the responses of neurons well-tuned to the stimulus orientation and direction while suppressing the responses of neurons not tuned for the orientation and/or the direction of the visual cue. As a result, sound modulation improved the representation of the orientation and the direction of the visual stimulus in V1 (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). If studies on cross modal interactions have mainly reported facilitatory interactions (<xref ref-type="bibr" rid="B41">Vroomen and De Gelder, 2000</xref>; <xref ref-type="bibr" rid="B30">Odgaard et al., 2004</xref>; <xref ref-type="bibr" rid="B23">Lippert et al., 2007</xref>; <xref ref-type="bibr" rid="B14">Gleiss and Kayser, 2014</xref>), others have shown context-dependent suppressive effects (<xref ref-type="bibr" rid="B20">Iurilli et al., 2012</xref>; <xref ref-type="bibr" rid="B18">Hidaka and Ide, 2015</xref>; <xref ref-type="bibr" rid="B25">Meijer et al., 2017</xref>; <xref ref-type="bibr" rid="B8">Deneux et al., 2019</xref>). Altogether, those studies suggest that the sign of sound modulation in V1 depends on the behavioral relevance of the visual and auditory stimuli.</p>
<p>To test this hypothesis, we performed calcium imaging in the V1 of mice alternating, during the same recording session, between the performance of a unimodal visual and an audiovisual task. To compare the representations of the visual stimuli in V1 between the unimodal and the audiovisual context, we tested a novel approach for this type of analysis: the shallow neural networks (SNNs). SNNs are simple neural networks having only one or two inner layers. Like other neural networks, they are classifiers that can be trained at identifying patterns with a very high proficiency (<xref ref-type="bibr" rid="B11">Fukushima, 1980</xref>). This approach was selected to fulfill the following criteria: (1) to be as biologically plausible as possible; (2) to use all the recorded neurons as an input instead of requiring the selection of &#x201C;active&#x201D; neurons; (3) to be able to classify V1 responses to orientations that do not belong to the training set; and (4) to allow determining which neurons carry the most weight in the classification of the visual stimulus. SNNs fulfill those four criteria as: their structure is inspired by the computational structure of the visual cortex (<xref ref-type="bibr" rid="B11">Fukushima, 1980</xref>); they do not require any criteria-based selection from the experimenter for including neurons; their output being a vector of probabilities assigned to each orientation of the training set, we can take advantage of the continuity of the orientation space and use circular statistics to decode any orientation; finally, their simplicity (i.e., their shallowness) allows for a straightforward access to the weight given in the classifying decisions to each individual recorded neuron.</p>
<p>We first tested the SNN approach using a calcium imaging dataset from a prior study to investigate how pure tones affect the representation of oriented stimuli in the V1 L2/3 of mice passively receiving the stimuli (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). We found that the weight assigned to each recorded neuron by the classifier during training was highly correlated with the neuron&#x2019;s tuning properties (preferred orientation and selectivity), suggesting that an optimal classifier uses the same features of the neuronal responses that we capture with the traditional approach of orientation tuning curves. Then, we used the trained SNNs to classify orientations that were not part of the training set. We showed that the presentation of a pure tone improved the visual stimulus representation. Those results reproduced the findings obtain when analyzing the same database using an active neurons selection approach (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). Then, we extended the method to our new dataset and showed that when pure tones have a behavioral relevance in the audiovisual task, the modulation of the representation of visual information in V1 can be bidirectional.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<title>Materials and Methods</title>
<p>All the procedures described below have been approved by the Institutional Animal Care and Use Committee (IACUC) of Rutgers University&#x2013;Newark, in agreement with the Guide for the Care and Use of Laboratory Animals (<xref ref-type="bibr" rid="B28">National Research Council of the National Academies, 2011</xref>).</p>
<sec id="S2.SS1">
<title>Surgery</title>
<sec id="S2.SS1.SSS1">
<title>Head-Bar Implants</title>
<p>Ten minutes after systemic injection of an analgesic (carprofen, Zoetis, Parsippany-Troy Hills, NJ, United States; 5 mg per kg of body weight), adult (3&#x2013;6 months old) male and female Gad2-IRES-Cre (Jackson stock #019022) &#x00D7; Ai9 (Jackson stock #007909) mice were anesthetized with isoflurane (5% induction, 1.2% maintenance) and placed in a stereotaxic frame. Body temperature was kept at 37&#x00B0;C using a feedback-controlled heating pad. Pressure points and incision sites were injected with lidocaine (2%). Eyes were protected from desiccation with artificial tear ointment (Dechra, Northwich, United Kingdom). Next, the skin covering the skull was incised and a custom-made lightweight metal head-bar was glued to the skull using Vetbond (3M, Saint Paul, MN, United States). In addition, a large recording chamber capable of retaining the water necessary for using a water-immersion objective was built using dental cement (Ortho-Jet, Lang, Dental, Wheeling, IL, United States). Mice recovered from surgery for 5 days, during which amoxicillin was administered in drinking water (0.25 mg/mL).</p>
</sec>
<sec id="S2.SS1.SSS2">
<title>Adeno-Associated Virus (AAV) Injection</title>
<p>After recovery from the head-bar surgery, mice were anesthetized using isoflurane as described above. A circular craniotomy (diameter = 3 mm) was performed above V1. The AAV vector AAV1.eSyn.GCaMP6f.WPRE.SV40 (UPenn Vector Core, Philadelphia, PA, United States) carrying the gene of the fluorescent calcium sensor GCaMP6f was injected at three sites 500 &#x03BC;m apart around the center of V1 (stereotaxic coordinates: &#x2212;4.0 mm AP, +2.2 mm ML from bregma) using a MicroSyringe Pump Controller Micro 4 (World Precision Instruments, Sarasota, FL, United States) at a rate of 30 nl/min. Injections started at a depth of 550 &#x03BC;m below the pial surface and the tip of the pipette was raised in steps of 100 &#x03BC;m during the injection, up to a depth of 200 &#x03BC;m below the dura surface. The total volume injected across all depths was 0.7 &#x03BC;l. After removal of the injection pipette, a 3-mm-diameter coverslip was placed over the dura, such that the coverslip fits entirely in the craniotomy and was flush with the skull surface. The coverslip was kept in place using Vetbond and dental cement. Mice were left to recover from the surgery for at least 3 weeks to obtain a satisfactory gene expression.</p>
</sec>
</sec>
<sec id="S2.SS2">
<title>Functional Imaging</title>
<sec id="S2.SS2.SSS1">
<title>Calcium Imaging Setup</title>
<p>During the last week of recovery, mice were trained to stay on a spherical treadmill consisting of a ball floating on a small cushion of air that allowed for full 2D movement (<xref ref-type="bibr" rid="B33">Polack et al., 2013</xref>). During three daily 20-min sessions, the mouse head-bar was fixed to a post holding the mouse on the apex of the spherical treadmill. Ball motion was tracked by an IR camera taking pictures of the ball at 30 Hz. Eye motion was monitored at 15 Hz using a second IR camera imaging the reflection of the eye on an infrared dichroic mirror. Functional imaging was performed at 15 frames per second using a resonant scanning two-photon microscope (Neurolabware, West Hollywood, CA, United States) powered by a Ti-Sapphire Ultra-2 laser (Coherent, Santa Clara, CA, United States) set at 910 nm. The microscope scanning mirrors were hermetically sealed in a chamber to bring the scanning hum below the room ambient noise (&#x003C;59 dBA). The laser beam was focused 200 microns below the cortical surface using a 16&#x00D7;, 0.8 NA Nikon water-immersion objective. The objective was tilted 30&#x00B0; such that the objective lens was parallel to the dura surface. Laser power was kept below 70 mW. Frames (512 &#x00D7; 796 pixels) were acquired using the software Scanbox developed by Neurolabware.</p>
</sec>
<sec id="S2.SS2.SSS2">
<title>Na&#x00EF;ve Imaging Session</title>
<p>Mice were placed head fixed in the functional imaging rig in front of a screen such that it covered the visual field of the right eye, contralateral to the craniotomy. Visual stimuli of the test block consisted of the presentation of one of two vertical sinewave gratings that drifted toward the right and were rotated clockwise by 45&#x00B0; and 135&#x00B0; (temporal frequency = 2 Hz, spatial frequency = 0.04 cycle per degree, contrast = 75%; duration: 3 s; intertrial interval: 3 s). Visual cues were presented in a pseudorandom order, such as the same stimulus could not be presented more than three times in a row. At the end of the imaging session, after a break of at least 5 min, we assessed the orientation tuning of the imaged neurons by presenting an orientation tuning block that consisted of the presentation of a series of drifting sinewave gratings (12 orientations evenly spaced by 30&#x00B0; and randomly permuted). The spatiotemporal parameters of the orientation tuning stimuli were identical to those for the task block except for their duration (temporal frequency = 2 Hz, spatial frequency = 0.04 cycle per degree, contrast = 75%; duration: 1.5 s; intertrial interval: 3 s). Auditory stimuli consisted of the presentation of one of two sine wave pure tones (10 kHz and 5 kHz; 78 dB; duration: 3 s). Each audiovisual trial resulted from the random combination of one of the two pure tones with one of the two drifting gratings (four possibilities: 5 kHz tone + 45&#x00B0; drifting grating, 10 kHz tone + 45&#x00B0; drifting grating, 5 kHz tone + 135&#x00B0; drifting grating, and 10 kHz tone + 135&#x00B0; drifting grating). As scanning was not synced to the stimuli, a photodiode located at the top left corner of the screen was used to detect the exact timing of the visual stimulus onset and offset. The photodiode signal was acquired along with the following signals: (1) a signal provided by the two-photon microscope, which indicated the onset of each frame, and (2) two analog signals encoding the orientation of the drifting grating. These signals were digitized (NiDAQ, National Instruments, Austin, TX, United States) and recorded with the software WinEDR (John Dempster, University of Strathclyde). Imaging sessions started by recording one thousand frames with the green and red channels. The red channel was used to exclude GABAergic neurons from the analysis.</p>
</sec>
<sec id="S2.SS2.SSS3">
<title>Behavioral Training</title>
<p>After the na&#x00EF;ve recording session, mice were water-deprived up to 85% of their body weight and acclimated to head fixation on a spherical treadmill in custom-built, soundproof training rigs. Each rig was equipped with a monitor (Dell), a water dispenser with a built-in lickometer (to monitor licking, infrared beam break). Data acquisition boards (National Instruments and Arduino) were used to actuate water delivery and vacuum reward retrieval as well as monitor animal licking. The monitor and data acquisition boards were connected to a computer that ran the custom-made training program scripted in MATLAB (MathWorks, Natick, MA, United States). Once animals reached the target weight and were acclimated to the training setup, they were trained to perform the orientation discrimination task. In this task, drifting sine-wave gratings oriented 45&#x00B0; below the vertical were paired with a water reward, and the animal was expected to lick (Go). Drifting gratings orthogonal to the Go stimulus signaled the absence of reward, and the animal was expected to withhold licking (NoGo, orientation 135&#x00B0;) during those trials. When the stimulus instructed the animal to lick, the water delivery had to be triggered by the mouse licking during the third second of the stimulus presentation. No water was dispensed in the no-lick condition or if the mouse failed to trigger water delivery in the lick condition. If the animal responded correctly [Hit or Correct Rejection (CR)], the intertrial interval was 3 s. If the animal responded incorrectly [Miss or False Alarm (FA)], the intertrial interval was increased to 9.5 s as negative reinforcement. Animals were considered experts if their performance during training sessions was greater than 1.7 (probability of chance behavior &#x003C;0.1%, Monte Carlo simulation; <xref ref-type="bibr" rid="B9">Einstein et al., 2017</xref>). All the mice were also trained to perform the same task but using the 5 kHz tone as a NoGo cue and the 10 kHz tone as the Go cue. Half of the mice started training with the visual task while the other half started training with the auditory task. The order of training did not impact the mice&#x2019;s performance.</p>
</sec>
<sec id="S2.SS2.SSS4">
<title>Recording Sessions</title>
<p>As training was performed on a training setup located in a different room, trained mice were habituated to perform the task on the imaging setup (typically for one or two sessions) until they could perform the task above the expert threshold. Recording sessions consisted of five blocks (see <xref ref-type="fig" rid="F5">Figure 5D</xref>). The first block contained unimodal (either visual or auditory; the modality was selected at random at the beginning of the session) and was followed by an audiovisual block. For trained mice, the reward in this audiovisual block was associated to the modality (visual or auditory) of the preceding unimodal block. Hence if the first block was unimodal visual, the second block was audiovisual with the visual cue indicating the presence or absence of reward. The third block was a unimodal block (visual if the first unimodal block was auditory, auditory if the first unimodal block was visual). The fourth block was an audiovisual block (same rule for the reward as for the second block). The last block was an orientation tuning block, consisting in the presentation in pseudorandom order of twelve evenly spaced oriented visual stimuli (0&#x00B0;, 30&#x00B0;, 60&#x00B0;, 90&#x00B0;, 120&#x00B0;, 150&#x00B0;, 180&#x00B0;, 210&#x00B0;, 240&#x00B0;, 270&#x00B0;, 300&#x00B0;, 330&#x00B0;; the 12 orientations needed to be presented before starting a new series).</p>
</sec>
</sec>
<sec id="S2.SS3">
<title>Data Analysis</title>
<p>All the analyses detailed below were performed using custom MATLAB scripts.</p>
<sec id="S2.SS3.SSS1">
<title>Imaging Data Pre-processing</title>
<p>Calcium imaging frames were realigned offline to remove movement artifacts using the Scanbox algorithm (Neurolabware). A region of interest (ROI) was determined for each neuron using a semi-automatic segmentation routine. For every frame, the fluorescence level was averaged across the pixels of the ROI. Potential contamination of the soma fluorescence by the local neuropil was removed by subtracting the mean fluorescence of a 2&#x2013;5 &#x03BC;m ring surrounding the neuron&#x2019;s ROI, excluding the soma of neighboring neurons, and then adding the median value across time of the subtracted background. We then computed the fractional fluorescence from the background-subtracted fluorescence data. The fractional fluorescence (&#x0394;<italic>F</italic>/<italic>F</italic> = (<italic>F</italic>&#x2013;<italic>F</italic><sub>0</sub>)/<italic>F</italic><sub>0</sub>), was calculated with <italic>F</italic><sub>0</sub> defined as the median of the raw fluorescence measured during every inter-trial interval. The response of neurons to a trial was then measured as the mean fractional fluorescence measured during the first second of the visual stimulus presentation minus the mean fractional fluorescence measured during the second and a half preceding the stimulus presentation. The orientation tuning curve of each neuron was computed using a resampling-based Bayesian method (<xref ref-type="bibr" rid="B5">Cronin et al., 2010</xref>; <xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>) from the area under the curve of the fractional fluorescence responses recorded during the different trials of the tuning curve blocks. The preferred orientation was defined as the peak of the orientation tuning curve. When a neuron was not direction selective (i.e., responding equally to the same oriented stimulus moving in opposite directions), the preferred orientation was defined as the orientation included in the range (0&#x00B0;&#x2013;180&#x00B0;). The responses of all the neurons to all the trials as well as the neurons&#x2019; tuning curve parameters were stored in a SQL database.</p>
</sec>
</sec>
<sec id="S2.SS4">
<title>Shallow Neural Networks</title>
<sec id="S2.SS4.SSS1">
<title>Rationale for the Choice of Shallow Neural Networks</title>
<p>The SNN approach was selected to be as biologically plausible as possible (criterium #1). Indeed, the structure of neural networks was inspired by the computational structure of the visual cortex (<xref ref-type="bibr" rid="B11">Fukushima, 1980</xref>), and it was shown that neural networks provide pertinent computational models of the visual cortex (<xref ref-type="bibr" rid="B22">Lindsay, 2021</xref>). Therefore, we can assume that the output of the SNN in this study might have a biological relevance. We also wanted the classifier to work on the entire recorded population and therefore not to require the selection of &#x201C;active&#x201D; neurons (criterium #2). Indeed, the activity of cortical neurons follows a long-tailed gamma or log-normal distribution (<xref ref-type="bibr" rid="B7">Decharms and Zador, 2000</xref>; <xref ref-type="bibr" rid="B42">Wohrer et al., 2013</xref>). As a result, most neurons&#x2019; evoked activity is very similar to their resting state and only a few neurons significantly increase their firing rate when a visual stimulus is presented (<xref ref-type="bibr" rid="B1">Barth and Poulet, 2012</xref>; <xref ref-type="bibr" rid="B42">Wohrer et al., 2013</xref>). Therefore, simple strategies of analysis such as averaging the activity across recorded neurons have limited interpretative power, as they are poorly sensitive to the change of activity of the minority of neurons responding to the stimulus. A common strategy used to circumvent this issue is to determine a threshold above which neurons are considered &#x201C;active.&#x201D; However, this approach reveals limitations when working on the modulation by the behavioral context of the V1 population activity. Indeed, those extrinsic modulatory factors inactivate some neurons while activating others (<xref ref-type="bibr" rid="B20">Iurilli et al., 2012</xref>; <xref ref-type="bibr" rid="B19">Ibrahim et al., 2016</xref>; <xref ref-type="bibr" rid="B25">Meijer et al., 2017</xref>; <xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>), complicating the comparison of the different &#x201C;active&#x201D; populations responding in the different behavioral contexts. To investigate sensory representations, some other analysis strategies such as dimensionality reduction (<xref ref-type="bibr" rid="B6">Cunningham and Yu, 2014</xref>; <xref ref-type="bibr" rid="B4">Carrillo-Reid et al., 2019</xref>) or decoding (<xref ref-type="bibr" rid="B34">Quian Quiroga and Panzeri, 2009</xref>; <xref ref-type="bibr" rid="B38">Stringer et al., 2021</xref>) avoid selecting neurons. These approaches capture the availability of the information about the stimulus feature embedded in the population activity using diverse metrics of statistical distance between different arrays of data. However, because the result of those computations is abstract, those strategies can only provide limited insights about the implementation by the biological networks of the computations realized by those methods. Indeed, if those techniques inform us about <italic>what</italic> information is present and can point out discrepancies between what an optimal decoder and an animal can discriminate (<xref ref-type="bibr" rid="B38">Stringer et al., 2021</xref>), they do not directly tackle the question of <italic>how</italic> the available information is used by the biological system. This is not the case of the SNN whose output allows taking advantage of the continuity of the orientation space to determine the representation in V1 of visual stimuli that do not belong to the classifier training stimuli. Finally, the simplicity of SNNs (i.e., their shallowness) allows to straightforwardly access the weights assigned to each recorded neuron in the classifying decisions. Here, we chose the <italic>Connection Weight Approach</italic> (<xref ref-type="bibr" rid="B31">Olden and Jackson, 2002</xref>; <xref ref-type="bibr" rid="B32">Olden et al., 2004</xref>) to determine which neurons carry the most weight in the classification of the visual stimulus.</p>
</sec>
<sec id="S2.SS4.SSS2">
<title>Implementation</title>
<p>The shallow neural network was a two-layer feedforward network with a sigmoid transfer function in the hidden layer and a softmax transfer function in the output. It was generated in MATLAB using the <italic>patternnet</italic> function. The input layer was made of 250 <italic>computational neurons</italic> that receive the evoked activity of 250 <italic>cortical neurons</italic>, the single hidden layer was made of 10 hidden computational neurons, and the output layer was composed of 12 computational neurons corresponding to the 12 orientations of the tuning block. The input layer was connected to 250 cortical neurons randomly selected in either the na&#x00EF;ve or trained mouse database using an SQL query. This sampling method pools cortical neurons from different mice. Pooling neurons across mice breaks the correlational structures between neurons. However, those correlations were found to have little influence on sensory information encoding. Indeed, although correlations are prevalent in the visual cortex, the additional information they provide is small (about 4% in an analysis window greater than 100 ms; <xref ref-type="bibr" rid="B16">Golledge et al., 2003</xref>) and offset by the redundancy arising from the neurons&#x2019; similar tuning properties (<xref ref-type="bibr" rid="B27">Montani et al., 2007</xref>). As a result, the spaces encoding sensory and behavioral variables are essentially orthogonal (<xref ref-type="bibr" rid="B39">Stringer et al., 2019</xref>; <xref ref-type="bibr" rid="B36">Rumyantsev et al., 2020</xref>). As the cortical neurons used for each SNN were selected from different recording sessions, the SNNs were trained using resampled trials from the tuning block (100 resampled trials for each of the 12 orientations). Those resampled trials consisted in the random selection for each selected cortical neuron of one trial corresponding to the presentation of that stimulus. The network was trained by scaled conjugate gradient backpropagation using the <italic>trainscg</italic> MATLAB function. For the SNN training, the data was split into training, validation, and test sets: 70% for training; 15% for cross-validation to validate that the network is generalizing and to stop training before overfitting, and 15% to independently test network generalization.</p>
</sec>
<sec id="S2.SS4.SSS3">
<title>Cortical Neuron&#x2019;s Connection Weights</title>
<p>We evaluated the relative weight of the cortical neurons connected to each input computational neuron using the <italic>Connection Weight Approach</italic> (<xref ref-type="bibr" rid="B31">Olden and Jackson, 2002</xref>; <xref ref-type="bibr" rid="B32">Olden et al., 2004</xref>). First, input-hidden-output connection weights were obtained as the product of input-hidden and hidden-output connection weights for each input and hidden computational neuron; then overall connection weights were defined as the sum of the input-hidden-output connection weights for each input variable (<xref ref-type="bibr" rid="B31">Olden and Jackson, 2002</xref>). This approach that uses raw input-hidden and hidden-output connection weights in the neural network provides the best methodology for accurately quantifying variable importance (<xref ref-type="bibr" rid="B32">Olden et al., 2004</xref>).</p>
</sec>
</sec>
<sec id="S2.SS5">
<title>Statistics</title>
<sec id="S2.SS5.SSS1">
<title>Permutation Tests</title>
<p>To determine if the mean across trials computed from two different pools was significantly different, we compared the value obtained from the distribution of 1,000 or 10,000 differences obtained when the pool labels were shuffled. The two-tailed confidence intervals of the null hypothesis at the alpha level 0.05 were defined as the 2.5 and 97.5 percentile of the distribution obtained from the permutations. The difference between the observed means was considered significant if located outside the confidence interval of the null distribution.</p>
</sec>
<sec id="S2.SS5.SSS2">
<title>Circular Statistics</title>
<p>Circular statistics were computed with the Circular Statistics Toolbox for MATLAB (<xref ref-type="bibr" rid="B2">Berens, 2009</xref>).</p>
</sec>
</sec>
</sec>
<sec id="S3" sec-type="results">
<title>Results</title>
<sec id="S3.SS1">
<title>Representation of the Visual Stimulus Orientation in the Na&#x00EF;ve Mouse V1</title>
<p>One of the main goals of this study was to determine the relevance of using an SNN to assess how the orientations of drifting gratings were represented at the population level by V1 layer 2/3 (L2/3) neurons. To test this approach, we used a dataset of two-photon calcium imaging experiments in which mice placed on a spherical treadmill in front of a screen and a speaker were shown visual, auditory, and audiovisual stimuli (<xref ref-type="fig" rid="F1">Figure 1A</xref>). During the recording sessions, three types of stimulus blocks were presented (<xref ref-type="fig" rid="F1">Figure 1B</xref>): unimodal blocks consisting of either visual or auditory stimuli (45&#x00B0; and 135&#x00B0; drifting gratings or auditory: 5 kHz or 10 kHz sinewave tones, respectively), an audiovisual block (where the two visual and the two auditory cues were randomly paired), and a tuning block (during which series of 12 different drifting gratings were presented). The first block of the session was a unimodal block either visual or auditory, followed by an audiovisual block. Then, the alternate unimodal block was presented (either auditory or visual, respectively) followed by a second audiovisual block (<xref ref-type="fig" rid="F1">Figure 1C</xref>). Each recording session ended with the presentation of a tuning block to allow determining the tuning-curves of the imaged neurons (<xref ref-type="fig" rid="F1">Figure 1C</xref>). Calcium imaging was performed while simultaneously tracking the locomotion and the pupil size (<xref ref-type="fig" rid="F1">Figures 1D,E</xref>) as locomotor activity and arousal (correlated to the pupil size) modulate the neuronal response of V1 neurons (<xref ref-type="bibr" rid="B29">Niell and Stryker, 2010</xref>; <xref ref-type="bibr" rid="B33">Polack et al., 2013</xref>; <xref ref-type="bibr" rid="B40">Vinck et al., 2015</xref>). We had already analyzed this database in a previous study and shown that sound modulation improves the representation of the orientation and direction of the visual stimulus in V1 L2/3 (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). We had also shown that arousal and locomotion are similar in the unimodal and audiovisual blocks in this dataset (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). The analytic method applied in that study was a thresholding approach used to determine which neurons were included in the analysis. In this study, we wanted to use a method that allows determining how the V1 population is representing the orientation of the visual stimulus without having to select &#x201C;responsive neurons&#x201D; in the recorded neuronal population. We decided to test SNNs as they are very effective for pattern classification (<xref ref-type="bibr" rid="B11">Fukushima, 1980</xref>) and therefore were a good candidate to identify the population activity patterns evoked by specific oriented stimuli. For this, we trained single hidden layer SNNs to identify the neuronal patterns evoked by 12 different drifting gratings evenly spaced in the orientation space (note that we will use the term &#x201C;orientation&#x201D; to indicate both the orientation and drifting direction of the gratings). The SNN output estimates the probability that the presented pattern belongs to each of the output categories. Hence, a SNN presented with a pattern that it has been trained to identify, theoretically returns an output of 1 for the corresponding category (30&#x00B0; and 60&#x00B0; in the example shown in <xref ref-type="fig" rid="F2">Figures 2A,B</xref>). Because of the continuity of the orientation space, we then assumed that the presentation of an oriented stimulus equidistant from two trained orientations such as 45&#x00B0; (which is equidistant from 30&#x00B0; and 60&#x00B0;) would be classified as 50% &#x201C;30&#x00B0; drifting grating&#x201D; and 50% &#x201C;60&#x00B0; drifting grating&#x201D; (<xref ref-type="fig" rid="F2">Figure 2B</xref>), as it activates a subpopulation of neurons responding both to the 30&#x00B0; and 60&#x00B0; stimuli (<xref ref-type="fig" rid="F2">Figure 2A</xref>). Hence, we trained SNNs to classify the neural patterns of subpopulations of V1 neurons (250 cortical neurons randomly picked in a database of 1,353 imaged neurons in eight mice). Each SNN was composed of an input layer of 250 computational neurons fully connected to a layer of 10 hidden computational neurons which were in turn fully connected to an output layer of 12 computational neurons (<xref ref-type="fig" rid="F2">Figure 2C</xref>). The training of the network was performed using 100 resampled trials for each of the twelve evenly spaced oriented visual stimuli presented during the tuning block (0&#x00B0;, 30&#x00B0;, 60&#x00B0;, 90&#x00B0;, 120&#x00B0;, 150&#x00B0;, 180&#x00B0;, 210&#x00B0;, 240&#x00B0;, 270&#x00B0;, 300&#x00B0;, 330&#x00B0;). Each resampled trial for an orientation corresponded to the random selection for each cortical neuron of one response to the presentation of that stimulus (mean &#x0394;F/F across the visual stimulus presentation). Once trained, the SNNs were able to accurately classify all the resampled trials (probability of correct classification with cross-validation &#x003E; 0.99). We then used the trained SNNs to classify 100 resampled trials collected when presenting visual stimuli of the unimodal block, i.e., drifting gratings having orientations that the SNNs were not trained to recognize (45&#x00B0; and 135&#x00B0;; <xref ref-type="fig" rid="F2">Figure 2D</xref>). The 100 outputs of each SNN were averaged and the circular mean of this mean output provided the orientation estimated by the SNN (<xref ref-type="fig" rid="F2">Figure 2E</xref>). The accuracy (precision index) of this represented orientation was measured as the projection of the circular mean vector onto the radial axis of the visual stimulus orientation (<xref ref-type="fig" rid="F2">Figure 2F</xref>). To determine the variability of the representation across the population of imaged V1 neurons, we repeated the analysis hundreds of times, creating each time a new SNN from a new pseudo-population of 250 V1 neurons.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p>Recording sessions in na&#x00EF;ve mice. <bold>(A)</bold> Schematic representation of the recording setup. <bold>(B)</bold> Stimuli presented during the unimodal, audiovisual, and tuning blocks. <bold>(C)</bold> Organization of the unimodal, audiovisual, and tuning blocks in a recording session. <bold>(D)</bold> Two-photon image of V1 neurons recorded in a na&#x00EF;ve mouse. The preferred orientation of the segmented neurons is indicated by the color scale in inset. White contours indicate segmented neurons that were not estimated orientation selective by the algorithm. <bold>(E)</bold> Example of the activity of neurons indicated in panel <bold>(D)</bold> during the presentation of the visual stimuli of the audiovisual block. The neuronal activity was recorded simultaneously with the locomotion of the animal as well as the pupil size (bottom traces).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-869705-g001.tif"/>
</fig>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p>Analysis of orientation representation in V1 using shallow neural networks. <bold>(A)</bold> Schematic representation of the distribution of the preferred orientations of V1 neurons. Each dot represents a neuron. The hue represents the preferred orientation of the neuron, and the hue intensity represents the amplitude of the neuronal response to the presentation of its preferred orientation. <bold>(B)</bold> Schematic representation of the response of a SNN trained to discriminate the V1 population response to the presentation of a 30&#x00B0; drifting grating (top panel) and to the presentation of a 60&#x00B0; drifting grating (middle panel). When the SNN is provided with the response of its input cortical neuron population for the presentation of a 45&#x00B0; drifting grating (bottom panel), its output should indicate an equal probability that the stimulus belongs to the &#x2018;30&#x00B0;&#x2019; and &#x2018;60&#x00B0;&#x2019; categories. <bold>(C)</bold> Schematic representation of the SNN training. <bold>(D)</bold> Schematic representation of the SNN testing with a 45&#x00B0; stimulus that does not belong to the training categories. <bold>(E)</bold> Average output across 100 trials of the presentation of a 45&#x00B0; drifting grating of a randomly selected trained SNN. The dot indicates the orientation and length of the circular mean vector computed from the mean distribution of the SNN output. <bold>(F)</bold> Precision index defined as the length of the vector resulting from the projection of the circular mean vector onto the axis of the visual stimulus orientation.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-869705-g002.tif"/>
</fig>
</sec>
<sec id="S3.SS2">
<title>Relationship Between Tuning Curves and Shallow Neural Network Weights</title>
<p>In an effort of comparison, our first goal was to assess whether the SNN would use the same features of the neurons&#x2019; response statistics that are captured by the traditional tuning curves. Therefore, we tested the hypothesis that the weight of the cortical neurons in the SNN decision corresponded to their orientation tuning properties. In a neural network, the relative contributions of the input variables to the predictive output depend primarily on the magnitude and direction of the connection weights between computational neurons. Input variables with larger connection weights represent greater intensities of signal transfer (<xref ref-type="bibr" rid="B31">Olden and Jackson, 2002</xref>). Therefore, they are more important in the prediction process compared to variables with smaller weights. To determine whether the orientation tuning of the cortical neurons would be a predictor of their connection weight in the SNN, we first determined the preferred orientation, orientation selectivity index (OSI) and direction selectivity index (DSI) for all the neurons of the dataset. For each neuron, the responses to the different visual stimuli of the tuning block (<xref ref-type="fig" rid="F3">Figure 3A</xref>) were fitted using a resampling-based Bayesian method (<xref ref-type="bibr" rid="B5">Cronin et al., 2010</xref>; <xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>; <xref ref-type="fig" rid="F3">Figures 3B,C</xref>). We then estimated the weights of every input cortical neuron for each of the 12 SNN outputs (corresponding to the 12 visual stimuli of the tuning block) using the <italic>Connection Weight Approach</italic> (<xref ref-type="bibr" rid="B31">Olden and Jackson, 2002</xref>; <xref ref-type="bibr" rid="B32">Olden et al., 2004</xref>), and repeated this measurement in 250 SNNs (250 inputs &#x00D7; 250 iterations = 62,500 datapoints). Finally, we sorted the cortical neurons by preferred orientation and displayed their connection weights for each of the 12 decision outputs (<xref ref-type="fig" rid="F3">Figure 3D</xref>). We found that the SNNs assigned the largest connection weights to neurons tuned to the visual stimulus presented (<xref ref-type="fig" rid="F3">Figures 3D,E</xref>). We then plotted the connection weights of cortical neurons as a function of their orientation selectivity (<xref ref-type="fig" rid="F3">Figure 3F</xref>) and direction selectivity indexes (<xref ref-type="fig" rid="F3">Figure 3G</xref>). Those two relationships were best fitted by an exponential curve indicating that cortical neurons with high orientation and/or direction selectivity had a much larger connection weights, and therefore a much larger impact in the SNN decision than most of the other cortical neurons, even though they represented only a fraction of the total neuronal pseudo-population in V1 (<xref ref-type="fig" rid="F3">Figures 3H,I</xref>). Hence, we show that SNNs classify the orientation of the visual stimuli of the tuning block by learning and using the orientation tuning properties of the V1 neurons.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p>SNN connection weights as a function of the preferred orientation, orientation selectivity and direction selectivity of the input neurons. <bold>(A)</bold> Activity of the neurons labeled in <xref ref-type="fig" rid="F1">Figure 1D</xref> during an 18-trial segment of the tuning block. <bold>(B)</bold> Tuning curves of the neurons 2 to 5 shown in panel <bold>(A)</bold>. Radial axis: area under the curve of the neuronal response. <bold>(C)</bold> Tuning curve of the cell 1 shown in panel <bold>(A)</bold>. Top: superimposition of the fractional fluorescence of the neuron for all the trials of the different stimuli of the tuning block. In red is the average of the fractional fluorescence across trials. Bottom: tuning curve (orange line) fitted to the data points (dots). Red crosses indicate the median value across trials. Inset: same tuning curve presented as a polar plot. <bold>(D)</bold> Connection weights (see color scale) of each input cortical neuron for the 12 SNN outputs (<italic>x</italic>-axis). The neurons were ranked by their preferred orientation (<italic>y</italic>-axis). Connection weights were normalized using a <italic>z</italic>-score normalization method. <bold>(E)</bold> Distribution of the preferred orientations in the input cortical neurons&#x2019; population. Presentation matching the presentation shown in panel <bold>(D)</bold>. <bold>(F)</bold> Distribution of the input neurons&#x2019; connection weights as a function of the V1 neurons&#x2019; orientation selectivity indexes. Dotted lines: exponential fit. <bold>(G)</bold> Distribution of the input neurons&#x2019; connection weights as a function of the V1 neurons&#x2019; orientation selectivity indexes. <bold>(H)</bold> Distribution of the orientation selectivity index in the recorded neuronal population. <bold>(I)</bold> Distribution of the direction selectivity index in the recorded neuronal population.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-869705-g003.tif"/>
</fig>
</sec>
<sec id="S3.SS3">
<title>Sound Modulation of Orientation Representation in Na&#x00EF;ve Mice V1</title>
<p>Once we had confirmed that SNNs were using the orientation tuning properties of the V1 neurons to classify the visual stimuli of the tuning block, we tested the hypothesis that the SNN approach could be used to determine how sound modulates the representation of the orientation representation in V1 L2/3. We trained 1,000 SNNs to classify the stimuli of the tuning block. Then, we presented the SNNs with the response of their input cortical neurons to the presentation of the 45&#x00B0; drifting gratings recorded during the unimodal visual block (average output of 100 resampled trials). The circular means of the 1,000 SNNs outputs were displayed on a polar plot (<xref ref-type="fig" rid="F4">Figure 4A</xref>, blue dots, see <xref ref-type="fig" rid="F2">Figures 2E,F</xref> for the approach). We repeated the same analysis for the 45&#x00B0; drifting gratings recorded during the audiovisual blocks when the visual stimulus was paired with the low tone (5 kHz, red) or the high tone (10 kHz, green). The same approach was used with the neuronal response to the presentation of the 135&#x00B0; drifting gratings (<xref ref-type="fig" rid="F4">Figure 4B</xref>; unimodal: blue, audiovisual 5 kHz: red, audiovisual 10 kHz, blue) and for the unimodal auditory tones (<xref ref-type="fig" rid="F4">Figure 4C</xref>). In the unimodal visual and audiovisual conditions, the output vectors of the SNNs were indicating the orientation of the visual stimulus (<xref ref-type="fig" rid="F4">Figures 4D&#x2013;F</xref>; circular mean &#x00B1; confidence interval; For 45&#x00B0;: unimodal visual: 42.3&#x00B0; &#x00B1; 1.0, visual + 5 kHz tone: 47.7&#x00B0; &#x00B1; 0.8, visual + 10 kHz tone: 48.0&#x00B0; &#x00B1; 0.7; For 135&#x00B0;: unimodal visual: 105.4&#x00B0; &#x00B1; 4.4, visual + 5 kHz tone: 123.0&#x00B0; &#x00B1; 2.7, visual + 10 kHz tone: 127.9&#x00B0; &#x00B1; 2.7), while in the unimodal auditory condition the output vectors were both similarly attracted toward 90&#x00B0;, i.e., at equidistance between 45&#x00B0; and 135&#x00B0; (5 kHz tone: 84.6&#x00B0; &#x00B1; 5.0, 5 kHz tone: 89.0&#x00B0; &#x00B1; 4.3; <xref ref-type="fig" rid="F4">Figure 4C</xref>). To determine how sounds modified the representation of the visual stimuli in V1, we compared the accuracy of the representation of the visual stimulus in the unimodal and audiovisual conditions by computing for each SNN the difference between the precision index (see <xref ref-type="fig" rid="F2">Figure 2F</xref>) obtained with the audiovisual responses and the precision index obtained with the unimodal responses. We then plotted the distribution of those differences as violin plots (<xref ref-type="fig" rid="F4">Figure 4G</xref>). We found that the precision index of the representation of the 45&#x00B0; and 135&#x00B0; stimuli was improved in the audiovisual conditions compared to the unimodal context (<xref ref-type="fig" rid="F4">Figure 4G</xref>; difference between 45&#x00B0; + 5 kHz and 45&#x00B0; unimodal; 5.4%; 45&#x00B0; + 10 kHz and 45&#x00B0; unimodal; 6.1%; 135&#x00B0; + 5 kHz and 135&#x00B0; unimodal; 7.6%; 135&#x00B0; + 10 kHz and 135&#x00B0; unimodal; 9.4%; <italic>p</italic> &#x003C; 0.0001 for all audiovisual combinations; random permutation test). We also compared the proportion of SNN outputs that changed direction (e.g., moving from a 225&#x00B0; output to 45&#x00B0;) when the visual stimulus was presented with one of the two sounds. We found that the representation of the stimulus direction was more accurate when the stimulus was presented with a tone (<xref ref-type="fig" rid="F4">Figure 4H</xref>; random permutation; <italic>p</italic> = 0.02, and <italic>p</italic> &#x003C; 0.0001, respectively for 45&#x00B0; and 135&#x00B0;combined with either tone). This improvement of the representation of the 45&#x00B0; and 135&#x00B0; visual stimuli was mainly due to the improvement of the SNNs that performed the worst, as illustrated by the quiver plots indicating how the SNNs outputs were modified by sound as a function of the output of the SNN for the unimodal stimulus (<xref ref-type="fig" rid="F4">Figures 4I,J</xref>, the base of the arrow corresponds to the unimodal stimulus while the arrowhead corresponds to the audiovisual stimulus). Altogether, those results replicated our previous findings that orientation is better represented in the V1 of na&#x00EF;ve mice when sounds are presented simultaneously with the oriented visual stimulus (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>).</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p>Sound modulation of the V1 population evoked response in na&#x00EF;ve mice. <bold>(A)</bold> Output of 1,000 SNNs made of 250 randomly selected V1 neurons to the presentation of a 45&#x00B0; drifting grating in the unimodal (blue) and audiovisual (5 kHz tone: red; 10 kHz: green) contexts. The orange arrow indicates the orientation of the presented stimulus. <bold>(B)</bold> Same representation as in panel <bold>(A)</bold> for the presentation of the 135&#x00B0; drifting grating. <bold>(C)</bold> Same representation as in panel <bold>(A)</bold> for the presentation of unimodal auditory stimuli (5 kHz tone, dark red; 10 kHz, dark green). <bold>(D)</bold> Distribution of the orientations indicated by the output circular mean of the 1,000 SNNs shown in panels <bold>(A,B)</bold> when the input is the neuronal activity evoked by the unimodal 45&#x00B0; (blue) and unimodal 135&#x00B0; (red) drifting grating. <bold>(E)</bold> Same representation as in panel <bold>(D)</bold> when the visual stimulus is paired with the 5 kHz tone. <bold>(F)</bold> Same representation as in panel <bold>(D)</bold> when the visual stimulus is paired with the 10 kHz tone. <bold>(G)</bold> Modulation by the 5 kHz (red background) and 10 kHz (green background) sounds of the precision indexes of the 1,000 SNN shown in panels (<bold>A,B)</bold>. The black bar indicates the mean of the distribution. <sup>&#x002A;&#x002A;&#x002A;</sup> Random permutation test (<italic>p</italic> &#x003C; 0.0001). <bold>(H)</bold> Comparison of the percent of SNN outputs shown in panels <bold>(A,B)</bold> changing direction when the visual stimulus was presented with one of the two sounds in panels <bold>(B,C)</bold> with the probability distribution of the same measure performed 10,000 times with shuffled data (from left to right: <italic>p</italic> = 0.02, <italic>p</italic> = 0.02, <italic>p</italic> &#x003C; 0.0001, <italic>p</italic> &#x003C; 0.0001). <bold>(I)</bold> Quiver plot of topographically clustered modulation vectors illustrating how SNNs with similar outputs for the presentation of the 45&#x00B0; drifting grating are modulated by the 5 kHz (red) and 10 kHz (green) sounds. <bold>(J)</bold> Same representation as in <bold>(I)</bold> for the presentation of the 135&#x00B0; drifting grating.</p></caption><graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-869705-g004.tif"/></fig>
</sec>
<sec id="S3.SS4">
<title>Sound Modulation of Orientation Representation in Mice Performing an Audiovisual Task</title>
<p>We then tested the hypothesis that the modulation of the representation of the orientation of visual stimuli by sounds depended on the relative importance of the auditory and visual stimuli for the completion of the task. To test this hypothesis, we used a new database in which mice were performing an audiovisual discrimination task using the same stimuli as the one presented to the na&#x00EF;ve mice. Water-restricted mice were placed on the same apparatus as the na&#x00EF;ve mice, but this time a lickometer was placed in front of their mouths (<xref ref-type="fig" rid="F5">Figure 5A</xref>). Mice were successfully trained at performing the unimodal visual and the unimodal auditory Go/NoGo task (<xref ref-type="fig" rid="F5">Figure 5B</xref>, the training order was randomly assigned). For those two tasks, mice were presented with a Go cue (for the visual task a 45&#x00B0; drifting grating; for the auditory task a 10 kHz tone) or a NoGo cue (visual task: 135&#x00B0;; auditory task: 5 kHz tone). The stimulus was presented for 3 s. Mice had to lick to obtain a reward when the Go signal was presented, and to withhold licking during the NoGo signal. The response window corresponded to the third second of the stimulus (<xref ref-type="fig" rid="F5">Figure 5C</xref>). Once trained to the first unimodal task, mice were trained to the other unimodal task. Then, when the expert level (<italic>D</italic>&#x2032; &#x003E; 1.7) at those two unimodal tasks was reached, mice were habituated to the audiovisual context (<xref ref-type="fig" rid="F5">Figure 5C</xref>). Each session of the audiovisual task started with a unimodal block (either visual or auditory) followed by an audiovisual block during which the modality of the preceding unimodal block was predicting the reward (<xref ref-type="fig" rid="F5">Figure 5D</xref>). This first audiovisual block was followed by a unimodal block using the other modality (either auditory or visual, respectively) then a second audiovisual block during which the modality of the second unimodal block was used to dispense the reward. To perform perfectly at the task, mice would have to perform a modality-specific attention task (attend visual for the first two blocks then auditory for the last two blocks in the example provided in <xref ref-type="fig" rid="F5">Figure 5D</xref>). Our analysis of the mouse behavior showed that mice used an alternate strategy (<xref ref-type="fig" rid="F5">Figure 5E</xref>). Indeed, they licked whenever one of the Go cues (auditory or visual) was presented, regardless of the identity of the rewarded modality for the current block (auditory rewarded block lick rate (median &#x00B1; m.a.d.): Go<sub><italic>v</italic></sub>-NoGo<sub><italic>a</italic></sub>: 71 &#x00B1; 12%; NoGo<sub><italic>v</italic></sub>-Go<sub><italic>a</italic></sub>: 92 &#x00B1; 5%; visual rewarded block lick rate: Go<sub><italic>v</italic></sub>-NoGo<sub><italic>a</italic></sub>: 94 &#x00B1; 5%; NoGo<sub><italic>v</italic></sub>-Go<sub><italic>a</italic></sub>: 68 &#x00B1; 20%). Therefore, we sorted the data by presented stimuli, pooling together audiovisual blocks where different modalities were rewarded. In the unimodal condition, mice licked systematically whenever the Go signal was presented (unimodal auditory hit rate: 92%, <italic>n</italic> = 10 mice, Wilcoxon test: <italic>p</italic> &#x003C; 0.0001; unimodal visual hit rate: 96%, <italic>n</italic> = 10 mice; Wilcoxon test: <italic>p</italic> &#x003C; 0.0001), and avoided licking in the presence of the NoGo signal (unimodal auditory False Alarm (FA) rate: 29%, <italic>n</italic> = 10 mice, Wilcoxon test: <italic>p</italic> &#x003C; 0.0001; unimodal visual FA rate: 32%, <italic>n</italic> = 10 mice, Wilcoxon test: <italic>p</italic> = 0.0004). In the audiovisual blocks, our analysis of the mouse behavior showed that mice used an alternative strategy. The performance of the mice at refraining from licking was improved when the auditory and visual NoGo cues were presented simultaneously, compared to the unimodal NoGo conditions (audiovisual NoGo FA rate: 19%, <italic>n</italic> = 10 mice, audiovisual NoGo vs. unimodal auditory NoGo: Wilcoxon test <italic>p</italic> = 0.0100; audiovisual NoGo vs. unimodal visual NoGo: Wilcoxon test <italic>p</italic> = 0.0009). We did not find an improvement in behavioral performance when the two Go signals were presented together, compared to the two unimodal conditions (hit rate 98%, <italic>n</italic> = 10 mice; audiovisual Go vs. unimodal auditory Go: Wilcoxon test <italic>p</italic> = 0.8109; audiovisual Go vs. unimodal visual Go: Wilcoxon test <italic>p</italic> = 0.2415), likely because mice already performed almost perfectly in the unimodal contexts. When the visual and auditory-visual cues were in conflict, mice clearly chose to lick (Go<sub><italic>visual</italic></sub>/NoGo<sub><italic>auditory</italic></sub>: hit rate = 79%, <italic>n</italic> = 10 mice; Wilcoxon test: <italic>p</italic> &#x003C; 0.0001; NoGo<sub><italic>auditory</italic></sub>/Go<sub><italic>visual</italic></sub>: hit rate = 82%, <italic>n</italic> = 10 mice; Wilcoxon test: <italic>p</italic> &#x003C; 0.0001). Hence, when the signal was conflicting (e.g., Go visual paired with NoGo auditory), mice licked by default (<xref ref-type="fig" rid="F5">Figure 5E</xref>). The apparent strategy of the animals was to seek the Go cue regardless of the modality, ignoring the current reward contingencies. Instead of the intended modality-specific attention Go/NoGo task, they engaged with the task as a cross-modal Go detection task (<xref ref-type="fig" rid="F5">Figure 5F</xref>).</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption><p>Recording sessions in trained mice. <bold>(A)</bold> Schematic representation of the recording setup. <bold>(B)</bold> Potential behavioral outcomes for a trial. <bold>(C)</bold> Trial time course. Inter Trial interval (I.T.I) after Hit and CR: 3 s; I.T.I after Miss and FA: 12.5 s. <bold>(D)</bold> Stimuli shown during the unimodal, audiovisual, and tuning blocks in a session starting with visual reward. Water drops indicate stimuli that are rewarded for that block. <bold>(E)</bold> Mouse licking probability for the different unimodal and audiovisual stimuli. <bold>(F)</bold> Activity of neurons during the presentation of the visual stimuli of the audiovisual block (auditory reward). The neuronal activity was recorded simultaneously with the licking activity (top trace), the locomotion of the animal as well as the pupil size (bottom traces).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-869705-g005.tif"/></fig>
<p>Using the SNN approach, we compared the representation of the Go and NoGo visual cues in the unimodal and audiovisual contexts (<xref ref-type="fig" rid="F6">Figures 6A,B</xref>). The orientation of the SNNs output vectors were similar for the unimodal and audiovisual blocks (circular mean &#x00B1; confidence interval; For 45&#x00B0;: unimodal visual: 61.7&#x00B0; &#x00B1; 4.3, visual + 5 kHz tone: 58.7&#x00B0; &#x00B1; 2.7, visual + 10 kHz tone: 57.8&#x00B0; &#x00B1; 2.8; <xref ref-type="fig" rid="F6">Figure 6A</xref>; For 135&#x00B0;: unimodal visual: 130.5&#x00B0; &#x00B1; 2.3, visual + 5 kHz tone: 128.7&#x00B0; &#x00B1; 2.9, visual + 10 kHz tone: 133.1&#x00B0; &#x00B1; 3.1; <xref ref-type="fig" rid="F6">Figure 6B</xref>). The precision of the representation of the visual Go signal (see <xref ref-type="fig" rid="F2">Figure 2F</xref>) was slightly but significantly improved by sound (<xref ref-type="fig" rid="F6">Figure 6C</xref>; difference between 45&#x00B0; + 5 kHz and 45&#x00B0; unimodal; 2.9%, <italic>p</italic> = 0.003; 45&#x00B0; + 10 kHz and 45&#x00B0; unimodal; 3.2%; <italic>p</italic> = 0.002, random permutation test). On the contrary, the representation of the NoGo signal was significantly less precise in the audiovisual context (<xref ref-type="fig" rid="F6">Figure 6C</xref>, 135&#x00B0; + 5 kHz and 135&#x00B0; unimodal; &#x2212;19.8%, <italic>p</italic> &#x003C; 0.0001; 135&#x00B0; + 10 kHz and 135&#x00B0; unimodal; &#x2212;17.1%; <italic>p</italic> &#x003C; 0.0001; random permutation test). This opposite modulation of the Go and NoGo orientation representation was associated with a comparable change of the representation of the direction of the drifting grating with a significant improvement of the representation of the direction of the Go drifting grating in the audiovisual context, and a deterioration of the representation of the direction of the NoGo drifting grating with sound (<xref ref-type="fig" rid="F6">Figure 6D</xref>). The differential modulation of the Go and NoGo cue representation by sound was particularly salient in quiver plots as most of the improvement of the Go cue representation was carried by SNNs having a poor accuracy in the unimodal context (<xref ref-type="fig" rid="F6">Figure 6E</xref>), while most of the modulation for the NoGo visual cue representation was due to highly accurate SNNs that saw their performance decrease in the audiovisual context (<xref ref-type="fig" rid="F6">Figure 6F</xref>). Altogether our results suggest that sounds can have a bidirectional impact on the orientation representation accuracy in V1, as the modulation interacted with the way the animals engaged in the task. For the sought-after Go visual stimulus, sound potentiated the orientation representation, while it degraded the representation of the NoGo visual stimulus that the animals tended to ignore.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption><p>Sound modulation of the V1 population evoked response in mice performing an audiovisual detection task. <bold>(A)</bold> Output of 1,000 SNNs made of 250 randomly selected V1 neurons to the presentation of a 45&#x00B0; drifting grating (Go cue) in the unimodal (blue) and audiovisual (5 kHz tone: red, NoGo cue; 10 kHz: green, Go cue) contexts. The orange arrow indicates the orientation of the presented stimulus. <bold>(B)</bold> Same representation as in panel <bold>(A)</bold> for the presentation of the 135&#x00B0; drifting grating (NoGo cue). <bold>(C)</bold> Modulation by the 5 kHz (red background, NoGo cue) and 10 kHz (green background, Go cue) sounds of the precision indexes of the 1,000 SNNs shown in panels <bold>(A,B)</bold>. The black bar indicates the mean of the distribution. Random permutation test (<sup>&#x002A;&#x002A;</sup><italic>p</italic> = 0.002; <sup>&#x002A;&#x002A;&#x002A;</sup><italic>p</italic> &#x003C; 0.0001). <bold>(D)</bold> Comparison of the proportion of SNN outputs shown in panels <bold>(A,B)</bold> changing direction when the visual stimulus was presented with one of the two sounds in panels <bold>(B,C)</bold> together with the probability distribution of the same measure performed 10,000 times with shuffled data (for all panels: <italic>p</italic> &#x003C; 0.0001). <bold>(E)</bold> Quiver plot of topographically clustered modulation vectors illustrating how SNNs with similar outputs for the presentation of the 45&#x00B0; drifting grating (Go cue) are modulated by the 5 kHz (red, NoGo) and 10 kHz (green, Go) sounds. <bold>(F)</bold> Same representation as in panel <bold>(E)</bold> for the presentation of the 135&#x00B0; drifting grating (NoGo cue).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-869705-g006.tif"/></fig>
</sec>
</sec>
<sec id="S4" sec-type="discussion">
<title>Discussion</title>
<p>In this study, our goal was to test the SNN approach as a tool to analyze the representation of the orientation of drifting gratings by the V1 neuronal population. As an example, and to provide a comparison with a more traditional analysis approach, we used this method on a previously published dataset investigating the modulation by sounds of orientation representation (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). We showed that: (1) SNNs with a unique hidden layer of 10 computational neurons can be trained to categorize the 12 orientations of the tuning block. (2) For each output node (corresponding to the different orientations of the tuning block), SNNs assign to each input node a connection weight that corresponds to the tuning of the input cortical neuron for that orientation. (3) The circular mean of the SNN output can be used to estimate the orientation of drifting gratings even when the SNN was not trained to categorize that orientation. (4) Using this approach, we confirmed that orientation representation is improved in na&#x00EF;ve mice when a sound is presented simultaneously with the visual stimulus. (5) Finally, we extended the approach to a new dataset and showed that in mice performing a cross-modal Go detection task the sound-induced modulation of the V1 orientation representation depends on the importance of the visual stimulus for the behavior. Indeed, sounds improved the representation of visual stimuli that acquired a high behavioral importance for the animals (the Go signal) while degrading the representation of the other visual stimuli (the NoGo signal).</p>
<p>This study is, to our knowledge, the first study using SNNs to evaluate the representation by the V1 neuronal population of the orientation of visual stimuli. The use of SNNs to analyze the structure of neuronal activity of the visual cortex brings convolutional networks back to their roots, as its original structure was inspired by the connectivity of the vertebrate visual system (<xref ref-type="bibr" rid="B11">Fukushima, 1980</xref>). Recently, convolutional networks started to be used to model sensory processing in V1 and were found to be more effective than other traditional methods (<xref ref-type="bibr" rid="B43">Zhang et al., 2019</xref>). In our study, the SNNs received the input of a subset of V1 L2/3 neurons and were trained to categorize their activity patterns evoked by the stimuli of the tuning block. Their high accuracy for classifying the trained input patterns quickly led us to abandon the idea of training SNNs to discriminate between the 45&#x00B0; and 135&#x00B0; visual stimuli of the unimodal block. Indeed, this approach leaves very little room for improvement as the classification of the trained stimuli is highly efficient. Moreover, training the SNN to classify the unimodal block stimuli would not be addressing the question of orientation representation, but estimate the capability of the SNN to use the neuronal activity to discriminate between the two stimuli; an approach similar to that of studies using linear discriminant analysis (<xref ref-type="bibr" rid="B38">Stringer et al., 2021</xref>). Thus, we devised the alternate strategy of using the output of SNNs trained to categorize the V1 neuronal activity evoked by the 12 orientations of the tuning block, allowing us to assess how the orientations of the unimodal and audiovisual stimuli were represented in V1. Indeed, as the output layer of the SNNs uses a softmax function, SNN outputs indicate the probability that the presented visual stimulus belongs to the different trained stimulus orientation categories. By linearizing the categorical outputs in the orientation space using the circular mean, we were able not only to estimate the orientation of the visual stimulus, but also the specificity of this neuronal pattern.</p>
<p>For this report, we chose to train hundreds of SNNs with different subsets of the cortical neurons present in our databases. Our goal in using this approach was to assess the statistical variability of the orientation representation across the V1 neuronal population. We found large variations in SNN performance depending on the subset of neurons used. The poor performance of some SNNs is likely due to samples with a small proportion of well-responsive neurons. This would explain why the presence of sounds is particularly efficient at improving the performance of SNNs poorly responding in the unimodal context (as shown in the quiver plot analysis <xref ref-type="fig" rid="F4">Figures 4I,J</xref>). Indeed, we had already shown that the improved representation of oriented stimuli in the audiovisual context is due to an increase in the response of tuned neurons and a decrease in response of neurons not tuned for the stimulus (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>), limiting the risks of misclassification due to outlying neuronal activities. The main advantage of the SNN approach compared to the approach used in our previous study (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>), is that all the neurons are now included in the analysis. During training, every computational input neuron is given a connection weight proportional to the importance of this neuron in the assessment of the classifier output (<xref ref-type="bibr" rid="B13">Garson, 1991</xref>; <xref ref-type="bibr" rid="B15">Goh, 1995</xref>). The possibility of determining the connection weights using the <italic>Connection Weight Approach</italic> (<xref ref-type="bibr" rid="B31">Olden and Jackson, 2002</xref>; <xref ref-type="bibr" rid="B32">Olden et al., 2004</xref>) is a great strength of SNNs. Hence, we show that for each categorical output (i.e., the 12 orientations of the tuning block) the largest weights are attributed to neurons having similar preferred orientations. Moreover, we show that neurons with the largest weight are neurons with the best orientation selectivity and/or best direction selectivity. Thanks to the <italic>Connection Weight Approach</italic>, it will be possible in future studies to determine which neuronal population drives the sound modulation of orientation representation, and whether it happens through mechanisms such as potentiation of tuned neurons, suppression of untuned neurons, or improved trial to trial reliability of the neurons.</p>
<p>We used the SNN approach on two databases that were generated to investigate how sound modulates the visually evoked neuronal activities in V1. Indeed, in the past decade, an increasing number of studies have shown that the presence of sounds modifies the response of neurons to the presentation of visual stimuli in the mouse V1 (<xref ref-type="bibr" rid="B20">Iurilli et al., 2012</xref>; <xref ref-type="bibr" rid="B19">Ibrahim et al., 2016</xref>; <xref ref-type="bibr" rid="B25">Meijer et al., 2017</xref>; <xref ref-type="bibr" rid="B8">Deneux et al., 2019</xref>; <xref ref-type="bibr" rid="B21">Kn&#x00F6;pfel et al., 2019</xref>; <xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>; <xref ref-type="bibr" rid="B12">Garner and Keller, 2021</xref>). Those studies are characterized by a large array of recording techniques (electrophysiology and functional imaging), different sounds and visual stimuli, and a great variety of analysis approaches. The database of na&#x00EF;ve mice used in this study was generated for our previous report in which we showed that the presence of pure tones improves the representation of the orientation and direction of the visual stimulus in V1 L2/3 by favoring the recruitment of a neuronal population better tuned to the visual stimulus orientation and direction than the population responding to the unimodal visual stimulus (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). Here, we confirmed using the SNN approach that the presence of pure tones improves the representation of the orientation of the visual stimuli. This new approach allows us to assess orientation representation by the whole population and not using a subset of selected &#x201C;active&#x201D; or &#x201C;responsive&#x201D; neurons (<xref ref-type="bibr" rid="B19">Ibrahim et al., 2016</xref>; <xref ref-type="bibr" rid="B25">Meijer et al., 2017</xref>; <xref ref-type="bibr" rid="B8">Deneux et al., 2019</xref>; <xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). We also confirmed that sound modulation is stronger in V1 neurons that poorly respond to the visual stimulus in the unimodal context, or that are biased toward the opposite direction (<xref ref-type="fig" rid="F4">Figures 4I,J</xref>; <xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>). Note that we had already shown that arousal and locomotion could not account for those results (<xref ref-type="bibr" rid="B24">McClure and Polack, 2019</xref>), and we therefore did not consider those parameters further in this study. Moreover, several studies have recently demonstrated that the modulation of the V1 neuronal activity by behavioral parameters such as locomotion and arousal are orthogonal to orientation encoding (<xref ref-type="bibr" rid="B17">Hajnal et al., 2021</xref>; <xref ref-type="bibr" rid="B39">Stringer et al., 2019</xref>).</p>
<p>We also present novel findings suggesting that sound modulation itself depends on the audiovisual context. Indeed, we show in mice performing a cross-modal Go detection task that the presence of sounds improves the representation of the Go visual cue orientation while degrading the representation of the NoGo visual cue orientation. This degradation of the NoGo visual cue orientation representation is mostly carried by the degradation of the best performing SNNs. This suggests that this effect is supported by a decrease in the responsiveness of highly tuned neurons. This result extends previous findings showing that incongruent audiovisual stimulation (a looming visual stimulus associated to a frequency-modulated tone) had a suppressive effect on V1 neuronal responses while congruent audiovisual stimuli did not significantly change the neuronal responses in V1 [(<xref ref-type="bibr" rid="B25">Meijer et al., 2017</xref>) but see also (<xref ref-type="bibr" rid="B12">Garner and Keller, 2021</xref>)]. In our behavioral paradigm, we did not find an effect of behavioral congruence as both NoGo and Go sounds similarly suppressed the NoGo visual cue while improving the Go visual cue orientation representation. Future experiments will be necessary to determine the cellular and network mechanisms underpinning the differential modulation of V1 visual processing by sound.</p>
</sec>
<sec id="S5" sec-type="data-availability">
<title>Data Availability Statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec id="S6">
<title>Ethics Statement</title>
<p>The animal study was reviewed and approved by the Institutional Animal Care and Use Committee (IACUC) of Rutgers University&#x2013;Newark.</p>
</sec>
<sec id="S7">
<title>Author Contributions</title>
<p>JM and P-OP designed the project. JM performed the experiments. P-OP performed the analyses with assistance from OBE, JC, and JM. P-OP wrote the manuscript with assistance from JC. All the authors revised the manuscript.</p>
</sec>
<sec id="conf1" sec-type="COI-statement">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="pudiscl1" sec-type="disclaimer">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
</body>
<back>
<sec id="S8" sec-type="funding-information">
<title>Funding</title>
<p>This work was funded by the Whitehall Foundation (Grant No. 2015-08-69), the Charles and Johanna Busch Biomedical Grant Program, and the National Institutes of Health &#x2013; National Eye Institute (Grant No. #R01 EY030860). JC was supported by a Fyssen Foundation postdoctoral fellowship.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Barth</surname> <given-names>A. L.</given-names></name> <name><surname>Poulet</surname> <given-names>J. F.</given-names></name></person-group> (<year>2012</year>). <article-title>Experimental evidence for sparse firing in the neocortex.</article-title> <source><italic>Trends Neurosci.</italic></source> <volume>35</volume> <fpage>345</fpage>&#x2013;<lpage>355</lpage>. <pub-id pub-id-type="doi">10.1016/j.tins.2012.03.008</pub-id> <pub-id pub-id-type="pmid">22579264</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Berens</surname> <given-names>P.</given-names></name></person-group> (<year>2009</year>). <article-title>CircStat: a MATLAB Toolbox for Circular Statistics.</article-title> <source><italic>J. Stat. Softw.</italic></source> <volume>31</volume> <fpage>1</fpage>&#x2013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1016/j.biopsycho.2022.108291</pub-id> <pub-id pub-id-type="pmid">35202742</pub-id></citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cappe</surname> <given-names>C.</given-names></name> <name><surname>Barone</surname> <given-names>P.</given-names></name></person-group> (<year>2005</year>). <article-title>Heteromodal connections supporting multisensory integration at low levels of cortical processing in the monkey.</article-title> <source><italic>Eur. J. Neurosci.</italic></source> <volume>22</volume> <fpage>2886</fpage>&#x2013;<lpage>2902</lpage>. <pub-id pub-id-type="doi">10.1111/j.1460-9568.2005.04462.x</pub-id> <pub-id pub-id-type="pmid">16324124</pub-id></citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Carrillo-Reid</surname> <given-names>L.</given-names></name> <name><surname>Han</surname> <given-names>S.</given-names></name> <name><surname>Yang</surname> <given-names>W.</given-names></name> <name><surname>Akrouh</surname> <given-names>A.</given-names></name> <name><surname>Yuste</surname> <given-names>R.</given-names></name></person-group> (<year>2019</year>). <article-title>Controlling Visually Guided Behavior by Holographic Recalling of Cortical Ensembles.</article-title> <source><italic>Cell</italic></source> <volume>178</volume> <fpage>447</fpage>&#x2013;<lpage>457e445</lpage>. <pub-id pub-id-type="doi">10.1016/j.cell.2019.05.045</pub-id> <pub-id pub-id-type="pmid">31257030</pub-id></citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cronin</surname> <given-names>B.</given-names></name> <name><surname>Stevenson</surname> <given-names>I. H.</given-names></name> <name><surname>Sur</surname> <given-names>M.</given-names></name> <name><surname>Kording</surname> <given-names>K. P.</given-names></name></person-group> (<year>2010</year>). <article-title>Hierarchical Bayesian modeling and Markov chain Monte Carlo sampling for tuning-curve analysis.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>103</volume> <fpage>591</fpage>&#x2013;<lpage>602</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00379.2009</pub-id> <pub-id pub-id-type="pmid">19889855</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cunningham</surname> <given-names>J. P.</given-names></name> <name><surname>Yu</surname> <given-names>B. M.</given-names></name></person-group> (<year>2014</year>). <article-title>Dimensionality reduction for large-scale neural recordings.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>17</volume> <fpage>1500</fpage>&#x2013;<lpage>1509</lpage>. <pub-id pub-id-type="doi">10.1038/nn.3776</pub-id> <pub-id pub-id-type="pmid">25151264</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Decharms</surname> <given-names>R. C.</given-names></name> <name><surname>Zador</surname> <given-names>A.</given-names></name></person-group> (<year>2000</year>). <article-title>Neural Representation and the Cortical Code.</article-title> <source><italic>Annu. Rev. Neurosci.</italic></source> <volume>23</volume> <fpage>613</fpage>&#x2013;<lpage>647</lpage>. <pub-id pub-id-type="doi">10.1146/annurev.neuro.23.1.613</pub-id> <pub-id pub-id-type="pmid">10845077</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Deneux</surname> <given-names>T.</given-names></name> <name><surname>Harrell</surname> <given-names>E. R.</given-names></name> <name><surname>Kempf</surname> <given-names>A.</given-names></name> <name><surname>Ceballo</surname> <given-names>S.</given-names></name> <name><surname>Filipchuk</surname> <given-names>A.</given-names></name> <name><surname>Bathellier</surname> <given-names>B.</given-names></name></person-group> (<year>2019</year>). <article-title>Context-dependent signaling of coincident auditory and visual events in primary visual cortex.</article-title> <source><italic>Elife</italic></source> <volume>2019</volume>:<issue>8</issue>. <pub-id pub-id-type="doi">10.7554/eLife.44006</pub-id> <pub-id pub-id-type="pmid">31115334</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Einstein</surname> <given-names>M. C.</given-names></name> <name><surname>Polack</surname> <given-names>P. O.</given-names></name> <name><surname>Tran</surname> <given-names>D. T.</given-names></name> <name><surname>Golshani</surname> <given-names>P.</given-names></name></person-group> (<year>2017</year>). <article-title>Visually Evoked 3-5 Hz Membrane Potential Oscillations Reduce the Responsiveness of Visual Cortex Neurons in Awake Behaving Mice.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>37</volume> <fpage>5084</fpage>&#x2013;<lpage>5098</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.3868-16.2017</pub-id> <pub-id pub-id-type="pmid">28432140</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Falchier</surname> <given-names>A.</given-names></name> <name><surname>Clavagnier</surname> <given-names>S.</given-names></name> <name><surname>Barone</surname> <given-names>P.</given-names></name> <name><surname>Kennedy</surname> <given-names>H.</given-names></name></person-group> (<year>2002</year>). <article-title>Anatomical evidence of multimodal integration in primate striate cortex.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>22</volume> <fpage>5749</fpage>&#x2013;<lpage>5759.s</lpage> <pub-id pub-id-type="doi">10.1523/JNEUROSCI.22-13-05749.2002</pub-id> <pub-id pub-id-type="pmid">12097528</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fukushima</surname> <given-names>K.</given-names></name></person-group> (<year>1980</year>). <article-title>Neocognitron: a self organizing neural network model for a mechanism of pattern recognition unaffected by shift in position.</article-title> <source><italic>Biol. Cybern.</italic></source> <volume>36</volume> <fpage>193</fpage>&#x2013;<lpage>202</lpage>. <pub-id pub-id-type="doi">10.1007/bf00344251</pub-id> <pub-id pub-id-type="pmid">7370364</pub-id></citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Garner</surname> <given-names>A. R.</given-names></name> <name><surname>Keller</surname> <given-names>G. B.</given-names></name></person-group> (<year>2021</year>). <article-title>A cortical circuit for audio-visual predictions.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>25</volume> <fpage>98</fpage>&#x2013;<lpage>105</lpage>. <pub-id pub-id-type="doi">10.1038/s41593-021-00974-7</pub-id> <pub-id pub-id-type="pmid">34857950</pub-id></citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Garson</surname> <given-names>G. D.</given-names></name></person-group> (<year>1991</year>). <article-title>Interpreting neural network connection weights.</article-title> <source><italic>Artif. Intell. Exp.</italic></source> <volume>6</volume> <fpage>46</fpage>&#x2013;<lpage>51</lpage>.</citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gleiss</surname> <given-names>S.</given-names></name> <name><surname>Kayser</surname> <given-names>C.</given-names></name></person-group> (<year>2014</year>). <article-title>Acoustic noise improves visual perception and modulates occipital oscillatory states.</article-title> <source><italic>J. Cogn. Neurosci.</italic></source> <volume>26</volume> <fpage>699</fpage>&#x2013;<lpage>711</lpage>. <pub-id pub-id-type="doi">10.1162/jocn_a_00524</pub-id> <pub-id pub-id-type="pmid">24236698</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Goh</surname> <given-names>A. T. C.</given-names></name></person-group> (<year>1995</year>). <article-title>Back-propagation neural networks for modeling complex systems.</article-title> <source><italic>Artif. Intell. Eng.</italic></source> <volume>9</volume> <fpage>143</fpage>&#x2013;<lpage>151</lpage>. <pub-id pub-id-type="doi">10.1016/0954-1810(94)00011-s</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Golledge</surname> <given-names>H. D.</given-names></name> <name><surname>Panzeri</surname> <given-names>S.</given-names></name> <name><surname>Zheng</surname> <given-names>F.</given-names></name> <name><surname>Pola</surname> <given-names>G.</given-names></name> <name><surname>Scannell</surname> <given-names>J. W.</given-names></name> <name><surname>Giannikopoulos</surname> <given-names>D. V.</given-names></name><etal/></person-group> (<year>2003</year>). <article-title>Correlations, feature-binding and population coding in primary visual cortex.</article-title> <source><italic>Neuroreport</italic></source> <volume>14</volume> <fpage>1045</fpage>&#x2013;<lpage>1050</lpage>. <pub-id pub-id-type="doi">10.1097/01.wnr.0000073681.00308.9c</pub-id> <pub-id pub-id-type="pmid">12802200</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hajnal</surname> <given-names>M. A.</given-names></name> <name><surname>Tran</surname> <given-names>D.</given-names></name> <name><surname>Einstein</surname> <given-names>M.</given-names></name> <name><surname>Martelo</surname> <given-names>M. V.</given-names></name> <name><surname>Safaryan</surname> <given-names>K.</given-names></name> <name><surname>Polack</surname> <given-names>P.-O.</given-names></name><etal/></person-group> (<year>2021</year>). <article-title>Continuous multiplexed population representations of task context in the mouse primary visual cortex.</article-title> <source><italic>bioRxiv</italic></source> <volume>2021</volume>:<issue>440666</issue>.</citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hidaka</surname> <given-names>S.</given-names></name> <name><surname>Ide</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Sound can suppress visual perception.</article-title> <source><italic>Sci. Rep.</italic></source> <volume>5</volume>:<issue>10483</issue>. <pub-id pub-id-type="doi">10.1038/srep10483</pub-id> <pub-id pub-id-type="pmid">26023877</pub-id></citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ibrahim</surname> <given-names>L. A.</given-names></name> <name><surname>Mesik</surname> <given-names>L.</given-names></name> <name><surname>Ji</surname> <given-names>X. Y.</given-names></name> <name><surname>Fang</surname> <given-names>Q.</given-names></name> <name><surname>Li</surname> <given-names>H. F.</given-names></name> <name><surname>Li</surname> <given-names>Y. T.</given-names></name><etal/></person-group> (<year>2016</year>). <article-title>Cross-Modality Sharpening of Visual Cortical Processing through Layer-1-Mediated Inhibition and Disinhibition.</article-title> <source><italic>Neuron</italic></source> <volume>89</volume> <fpage>1031</fpage>&#x2013;<lpage>1045</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2016.01.027</pub-id> <pub-id pub-id-type="pmid">26898778</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Iurilli</surname> <given-names>G.</given-names></name> <name><surname>Ghezzi</surname> <given-names>D.</given-names></name> <name><surname>Olcese</surname> <given-names>U.</given-names></name> <name><surname>Lassi</surname> <given-names>G.</given-names></name> <name><surname>Nazzaro</surname> <given-names>C.</given-names></name> <name><surname>Tonini</surname> <given-names>R.</given-names></name><etal/></person-group> (<year>2012</year>). <article-title>Sound-driven synaptic inhibition in primary visual cortex.</article-title> <source><italic>Neuron</italic></source> <volume>73</volume> <fpage>814</fpage>&#x2013;<lpage>828</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2011.12.026</pub-id> <pub-id pub-id-type="pmid">22365553</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kn&#x00F6;pfel</surname> <given-names>T.</given-names></name> <name><surname>Sweeney</surname> <given-names>Y.</given-names></name> <name><surname>Radulescu</surname> <given-names>C. I.</given-names></name> <name><surname>Zabouri</surname> <given-names>N.</given-names></name> <name><surname>Doostdar</surname> <given-names>N.</given-names></name> <name><surname>Clopath</surname> <given-names>C.</given-names></name><etal/></person-group> (<year>2019</year>). <article-title>Audio-visual experience strengthens multisensory assemblies in adult mouse visual cortex.</article-title> <source><italic>Nat. Comm.</italic></source> <volume>10</volume>:<issue>5684</issue>. <pub-id pub-id-type="doi">10.1038/s41467-019-13607-2</pub-id> <pub-id pub-id-type="pmid">31831751</pub-id></citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lindsay</surname> <given-names>G. W.</given-names></name></person-group> (<year>2021</year>). <article-title>Convolutional Neural Networks as a Model of the Visual System: Past. Present, and Future.</article-title> <source><italic>J. Cogn. Neurosci.</italic></source> <volume>33</volume> <fpage>2017</fpage>&#x2013;<lpage>2031</lpage>. <pub-id pub-id-type="doi">10.1162/jocn_a_01544</pub-id> <pub-id pub-id-type="pmid">32027584</pub-id></citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lippert</surname> <given-names>M.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name> <name><surname>Kayser</surname> <given-names>C.</given-names></name></person-group> (<year>2007</year>). <article-title>Improvement of visual contrast detection by a simultaneous sound.</article-title> <source><italic>Brain Res.</italic></source> <volume>1173</volume> <fpage>102</fpage>&#x2013;<lpage>109</lpage>. <pub-id pub-id-type="doi">10.1016/j.brainres.2007.07.050</pub-id> <pub-id pub-id-type="pmid">17765208</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mc&#x003C;/snm&#x003E;Clure</surname> <given-names>J. P.</given-names><suffix>Jr.</suffix></name> <name><surname>Polack</surname> <given-names>P. O.</given-names></name></person-group> (<year>2019</year>). <article-title>Pure tones modulate the representation of orientation and direction in the primary visual cortex.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>121</volume> <fpage>2202</fpage>&#x2013;<lpage>2214</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00069.2019</pub-id> <pub-id pub-id-type="pmid">30969800</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Meijer</surname> <given-names>G. T.</given-names></name> <name><surname>Montijn</surname> <given-names>J. S.</given-names></name> <name><surname>Pennartz</surname> <given-names>C. M. A.</given-names></name> <name><surname>Lansink</surname> <given-names>C. S.</given-names></name></person-group> (<year>2017</year>). <article-title>Audiovisual Modulation in Mouse Primary Visual Cortex Depends on Cross-Modal Stimulus Configuration and Congruency.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>37</volume> <fpage>8783</fpage>&#x2013;<lpage>8796</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0468-17.2017</pub-id> <pub-id pub-id-type="pmid">28821672</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Molholm</surname> <given-names>S.</given-names></name> <name><surname>Sehatpour</surname> <given-names>P.</given-names></name> <name><surname>Mehta</surname> <given-names>A. D.</given-names></name> <name><surname>Shpaner</surname> <given-names>M.</given-names></name> <name><surname>Gomez-Ramirez</surname> <given-names>M.</given-names></name> <name><surname>Ortigue</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2006</year>). <article-title>Audio-visual multisensory integration in superior parietal lobule revealed by human intracranial recordings.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>96</volume> <fpage>721</fpage>&#x2013;<lpage>729</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00285.2006</pub-id> <pub-id pub-id-type="pmid">16687619</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Montani</surname> <given-names>F.</given-names></name> <name><surname>Kohn</surname> <given-names>A.</given-names></name> <name><surname>Smith</surname> <given-names>M. A.</given-names></name> <name><surname>Schultz</surname> <given-names>S. R.</given-names></name></person-group> (<year>2007</year>). <article-title>The role of correlations in direction and contrast coding in the primary visual cortex.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>27</volume> <fpage>2338</fpage>&#x2013;<lpage>2348</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.3417-06.2007</pub-id> <pub-id pub-id-type="pmid">17329431</pub-id></citation></ref>
<ref id="B28"><citation citation-type="journal"><collab>National Research Council of the National Academies.</collab> (<year>2011</year>). <source><italic>Guide for the Care and Use of Laboratory Animals</italic></source>, 8th Edn. <publisher-loc>Washington, DC</publisher-loc>: <publisher-name>National Academies Press</publisher-name>.</citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Niell</surname> <given-names>C. M.</given-names></name> <name><surname>Stryker</surname> <given-names>M. P.</given-names></name></person-group> (<year>2010</year>). <article-title>Modulation of visual responses by behavioral state in mouse visual cortex.</article-title> <source><italic>Neuron</italic></source> <volume>65</volume> <fpage>472</fpage>&#x2013;<lpage>479</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2010.01.033</pub-id> <pub-id pub-id-type="pmid">20188652</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Odgaard</surname> <given-names>E. C.</given-names></name> <name><surname>Arieh</surname> <given-names>Y.</given-names></name> <name><surname>Marks</surname> <given-names>L. E.</given-names></name></person-group> (<year>2004</year>). <article-title>Brighter noise: sensory enhancement of perceived loudness by concurrent visual stimulation.</article-title> <source><italic>Cogn. Affect Behav. Neurosci.</italic></source> <volume>4</volume> <fpage>127</fpage>&#x2013;<lpage>132</lpage>. <pub-id pub-id-type="doi">10.3758/cabn.4.2.127</pub-id> <pub-id pub-id-type="pmid">15460919</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Olden</surname> <given-names>J. D.</given-names></name> <name><surname>Jackson</surname> <given-names>D. A.</given-names></name></person-group> (<year>2002</year>). <article-title>Illuminating the &#x201C;black box&#x201D;: a randomization approach for understanding variable contributions in artificial neural networks.</article-title> <source><italic>Ecolog. Model.</italic></source> <volume>154</volume> <fpage>135</fpage>&#x2013;<lpage>150</lpage>. <pub-id pub-id-type="doi">10.1016/s0304-3800(02)00064-9</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Olden</surname> <given-names>J. D.</given-names></name> <name><surname>Joy</surname> <given-names>M. K.</given-names></name> <name><surname>Death</surname> <given-names>R. G.</given-names></name></person-group> (<year>2004</year>). <article-title>An accurate comparison of methods for quantifying variable importance in artificial neural networks using simulated data.</article-title> <source><italic>Ecolog. Model.</italic></source> <volume>178</volume> <fpage>389</fpage>&#x2013;<lpage>397</lpage>. <pub-id pub-id-type="doi">10.1016/j.ecolmodel.2004.03.013</pub-id></citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Polack</surname> <given-names>P. O.</given-names></name> <name><surname>Friedman</surname> <given-names>J.</given-names></name> <name><surname>Golshani</surname> <given-names>P.</given-names></name></person-group> (<year>2013</year>). <article-title>Cellular mechanisms of brain state-dependent gain modulation in visual cortex.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>16</volume> <fpage>1331</fpage>&#x2013;<lpage>1339</lpage>. <pub-id pub-id-type="doi">10.1038/nn.3464</pub-id> <pub-id pub-id-type="pmid">23872595</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Quian Quiroga</surname> <given-names>R.</given-names></name> <name><surname>Panzeri</surname> <given-names>S.</given-names></name></person-group> (<year>2009</year>). <article-title>Extracting information from neuronal populations: information theory and decoding approaches.</article-title> <source><italic>Nat. Rev. Neurosci.</italic></source> <volume>10</volume> <fpage>173</fpage>&#x2013;<lpage>185</lpage>. <pub-id pub-id-type="doi">10.1038/nrn2578</pub-id> <pub-id pub-id-type="pmid">19229240</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rockland</surname> <given-names>K. S.</given-names></name> <name><surname>Ojima</surname> <given-names>H.</given-names></name></person-group> (<year>2003</year>). <article-title>Multisensory convergence in calcarine visual areas in macaque monkey.</article-title> <source><italic>Int. J. Psychophysiol.</italic></source> <volume>50</volume> <fpage>19</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1016/s0167-8760(03)00121-1</pub-id> <pub-id pub-id-type="pmid">14511833</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rumyantsev</surname> <given-names>O. I.</given-names></name> <name><surname>Lecoq</surname> <given-names>J. A.</given-names></name> <name><surname>Hernandez</surname> <given-names>O.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Savall</surname> <given-names>J.</given-names></name> <name><surname>Chrapkiewicz</surname> <given-names>R.</given-names></name><etal/></person-group> (<year>2020</year>). <article-title>Fundamental bounds on the fidelity of sensory cortical coding.</article-title> <source><italic>Nature</italic></source> <volume>580</volume> <fpage>100</fpage>&#x2013;<lpage>105</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-020-2130-2</pub-id> <pub-id pub-id-type="pmid">32238928</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>C.</given-names></name> <name><surname>Sandberg</surname> <given-names>K.</given-names></name> <name><surname>Andersen</surname> <given-names>L. M.</given-names></name> <name><surname>Blicher</surname> <given-names>J. U.</given-names></name> <name><surname>Rees</surname> <given-names>G.</given-names></name></person-group> (<year>2017</year>). <article-title>Human Occipital and Parietal GABA Selectively Influence Visual Perception of Orientation and Size.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>37</volume> <fpage>8929</fpage>&#x2013;<lpage>8937</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.3945-16.2017</pub-id> <pub-id pub-id-type="pmid">28821653</pub-id></citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stringer</surname> <given-names>C.</given-names></name> <name><surname>Michaelos</surname> <given-names>M.</given-names></name> <name><surname>Tsyboulski</surname> <given-names>D.</given-names></name> <name><surname>Lindo</surname> <given-names>S. E.</given-names></name> <name><surname>Pachitariu</surname> <given-names>M.</given-names></name></person-group> (<year>2021</year>). <article-title>High-precision coding in visual cortex.</article-title> <source><italic>Cell</italic></source> <volume>184</volume> <fpage>2767.e</fpage>&#x2013;<lpage>2778.e</lpage>. <pub-id pub-id-type="doi">10.1016/j.cell.2021.03.042</pub-id> <pub-id pub-id-type="pmid">33857423</pub-id></citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stringer</surname> <given-names>C.</given-names></name> <name><surname>Pachitariu</surname> <given-names>M.</given-names></name> <name><surname>Steinmetz</surname> <given-names>N.</given-names></name> <name><surname>Reddy</surname> <given-names>C. B.</given-names></name> <name><surname>Carandini</surname> <given-names>M.</given-names></name> <name><surname>Harris</surname> <given-names>K. D.</given-names></name></person-group> (<year>2019</year>). <article-title>Spontaneous behaviors drive multidimensional, brainwide activity.</article-title> <source><italic>Science</italic></source> <volume>364</volume>:<issue>255</issue>. <pub-id pub-id-type="doi">10.1126/science.aav7893</pub-id> <pub-id pub-id-type="pmid">31000656</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vinck</surname> <given-names>M.</given-names></name> <name><surname>Batista-Brito</surname> <given-names>R.</given-names></name> <name><surname>Knoblich</surname> <given-names>U.</given-names></name> <name><surname>Cardin</surname> <given-names>J. A.</given-names></name></person-group> (<year>2015</year>). <article-title>Arousal and locomotion make distinct contributions to cortical activity patterns and visual encoding.</article-title> <source><italic>Neuron</italic></source> <volume>86</volume> <fpage>740</fpage>&#x2013;<lpage>754</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2015.03.028</pub-id> <pub-id pub-id-type="pmid">25892300</pub-id></citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vroomen</surname> <given-names>J.</given-names></name> <name><surname>De Gelder</surname> <given-names>B.</given-names></name></person-group> (<year>2000</year>). <article-title>Sound enhances visual perception: cross-modal effects of auditory organization on vision.</article-title> <source><italic>J. Exp. Psychol. Hum. Percept. Perform.</italic></source> <volume>26</volume> <fpage>1583</fpage>&#x2013;<lpage>1590</lpage>. <pub-id pub-id-type="doi">10.1037//0096-1523.26.5.1583</pub-id> <pub-id pub-id-type="pmid">11039486</pub-id></citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wohrer</surname> <given-names>A.</given-names></name> <name><surname>Humphries</surname> <given-names>M. D.</given-names></name> <name><surname>Machens</surname> <given-names>C. K.</given-names></name></person-group> (<year>2013</year>). <article-title>Population-wide distributions of neural activity during perceptual decision-making.</article-title> <source><italic>Prog. Neurobiol.</italic></source> <volume>103</volume> <fpage>156</fpage>&#x2013;<lpage>193</lpage>. <pub-id pub-id-type="doi">10.1016/j.pneurobio.2012.09.004</pub-id> <pub-id pub-id-type="pmid">23123501</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Lee</surname> <given-names>T. S.</given-names></name> <name><surname>Li</surname> <given-names>M.</given-names></name> <name><surname>Liu</surname> <given-names>F.</given-names></name> <name><surname>Tang</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Convolutional neural network models of V1 responses to complex patterns.</article-title> <source><italic>J. Comput. Neurosci.</italic></source> <volume>46</volume> <fpage>33</fpage>&#x2013;<lpage>54</lpage>. <pub-id pub-id-type="doi">10.1007/s10827-018-0687-7</pub-id> <pub-id pub-id-type="pmid">29869761</pub-id></citation></ref>
</ref-list>
</back>
</article>
