3,850 Matching Annotations
  1. Oct 2024
    1. Out on the street, the largest riot since Conscription was passed in 1944 (bringing in the draft for the final year of the Second World War) broke out along a seven-block length of Rue Ste. Catherine, featuring overturned cars, smashed windows, a shot fired from somewhere and 137 arrests.

      I am surprised I have never heard of this riot or this piece of history.

    1. He found the house gone to decay—the roof fallen in, the windows shattered, and the doors off the hinges. A half-starved dog that looked like Wolf was skulking about it. Rip called him by name, but the cur snarled, showed his teeth, and passed on. This was an unkind cut indeed. “My very dog,” sighed Rip, “has forgotten me!”

      Oh damn this is kinda sad.

    1. Just as the environments in first-person shooters exist to support action-packed combat, the environments in most walking sims are designed to be platforms for understanding and empathizing with characters. In games like Dear Esther, Virginia (2016), What Remains of Edith Finch (2017), and many others, 3D game worlds come to be understood as metaphorical spaces offering windows into the minds and stories of the people within them.

      This very feature of these digital literature games that not only tells stories but also allows players to actively engage in them is what builds our understanding and empathy for the video game characters. As players, to fully understand the extent of someone else’s struggles, we must put ourselves into their shoes and this form of game literally allows this metaphor to come to life.

    1. “Because I always try so hard to win and had my troubles in Boston, I was suspended. At playoff time, it hurts not to be in the game with the boys. However, I want to do what is good for the people of Montreal and the team. So that no further harm will be done, I would like to ask everyone to get behind the team and to help the boys win from the Rangers and Detroit. I will take my punishment and come back next year to help the club and younger players to win the Cup.” His words had a palliative effect. The next night nobody threw galoshes, nobody broke any more windows, nobody stopped streetcars.

      I'm glad the people listened to his words and that he was so calm about it even after probably being upset that he was suspended.

  2. Sep 2024
    1. At such times I could not help remarking and admiring (although from his rich ideality I had been prepared to expect it{w}) a peculiar analytic ability in Dupin. He seemed, too, to take an eager delight in its exercise — if not exactly in its display — and did not hesitate{x} to confess the pleasure thus derived. He boasted to me, with a low chuckling laugh, that most men, in respect to himself, wore windows in their bosoms,(8) and was wont to follow up such assertions by direct and very startling proofs of his intimate knowledge of my own. His manner at these moments was frigid and abstract; his eyes were vacant in expression; while his voice, usually a rich tenor, rose into a treble which would have sounded petulantly but for the deliberateness and entire distinctness of the enunciation. Observing him in these moods, I often dwelt meditatively upon the old philosophy of the Bi-Part Soul,(9) and amused myself with the fancy of a double Dupin — the creative and the resolvent.

      I think the author is building up the charming characteristic for Dupin. To make readers like him more.<br /> Dupin was described as a well mannered man with money.

    1. tog vi dessutom fram en mall för att kunna hämta data från dessa värmepumpar på ett skalbart sätt via den infrastruktur för datainsamling som redan finns på plats inom Fabege.

      hänvisa till Figur 6 och bilagan.

      Tabellen i bilagan behöver förklaras. Annars tror jag att det blir lite kinesiska.

    2. Inom projektet genomfördes även försök till modellering av kretsar med de föreslagna datapunkterna i enlighet med Brick Schema. Vi anser att det går att modellera en krets i enlighet med den föreslagna omfattningen, men det kommer att kräva en utökning av de nuvarande standardiserade klasserna inom Brick Schema, i enlighet med de befintliga rekommendationerna för hur systemet vid behov ska kunna byggas ut.

      Varför redovisar vi inte denna del av arbetet som är så relevant för denna rapport?

    3. den låga dataupplösningen (1 timme) anses dessa resultat inte vara tillförlitliga

      här emotsäger vi oss själva. Ovan i sektion 1 skrev vi att 1 timme är OK. Så vi plockar bort den texten och håller oss fast till 5 minuter som rekommenderad upplösning.

    4. Bilderna visar att även om värmepumpen inte använder uteluft som energikälla1, finns det trender mellan prestanda och uteluftstemperatur. Man ser dessutom att dessa trender verkar skilja sig åt lite mellan olika perioder. COP verkar i synnerhet vara lägre under 2023-2024 jämfört med tidigare period. Samma skillnad ser man dock i Carnot-effektivitetens kurva vilket tyder på att denna inte beror på värmepumparna själva utan ändrade temperatur- och/eller flödesförutsättningar vid förångaren och kondensorn. Orsaken bör undersökas vidare. 1 Att använda uteluften är ett sätt att normalisera för energibehov.Att det finns trender möjliggör användningen av maskininlärningsmodeller inte bara för felsökning, utan även för att förutsäga prestanda, exempelvis för driftoptimering. En sådan modell hade kunnat flagga för att COP under 2023 var lägre än referensen under 2022. Maskininlärningsmodeller kan också användas för att optimera driften för att maximera effektivitet och lönsamhet.

      Eftersom uteluften blir kallare vintertid är det rimligt att maskinerna måste höja framledningstemperaturen, vilket gör att COPvärme minskar. Motsatsen gäller sommartid, att maskinerna jobba med sämre COPkyla när framledningstemperaturen på Kylkretsen behöver vara lägre pga hög utelufttemperatur. Jag tror inte vattenflödet genom förångaren och kondensor ändras runt baslastmaskiner men det är klart något att dubbelkolla.

      Här har vi kanske en intressant potentiellt mervärde. Beroende på var i systemet maskinerna kopplas kan man ev lyckas få de att jobba med mer jämnt COP.

    5. Eftersom det saknas energimätare för enskilda värmepumpar är det inte möjligt att beräkna prestandan för varje enskild värmepump.

      menar vi att energimätarna ovan mäter total för alla tre maskiner? bör förtydligas

    6. Detta kan indikera att börvärdena för systemet inte är optimalt inställda för att hantera dellastförhållanden.

      Denna slutsats måste också fram i början av rapporten

    7. Frekvenssignal

      Denna variabel/mätning är lite otydligt. Vad är det som mäts via denna signal? Kompressorfrekvens? hur relaterar vi det till Hertz till exempel? Värt att förklara för läsaren.

    8. Detta antyder att driften kan optimeras genom att säkerställa att värmepumparna opererar inom detta område så ofta som möjligt, t.ex. genom att anpassa vid vilken frekvens uppväxlingen mellan KVP01 och KVP02 sker

      Detta är en bra slutsats. Har den tagits upp i första kapitlet? Det borde vi göra!

    9. För att uppnå dessa mål föreslår vi att följande åtgärder vidtas i kommande sprintar:

      Ändra till "Vi föreslår följande plan i eventuella kommande sprintar:" Ta bort "för att uppnå dessa mål".

    10. bilagan

      BILAGA 1.

      Eftersom Bilaga1 är så viktigt så tänker jag att vi tar innehållet där direkt i denna sektion. En förklaring av Bilaga 1ger mer konkret idé på det ev fortsatta arbete.

    11. Implementera enkla maskininlärningsmodeller för att jämföra värmepumparnas faktiska prestanda med förväntad prestanda. Detta ger oss möjlighet att snabbt upptäcka avvikelser och möjliggör prediktivt underhåll, vilket kan minska driftstopp samt optimera energianvändningen och minska inköp av energi. Automatisera en del av energianalysprocessen genom att hitta gemensamma mönster i sensorer, tillgänglig data och beräkningsalgoritmer för olika anläggningar. Detta kan göras med hjälp av en generisk data modell och kan leda till en kostnadseffektiv energianalys på fastighetsbeståndsnivå. Jämföra kompressorernas verkningsgrad med tillverkarens för att upptäcka en möjlig prestandanedgradering över tid och planera underhåll. Automatisera data kvalité kontrol och filtering för bättre underlag till energianalys

      Även om denna text är bra så säger det ingenting till Fabege. Kanske kan skrivas i slutet av dokumentet

    12. framtida potential för ytterligare effektiviseringar

      En fråga: går det inte att komma med ett någorlunda kvalificerat påstående i stil med: "vi bedömer att en aktiv och datadriven övervakning av energicentraler som motsvarar Apotekaren och Bilia kan resultera i en förbättring av COP med X procentenheter, samt möjligheten att flytta en effektmängd om Y kW inköpt el och Z kW inköpt fjärrvärme på en axel om Å timmar, för att åstadkomma en plattare effektsignatur för den inköpta elen, såväl som för den inköpta fjärrvärmen."

      Jag vet att det är en massa antaganden som behöver göra och att det behövs mycket data och analyser, men nu har jag lyxen att låtsas vara en IT-kille utan koll på byggnadsfysik... om det skulle gå att komma med en sådan hypotes som inte är tagen ur luften, utan tillräckligt väl underbyggd för att hålla för Thomas och Carolines granskning, tror jag att vi har goda förutsättningar för att fortsätta, eftersom de prioriterar effektfrågan inom Fabege.

    13. Bilaga 1: Felsökning

      Denna bilaga blir ett bra säljmaterial om ni förklarar tabellen. Ni kan även koppla det till era observationer i BILIA. Här har vi bra potentiella mervärden som borde lyftas upp i rapporten.

    14. llustrerar till exempel hur smutsiga värmeväxlare eller felaktig mängd köldmedium kan identifieras genom dataanalys

      toppen om ni åtminstone kan förklara några av punkterna. T.ex. ett köldmedieläckage skulle visa sig genom en minskad underkylning? ...

    15. Mindre slutsatser kunde dras från analysen av Apotekaren, där mindre data fanns tillgänglig. Vi identifierade dock tendenser mellan prestanda och utomhustemperatur, vilket indikerar att det är möjligt att skapa modeller för driftoptimering tack vare prediktiv styrning

      denna punkt är otydlig. Formulera om om det verkligen är en relevant slutsats så som rubriken indikerar. Det vore toppen om vi kan presentera något relevant i Apotekaren, lik det vi poängterar i Bilia.

    16. på minst en gång per timme, och helst

      en gång per timme går vi miste om många värmepumpscykler. Jag föreslår att vi endast rekommenderar var 5:e minut.

    17. sprintarna

      kanske bra en en parentes eller pop up som förklarar att sprint är "en bestämd tidsperiod under vilken ett team arbetar för att uppnå specifika leveranser." (googlade)

    18. Detta tyder på att maskininlärningsmodeller kan användas inte bara för felsökning i systemet, utan även för att förutsäga prestanda, exempelvis för driftoptimering.

      Bra, beskriv kort hur

    19. Denna rapport sammanfattar resultaten från ett utvecklingsprojekt om digitalisering av energiprestanda som Bengt Dahlgren GEO genomför för Fabege i samarbete med iQuest. Sammanfattningen gäller de första tre sprintarna, vilka fokuserade på energiprestandan hos värmepumpar.

      Tänk så här: Thomas vill fortsätta arbete med oss men han måste kunna visa till sina chefer att vårt arbete get ett mervärde. Tänk att så om han ska använda denna rapport för att presentera arbetet till ledningen och försöka få dem att godkänna ett fortsättningsuppdrag. Ledningen har inte superbra koll på tekniska detaljer men förstår begrepp och övergripande koncept.

    20. Detta kan indikera att börvärdena för systemet inte är optimalt inställda.

      Var det slutsatsen? Inget om att värmepumparna var överdimensionerade t.ex.? Minns inte exakt vad vi sa där

    21. Fig.1 och Fig.2 visar respektivt de beräknade COP och Carnot-effektivitet som funktion av kompressorns frekvens för båda värmepumparna. Bilder visar att KVP01 presterar bättre än KVP02 vid samma frekvens, vilket indikerar att det kan finnas ett problem med KVP02.

      Vi måste kommentera dessa figurer mer. Hur ska Fabege tolka dem? Vilken information visar dem? jag saknar delen men max. prestanda vid 80% max.signal (runt 9-10 VDC).

    1. The need to protect working-class women was starkly illustrated in 1911 when the Triangle Shirtwaist Factory in Manhattan caught fire. The doors of the factory had been chained shut to prevent women employees from taking unauthorized breaks. The managers who held the keys had saved themselves when the fire broke out, but they left over two hundred women locked in the factory. A rickety fire escape ladder on the side of the building collapsed immediately. Women lined the rooftop and crowded the windows of the ten-story building to avoid the flames and smoke. Many jumped, landing in what newspaper reports described as a “mangled, bloody pulp”. Life nets held by firemen tore at the impact of the falling bodies. Among the onlookers, “women were hysterical, scores fainted; men wept [and] hurled themselves against the police lines.” By the time the fire burned itself out, 71 workers were injured and 146 had died.

      This is my first time learning about this I never knew people back in the day could be this cruel.

    2. The need to protect working-class women was starkly illustrated in 1911 when the Triangle Shirtwaist Factory in Manhattan caught fire. The doors of the factory had been chained shut to prevent women employees from taking unauthorized breaks. The managers who held the keys had saved themselves when the fire broke out, but they left over two hundred women locked in the factory. A rickety fire escape ladder on the side of the building collapsed immediately. Women lined the rooftop and crowded the windows of the ten-story building to avoid the flames and smoke. Many jumped, landing in what newspaper reports described as a “mangled, bloody pulp”. Life nets held by firemen tore at the impact of the falling bodies. Among the onlookers, “women were hysterical, scores fainted; men wept [and] hurled themselves against the police lines.” By the time the fire burned itself out, 71 workers were injured and 146 had died.

      I learned about this in 7th grade and it was so tragic that so many died because the owners swapped the door. The managers saved themselves but left so many trapped in the building.

    3. The need to protect working-class women was starkly illustrated in 1911 when the Triangle Shirtwaist Factory in Manhattan caught fire. The doors of the factory had been chained shut to prevent women employees from taking unauthorized breaks. The managers who held the keys had saved themselves when the fire broke out, but they left over two hundred women locked in the factory. A rickety fire escape ladder on the side of the building collapsed immediately. Women lined the rooftop and crowded the windows of the ten-story building to avoid the flames and smoke. Many jumped, landing in what newspaper reports described as a “mangled, bloody pulp”. Life nets held by firemen tore at the impact of the falling bodies. Among the onlookers, “women were hysterical, scores fainted; men wept [and] hurled themselves against the police lines.” By the time the fire burned itself out, 71 workers were injured and 146 had died.

      Sick.

    1. To fabricate SiCellA materials

      for - scalable fabrication technique - sustainable building - insulation - windows - beer hops aerogel

      scalable fabrication technique - sustainable building - insulation - windows - beer hops aerogel - This fabrication process is highly scalable (Fig. 2a–c) and compatible with roll-to-roll processing, - It combines simple steps: - Moulding to define the volume of the desired hydrogel, - Solvent exchanges at modestly elevated temperatures<br /> - Rolling and drying the gels atop plastic support in rolls - step 1 - Process wood-pulp-derived cellulose nanofibres - via 2,2,6,6-tetramethylpiperidine-1-oxyl radical (TEMPO)-mediated oxidation of native cellulose25,35,36,37 - (Methods and Supplementary Fig. 3). - This process affects the surface charges associated with the carboxylate anion preclude aggregation25,30 of the nanofibres, - which form stable aqueous colloidal dispersions at varying concentrations - step 2 - Pour solution into moulds of desired shapes and sizes - (Methods and Supplementary Figs. 3 and 4) - step 3 - Adding acid interlinks these nanofibres by hydrogen bonds between the carboxyl groups, - transforming the colloidal dispersion into a hydrogel - (Supplementary Fig. 4b) - with a network of sparse nanofibres - step 4 - Exchange the fluid medium within the gel by - replacing water with isopropanol or ethanol - (Fig. 2a,b and Supplementary Fig. 4c,d), - Step 5 - Super critically dry it to form an aerogel - (Fig. 2c and Supplementary Fig. 5a–d)

    1. Augmented e-books make use of hyperlinks (both within and outside the browsed documents), create popup windows, play videos and sounds, might include interactive or live graphs, and even change according to user interaction (either through individual choices or according to a community’s practice).

      Pourrait-on les qualifier de littérature ergodique d'après la définition qu'en fait Aspen J. Aarseth (soit une littérature dans laquelle « un effort non-trivial est requis pour permettre au lecteur de traverser le texte » (Cybertext. Perspectives on Ergodic Literature, 1997, p. 2)) ?

    1. - Joseph Lehman

      I've never heard the phrase "overton window" before, despite this, the effects of overton windows greatly impacts every American. Moving in accordance with the window is exactly expected of lawmakers trying to win the vote of as many Americans as they can.

    1. Imagine that you and I are working in the same company. I tell you there’s a new project for us two to work on. I explain it to you and you get reasonably excited. And then I tell you that I’ve started a new “bloorp” in BloorpyBase, a piece of software from 2012 that almost nobody uses. You grudgingly install BloorpyBase. The app doesn’t use the same keyboard shortcuts you’re used to. The shortcut normally assigned to adding a comment instead minimizes all windows. Sigh. You try to link some exploratory source code to it, but BloorpyBase only works with Mercurial. Sigh. You read some of my initial thoughts and try to respond but you don’t know what’s the best way to do it. Should you create another bloorp? Should you make a suggestion, or an edit? You spend half an hour reading a “How To Bloorp” guide on the internet but come back empty handed. Sigh.

      This is too real...

    1. Author response:

      The following is the authors’ response to the original reviews.

      We are grateful for the many positive comments. Moreover, we appreciate the recommendations to improve the manuscript; particularly, the important discussion points raised by reviewer 1 and the comments made by reviewer 2 concerning an extended quantification of how near-spike input conductances vary across individual spikes. We have performed several new detailed analyses to address reviewer 2’s comments. In particular, we now provide for all relevant postsynaptic cells the complete distributions of the excitatory and inhibitory input conductance changes that occur right before and after postsynaptic spiking, and we provide corresponding distributions of non-spiking regions as a reference. We performed these analyses separately for different baseline activity levels. Our new results largely support our previous conclusions but provide a much more nuanced picture of the synaptic basis of spiking. To the best of our knowledge, this is the first time that parallel information on input excitation, inhibition and postsynaptic spiking is provided for individual neurons in a biological network. We would argue that our new results further support the fundamental notion that even a reductionist neuronal culture model can give rise to sophisticated network dynamics with spiking – at least partially – triggered by rapid input fluctuations, as predicted by theory. Moreover, it appears that changes in input inhibition are a key mechanism to regulate spiking during spontaneous recurrent network activity. It will be exciting to test whether this holds true for neural circuits in vivo.

      In the following section, we address the reviewers’ comments individually.

      Reviewer 1:

      In this study the authors develop methods to interrogate cultured neuronal networks to learn about the contributions of multiple simultaneously active input neurons to postsynaptic activity. They then use these methods to ask how excitatory and inhibitory inputs combine to result in postsynaptic neuronal firing in a network context.

      The study uses a compelling combination of high-density multi-electrode array recordings with patch recordings. They make ingenious use of physiology tricks such as shifting the reversal potential of inhibitory inputs, and identifying inhibitory vs. excitatory neurons through their influence on other neurons, to tease apart the key parameters of synaptic connections.

      We thank the reviewer for acknowledging our efforts to develop an approach to investigate the synaptic basis of spiking in biological neurons and for appreciating the technical challenges that needed to be overcome.

      The method doesn't have complete coverage of all neurons in the culture, and it appears to work on rather low-density cultures so the size of the networks in the current study is in the low tens.

      (1) It would be valuable to see the caveats associated with the small size of the networks examined here.

      (2) It would be also helpful if there were a section to discuss how this approach might scale up, and how better network coverage might be achieved.

      These are indeed very important points that we should have discussed in more detail. Maximizing the coverage of neurons is critical to our approach, as it determines the number of potential synaptic connections that can be tested. The number of cells that we seeded onto our HD-MEA chip was chosen to achieve monolayer neuronal cultures. As detailed in ‘Materials and Methods -> Electrode selection and long-term extracellular recording of network spiking’, the entire HD-MEA chip (all 26'400 electrodes) was scanned for activity at the beginning of each experiment, and electrodes that recorded spiking activity were subsequently selected. While it is possible that some individual neurons escape detection, since they were not directly adjacent to an electrode, we estimate that a large majority of the active neurons in the culture was covered by our electrode selection method. New generations of CMOS HD-MEAs developed in our laboratory and other groups feature higher electrode densities, larger recording areas, and larger sets of electrodes that can be simultaneously recorded from (e.g., DOI:

      10.1109/JSSC.2017.2686580 & 10.1038/s41467-020-18620-4). These features will substantially improve the coverage of the network and also allow for using larger neuronal networks. As suggested by reviewer 1, we added these points to the Discussion section of the revised manuscript.

      The authors obtain a number of findings on the conditions in which the dynamics of excitatory and inhibitory inputs permit spiking, and the statistics of connectivity that result in this. This is of considerable interest, and clearly one would like to see how these findings map to larger networks, to non-cortical networks, and ideally to networks in-vivo. The suite of approaches discussed here could potentially serve as a basis for such further development.

      (3) It would be useful for the authors to suggest such approaches.

      We are confident that our suite of approaches will open important avenues to study the E & I input basis of postsynaptic spiking in other circuits beyond the in vitro cortical networks studied here. In fact, CMOS HD-MEA probes have been successfully combined with patch clamping in vivo (DIO: 10.1101/370080) and, in principle, the strategies and software tools introduced in our study would be equally applicable in an in vivo context. However, currently available in vitro CMOS HD-MEAs still surpass their in vivo counterparts (e.g., Neuropixels probes) in terms of electrode count. Moreover, using in vitro neural networks enables easy access and better network coverage compared to in vivo conditions. These are the main reasons why we chose an in vitro network for our investigation. We added these points to the Discussion section of the revised manuscript.

      (4) The authors report a range of synaptic conductance waveforms in time. Not surprisingly, E and I look broadly different. Could the authors comment on the implications of differences in time-course of conductance profiles even within E (or I) synapses? Is this functional or is it an outcome of analysis uncertainty?

      We are grateful to the reviewer for raising this interesting point. On the one hand, the onsets of the synaptic conductance waveform estimates were strikingly different between E and I synapses (see Fig. 8D). Furthermore, the rise and decay phases of synaptic currents were distinct for E vs. I inputs (Fig. 4C). We think that these differences are not just due to analysis uncertainty because both these observations are consistent with previously described properties of E and I inputs: Synaptic GABAergic I currents are typically slower compared to Glutamatergic E currents with respect to both rising and decay phase (DOI: 10.1126/science.abj586). Moreover, the relatively small onset latencies for I inputs that we observed are consistent with the well-known local action of inhibition. This finding was also consistent with smaller PRE-POST distances and general differences in neurite characteristics of E compared to I cells (Fig. S2).

      One of the challenges in doing such studies in a dish is that the network is simply ticking away without any neural or sensory context to work on, nor any clear idea of what its outputs might mean. Nevertheless, at a single-neuron level one expects that this system might provide a reasonable subset of the kinds of activity an individual cell might have to work on.

      (5) Could the authors comment on what subsets of network activity is, and is not, likely to be seen in the culture?

      (6) Could they indicate what this would mean for the conclusions about E-I summation, if the in-vivo activity follows different dynamics?

      We agree that there are natural limitations to a reductionist model, such as a dissociated cell culture. One may argue that neuronal cultures bear some similarities with neural networks formed during early brain development, where network formation is primarily driven by intrinsic, self-organizational capabilities. While such a self-organization is likely constrained in a 2D culture, it has been shown that several important circuit mechanisms that are observed in vivo are preserved in 2D dissociated cultures. For example, dissociated neuronal cultures can maintain E-I balance and achieve active decorrelation (DOI: 10.1038/nn.4415). In addition, in terms of activity levels, the sequences of heightened and more quiescent network spiking bear similarities with cortical Up-Down state oscillations observed during slow-wave sleep. To what extent individual circuit connectivity motifs and more nuanced network dynamics, found in vivo, can be recapitulated in vitro, is still not clear. However, combining our and previous work (especially DOI: 10.1038/nn.4415), we believe that there is sufficient evidence to justify work such as ours. On the one hand, identifying in simple cell culture models features of network dynamics and microcircuits known (or predicted) to exist in vivo is a testimony of neuronal self-organizing capabilities. On the other hand, our in vitro results will allow for more directed testing of equivalent mechanisms in vivo.

      Reviewer 2:

      The authors had two aims in this study. First, to develop a tool that lets them quantify the synaptic strength and sign of upstream neurons in a large network of cultured neurons. Second, they aimed at disentangling the contributions of excitatory and inhibitory inputs to spike generation.

      For the quantification of synaptic currents, their methods allows them to quantify excitatory and inhibitory currents simultaneously, as the sign of the current is determined by the neuron identity in the high-density extracellular recording. They further made sure that their method works for nonstationary firing rates, and they did a simulation to characterize what kind of connections their analysis does not capture. They did not include the possibility of (dendritic) nonlinearities or gap junctions or any kind of homeostatic processes.

      Thank you for the concise summary of our aims and of the features of our method. Indeed, we did not model nonlinear synaptic interactions, short-term plasticity etc., as there is likely a spectrum of possible interaction rules. Importantly, non-linear synaptic interactions were reduced by performing synaptic measurements in voltage-clamp mode.

      We do not anticipate that this would impact our connectivity inference per se. However, the presence of a significant number of nonlinear events would imply that some deviations between reconstructed and measured patch current traces were to be expected even if all incoming monosynaptic connections were identified. In the future, it will be exciting to add to our current experimental protocol a simultaneous HD-MEA & patch-clamp recording, in which the membrane potential is measured in current-clamp mode. Following application of our synaptic input-mapping procedure, one could, in this way, directly assess input-sequence dependent non-linear synaptic integration during spontaneous network activity.

      I see a clear weakness in the way that they quantify their goodness of fit, as they only report the explained variance, while their data are quite nonstationary. It could help to partition the explained variance into frequency bands, to at least separate the effects of a bias in baseline, the (around 100 Hz) band of synaptic frequencies and whatever high-frequency observation noise there may be. Another weak point is their explanation of unexplained variance by potential activation of extrasynaptic receptors without providing evidence. Given that these cultures are not a tissue and diffusion should be really high, this idea could easily be tested by adding a tiny amount of glutamate to the culture media.

      As suggested by the reviewer, we have now partitioned the current traces into frequency bands and separately assessed the goodness-of-fit. We have updated Fig. 3C accordingly:

      The following sentence was added to the main text:

      “We separately compared slow baseline changes (< 3 Hz), fast synaptic activity (3 - 200 Hz) and putative high-frequency noise (> 200 Hz), yielding a median variance explained of approximately 60% in the 3 - 200 Hz range (Fig. 3C).”

      Importantly, the variance explained in the frequency range of synaptic activity remains high. We would also like to point out that, even if all synaptic input connections were identified, one would expect some deviations between measured and reconstructed current trace. This is because the reconstructed trace is based on average input current waveforms and in the measured trace there may be synaptic transmission failures.

      We agree that the offered explanation for unexplained variance by activation of extrasynaptic receptors is fairly speculative. As it was not a crucial discussion point, we have therefore removed the statement.

      For the contributions of excitation and inhibition to neuronal spiking, the authors found a clear reduction of inhibitory inputs and increase of excitation associated with spiking when averaging across many spikes. And interestingly, the inhibition shows a reversal right after a spike and the timescale is faster during higher network activity. While these findings are great and provide further support that their method is working, they stop at this exciting point where I would really have liked to see more detail.

      Thank you for acknowledging our main results concerning the synaptic basis of spiking. We attempted to integrate in one manuscript a suite of new approaches, in addition to the respective applications. We, therefore, tried to strike the appropriate level of detail in presenting our findings. With regard to our analyses of which synaptic input events regulate postsynaptic spiking, we agree with reviewer 2’s assessment that more detail concerning the variability across individual spikes would be helpful. In the following parts, we detail multiple new analyses that we have included in the revised manuscript to address reviewer 2’s comments.

      A concern, of course, is that the network bursts in cultures are quite stereotypical, and that might cause averages across many bursts to show strange behaviour. So what I am missing here is a reference or baseline or null hypothesis. How does it look when using inputs from neurons that are not connected? And then, it looks like the E/(E+I) curve has lots of peaks of similar amplitude (that could be quantified...), so why does the neuron spike where it does? If I would compare to the peak (of similar amplitude) right before or right after (as a reference) are there some systematic changes? Is maybe the inhibition merely defining some general scaffold where spikes can happen and the excitation causes the spike as spiking is more irregular?

      The averaged trace reveals a different timescale for high and low activity states. But does that reflect a superposition of EPSCs in a single trial or rather a different jittering of a single EPSC across trials? For answering this question, it would be good to know the variance (and whether/ how much it changes over time). Maybe not all spikes are preceded by a decrease in inhibition. Could you quantitify (correlate, scatterplot?) how exactly excitation and inhibition contributions relate for single postsynaptic spikes (or single postsynaptic non-spikes)? After all, this would be the kind of detail that requires the large amount of data that this study provides.

      First of all, we are very grateful for the reviewer’s thorough assessment of our work and for the many valuable suggestions to improve it. We are convinced that we have addressed with our new analyses and the updated manuscript all issues raised here. One of the main findings from our original manuscript was that a rapid and brief change in input conductance (and particularly a reduction in inhibition) is an important spike trigger/regulator. We followed the reviewer’s suggestion and now provide scatter plots and distributions of the pre- (and post-spike) changes in input excitation and inhibition for individual postsynaptic spikes. A quantification of the peaks in the noisy E/(E+I) traces was not always trivial, which is why we reasoned that an assessment of the respective E and I changes is better suited. Moreover, as an unbiased reference, we generated separately for each postsynaptic cell a corresponding distribution of changes in input conductance in non-spiking periods (using random time points). We included our new results and updated figures in our responses to the specific reviewer comments below.

      For the first part, the authors achieved their goal in developing a tool to study synaptic inputs driving subthreshold activity at the soma, and characterizing such connections. For the second part, they found an effect of EPSCs on firing, but they barely did any quantification of its relevance due to the lack of a reference.

      With the availability of Neuropixels probes, there is certainly use for their tool in in vivo applications, and their statistical analysis provides a reference for future studies.

      The relevance of excitatory and inhibitory currents on spiking remains to be seen in an updated version of the manuscript.

      Thank you. Please see our new analyses below. Our new findings are in agreement with the main conclusions of the original manuscript. We provide evidence that rapid pre-spike changes in input conductance are observed across most individual spikes and that these rapid changes occur significantly more often before measured spikes than in non-spiking periods.

      I feel that specifically Figures 6 and 7 lack relevant detail and a consistent representation that would allow the reader to establish links between the different panels. The analysis shows very detailed examples, but then jumps into analyses that show population averages over averaged responses, losing or ignoring the variability across trials. In addition, while their results themselves pass a statistical test, it is crucial to establish some measure of how relevant these results are. For that, I would really want to know how much spiking would actually be restricted by the constraints that would be posed by these results, i.e. would this be reflected in tiny changes in spiking probabilities, or are there times when spiking probabilities are necessarily high, or do we see times when we would almost certainly get a spike, but neurons can fire during other times as well.

      I would agree that a detailed, quantitative analysis of this question is beyond the scope of this paper, but a qualitative analysis is feasible and should be done.

      Please see our revised Figure 6. We have rearranged some of the original panels and removed one example of mean conductance profiles. Moreover, we removed a panel with analysis results based on mean conductances that is now obsolete, as more detailed analyses are provided (which are in agreement with the original findings). Analyses from panels (A-F) are mostly unchanged. Panels (G-J) show the new results.

      The following paragraphs, which were added to the main text of the revised manuscript, describe our new findings:

      “For a more nuanced picture of which synaptic events are associated with postsynaptic spiking, we next quantified the changes in input excitation and inhibition that preceded individual postsynaptic spikes. In our analysis, we first focused on periods with high synaptic input activity. As previously discussed, cortical neurons in vivo typically receive and integrate barrages of input activation, similar to the high-activity events that we observed here (e.g., the event depicted in Fig. 6A, right). In Fig. 6G/H, individual pre-spike changes in input conductance are shown for two example postsynaptic neurons (plots labeled ‘spiking’, right). To assess how specific these conductance changes were to spiking periods, we also quantified the changes in input conductance that occurred during non-spiking periods as a reference (we used random time points from high-activity events excluding time points adjacent to measured spike times; we upscaled the number of measured spikes by 10x; the respective plots were labeled ‘non-spiking’). Spikes of both example neurons exhibited – compared to non-spiking regions – significantly more often a pre-spike decrease in inhibition, consistent with the mean conductance profiles. Precisely how an increase (top-right quadrants in Fig. 6G/H) or decrease (bottom-left quadrants) in both I and E conductance influenced the neuronal membrane potential is difficult to predict. However, if rapid changes in input conductance had a significant role in triggering spikes, one would expect that fewer spikes would exhibit a hyperpolarizing pre-spike increase in I and decrease in E (top-left quadrant) compared to the non-spiking period. Conversely, a decrease in I and an increase E (bottom-right quadrants) would likely result in a membrane potential depolarization so that more spikes should feature the corresponding pre-spike conductance changes compared to non-spiking periods. These relative shifts are precisely what can be observed in the plots of the two example neurons (Fig. 6G/H) and, in fact, across recordings (Fig. 6I). Finally, we compared the distributions of pre-spike changes in input inhibition and excitation of each postsynaptic neuron (Fig. 6J). Further indicating a pivotal role of inhibition in triggering spikes, 6 out of 7 neurons exhibited a clear decrease in the mean values (and medians) of pre-spike changes in inhibition compared to non-spiking periods. Interestingly, the 3 out of 7 neurons with an increase in excitation showed the smallest decrease in inhibition (or even an increase in inhibition in case of neuron #7). This latter observation suggests a matching of E and I inputs and cell-specific relative contributions of E and I conductance changes in triggering spikes.

      Theoretically, neuronal spiking could be driven by a prolonged suprathreshold depolarization (Petersen and Berg 2016; Renart et al. 2007) or, in more favorable subthreshold regimes, by fast synaptic input fluctuations (Ahmadian and Miller 2021; Amit and Brunel 1997; Brunel 2000; Van Vreeswijk and Sompolinsky 1996). In this section, we demonstrated that the majority of investigated neurons featured – during high-activity periods – a significant number of spikes that were associated with rapid pre-spike changes in input conductances. These findings suggest that even simple neuronal cultures can self-organize to form circuits exhibiting sophisticated spiking dynamics.”

      Our new analyses detailed in Fig. 6 show that there are also presumably depolarizing events (e.g., decrease in I and increase in E) in non-spiking regions. In future studies, it will be interesting to examine what distinguishes these events from spike-inducing events of similar magnitude – one possibility is a dependency on specific input-activation sequences.

      During the first days and weeks of developing neuronal cultures, spiking activity rapidly shifts from synapse-independent activity patterns to spiking dynamics that do depend on synaptic inputs and are progressively organized in network-wide high-activity events (DOI: 10.1016/j.brainres.2008.06.022). In our study, cultures at days-in-vitro 15-18 were used, and approximately 15% of the spikes occurred during high-activity events with relatively strong E and I input activity. In addition, spikes that occurred during low-activity events were at least partially regulated by synaptic input (see answers below related to Fig. 7).

      In the following, I am detailing what I would consider necessary to be done about these two Figures:

      Figure 6C is indeed great, though I don't see why the authors would characterize synchrony as low. When comparing with Figure 4B, I'd think that some of these values are quite high. And it wouldn't help me to imagine error bars in panel 6D.

      We have removed our characterization as ‘low’ from the text. One important difference between our synchrony measure (STTC) and the quantification of spike-transmission probability (STP) is the ‘lag’ of a few milliseconds for the STP quantification window to account for synaptic delay.

      Figure 6B is useful, but could be done better: The autocovariance of a shotnoise process is a convolution of the autocovariance of underlying point process and the autocovariance of the EPSC kernel. So one would want to separate those to obtain a better temporal resolution. But a shotnoise process has well defined peaks, and the time of these local maxima can be estimated quite precisely. Now if I would do a peak triggered average instead of the full convolution, I would do half of the deconvolution and obtain a temporally asymmetric curve of what is expected to happen around an EPSC. Importantly, one could directly see expected excitation after inhibition or expected inhibition after excitation, and this visualization could be much better and more intuitively compared to panel 6E.

      We appreciate the reviewer’s suggestion to present these results in a more sophisticated way. We would like to propose to stick with the original analysis to have it comparable with related analyses from the literature (e.g., DOI: 10.1038/nn.2105). Therefore, we hope the reviewer finds it acceptable that we leave the presentation of the data in its original form and potentially follow up in future work with the analysis strategy proposed by the reviewer.

      Panel D needs some variability estimate (i.e. standard deviation or interquartile range or even a probability density) for those traces.

      Figure 6E: Please use more visible colors. A sensitivity analysis to see traces for 2E/(2E+I) and E/(E+2I) would be great.

      Figure 6F: with an updated panel B, we should be able to have a slope for average inhibition after excitation for each of these cells. A second panel / third column showing those slopes would be of interest. It would serve as a reference for what could be expected from E-I interactions alone.

      With regard to the variability estimate in D, we now provide multiple panels characterizing the variability. For one, Fig. 6H contains a scatter plot of the pre-spike changes in input conductance across all individual postsynaptic spikes from the example cell shown in D. Moreover, in Fig. 7A, we show from the same example cell the standard deviations associated with the mean conductance traces separately for spikes that occurred during low- and high-activity states. For better visibility and because the separation according to activity states is more informative, we kept the original presentation of panel D (however, removing one example cell). In addition, we show the same mean traces from panel D with the respective standard deviations (across all spikes) in Supplementary Figure S3.

      Colors in Fig. 6E are adjusted, as requested.

      We have removed panel Fig. 6F as we now provide more detailed analyses at single-spike level (see Fig. 6G-J).

      Figure 6G: Could the authors provide an interquartile range here?

      With regard to the aligned input-output data from original panel Fig. 6G, now in panel Fig. 6F in the updated figure version, we show all individual traces that were averaged: the E/I traces from panel Fig. 6E and the three action potential waveforms from Supplementary Figure S5. Therefore, we chose to present the means only for better visibility.

      Figure 7A: it may be hard to squeeze in variability estimates here, but the information on whether and how much variance might be explained is essential. Maybe add another panel to provide a variability estimate? The variability estimate in panel 7B and 7D only reflect variability across connections, and it would be useful to add panels for the time courses of the variability of g (or E/(E+I) respectively).

      We now include the standard deviations across the input conductance traces in the updated Fig. 7A, as requested. We have also simplified Fig. 7 and performed the analysis using the 6 out of 7 neurons that, based on our new analysis (Fig. 6J) displayed a clear reduction in pre-spike inhibition, relative to the reference distribution. For a complete overview of the state-dependent changes in input conductance that are associated with individual postsynaptic spikes, we have included a new supplementary figure (Fig. S6). Fig. S6 also includes a characterization of the changes in input inhibition that occur right after postsynaptic spiking. In addition, Fig. S6D shows the standard deviations corresponding to the mean input conductance traces of all cells – separately for high- and low-activity periods.

      We added the following paragraph to the main text of the revised manuscript:

      “How can these deviations in the mean conductance profiles be explained? To answer this question, we further quantified – separately for low and high g states – the changes in input inhibition that occurred right before and after individual postsynaptic spikes (Fig. S6). This single-spike analysis suggested that, during high g states, most spikes experienced a post-spike increase and pre-spike decrease in inhibition (see also Fig. 6J). On the other hand, low g states were characterized by sparse synaptic input (e.g., see reconstruction in Fig. 6A). Therefore, many of the spikes that occurred during low g states were associated with little change in input conductance (note medians of approximately zero in Fig. S6A/C). Nevertheless, a considerable fraction of spikes (often > 25%) from low g states were also associated with a post-spike increase and pre-spike drop in inhibition. It, therefore, appears that even the sparse inhibitory inputs of low g states could influence spike timing. Moreover, the post-spike increases in input inhibition during low g states suggest that there were strong regulatory inhibitory circuits in place. However, limited activity levels during low g states presumably introduced an increased jitter of these spike-associated changes in input inhibition.

      In summary, the input inhibition of high-conductance states provides reliable and narrow windows-of-spiking opportunity. In addition, even during periods of sparse activity, there are rudimentary synaptic mechanisms in place to regulate spike timing.”

      As a suggestion for further analysis, though I am well aware that this is likely beyond the scope of this manuscript, I'd suggest the following analysis:

      I would split the data into the high and low activity states. Then I would compute the average of E/(E+I) values for spikes. Assuming that spikes tend to happen for local maxima of E/(E+I) I would find local maxima for periods without spike such that their average is equal to the value for actual spikes. Finally, I would test for a systematic difference in either excitation or inhibition.

      If there is no difference, you can make the claim that synaptic input does not guarantee a spike, and compare to a global average of E/(E+I).

      We are grateful for the fantastic suggestions for future analysis. We look forward to conducting these analyses in a more detailed follow-up characterization.

      In addition to the major alterations detailed above, we performed smaller corrections (e.g., spelling mistakes, inaccuracies) in some parts of the manuscript.

    1. He found the house gone to decay—the roof fallen in, the windows shattered, and the doors off the hinges. A half-starved dog that looked like Wolf was skulking about it. Rip called him by name, but the cur snarled, showed his teeth, and passed on. This was an unkind cut indeed. “My very dog,” sighed Rip, “has forgotten me!”

      the big realization

    1. With Blue - uncertain - stumbling Buzz - Between the light - and me - And then the Windows failed - and then I could not see to see -

      After reading this the way I took it was that a windows job when its closed it to keep things out of the house and only bring in the natural light from outside. It seems as she was trying to catch the fly and at that point the fly flies out the open window which is why it says the windows fail. Chasing the fly she looks into the light blinging her and losing the fly

    2. And then the Windows failed - and then I could not see to see -

      These lines are very symbolic and represent last a few feelings of a person dying. Last sensory/motor skills of the brain. Last connection with nature. Last connection with this living world before pass to soul world. Such as in this these lines represent every single last touch, hearing and seeing sign. Although, the line "I could not see to see" might express the last late hope of living.

    3. And then the Windows failed - and then I could not see to see -

      "The Windows" in this metaphor refers to the speaker's eyes or ability to see. When the speaker gets closer to passing away, the "Windows failed," signifying his losing vision as his life approaches to an end. By associating the closure of windows with the actual act of dying, such as losing consciousness and sight, this metaphor suggests a barrier separating the individual from the outside world as they approach death. The use of the phrase "I could not see to see" highlights the speaker's complete loss of vision and reinforces their incapacity to notice anything beyond this last instant.

    1. Author response:

      The following is the authors’ response to the original reviews.

      We are deeply appreciative of the reviewers' insightful comments and constructive feedback on our manuscript. In response, we have implemented substantial revisions to enhance the clarity and impact of our work. Key changes include: 

      Reframing: We have shifted our focus from cognitive control to attention and memory processes, aligning more closely with our experimental design. This reframing is reflected throughout the manuscript, including additional citations highlighting the triple network model's involvement in memory processing. To reflect this change, we have updated the title to "Causal dynamics of salience, default mode, and frontoparietal networks during episodic memory formation and recall: A multi-experiment iEEG replication".

      Control analyses using resting-state epochs: We have conducted new analyses comparing task periods to resting baseline epochs. These results demonstrate enhanced directed information flow from the anterior insula to both the default mode and frontoparietal networks during encoding and recall periods compared to resting state across all four experiments. This finding underscores the anterior insula's critical role in memory and attention processing.

      Control analysis using the inferior frontal gyrus: To address specificity concerns, we performed control analyses using the inferior frontal gyrus as a comparison region. This analysis confirms that the observed directed information flow to the default mode and frontoparietal networks is specific to the anterior insula, rather than a general property of task-engaged brain regions.

      These revisions, combined with our rigorous methodologies and comprehensive analyses, provide compelling support for the central claims of our manuscript. We believe these changes significantly enhance the scientific contribution of our work.

      Our point-by-point responses to the reviewers' comments are provided below.

      Reviewer 1:

      -  The authors present results from an impressively sized iEEG sample. For reader context, this type of invasive human data is difficult and time-consuming to collect and many similar studies in high-level journals include 5-20 participants, typically not all of whom have electrodes in all regions of interest. It is excellent that they have been able to leverage open-source data in this way. 

      -  Preprocessing of iEEG data also seems sensible and appropriate based on field standards. 

      -  The authors tackle the replication issues inherent in much of the literature by replicating findings across task contexts, demonstrating that the principles of network communication evidenced by their results generalize in multiple task memory contexts. Again, the number of iEEG patients who have multiple tasks' worth of data is impressive. 

      We thank the reviewer for the encouraging comments and appreciate the positive feedback.  

      (1.1) The motivation for investigating the tripartite network during memory tasks is not currently well-elaborated. Though the authors mention, for example, that "the formation of episodic memories relies on the intricate interplay between large-scale brain networks (p. 4)", there are no citations provided for this statement, and the reader is unable to evaluate whether the nodes and networks evidenced to support these processes are the same as networks measured here. 

      Recommendation: Detail with citations the motivation for assessing the tripartite network in these tasks. Include work referencing network-level and local effects during encoding and recall.

      We appreciate the reviewer's feedback and suggestions for improving our framing. We have substantially expanded and revised the Introduction to elaborate on the motivation for investigating the tripartite network during memory tasks, supported by relevant citations.

      We now provide a stronger rationale for examining these networks in the context of episodic memory, emphasizing that while the tripartite network has been extensively studied in cognitive control tasks, growing evidence suggests its relevance to episodic memory as a domain-general network. We cite several key studies that demonstrate the involvement of the salience, default mode, and frontoparietal networks in memory processes, including work by Sestieri et al. (2014) and Vatansever et al. (2021), which show the consistent engagement of these networks during memory tasks. We have also included references to studies examining network-level and local effects during encoding and recall, such as the work by Xie et al. (2012) on disrupted intrinsic connectivity in amnestic mild cognitive impairment, and Le Berre et al. (2017) on the role of insula connectivity in memory awareness (pages 4-5).

      Furthermore, we have clarified how our study aims to address gaps in the current understanding by investigating the electrophysiological basis of these network interactions during memory formation and retrieval, which has not been explored in previous research. This expanded framing provides a clearer motivation for our investigation and places our study within the broader context of memory and network neuroscience research (pages 3-6).  

      (1.2) In addition, though the tripartite network has been proposed to support cognitive control processes, and the neural basis of cognitive control is the framed focus of this work, the authors do not demonstrate that they have measured cognitive control in addition to simple memory encoding and retrieval processes. Tasks that have investigated cognitive control over memory (such as those cited on p. 13 - Badre et al., 2005; Badre & Wagner, 2007; Wagner et al., 2001; Wagner et al., 2005) generally do not simply include encoding, delay, and recall (as the tasks used here), but tend to include some manipulation that requires participants to engage control processes over memory retrieval, such as task rules governing what choice should be made at recall (e.g., from Badre et al., 2005 Fig. 1: congruency of match, associative strength, number of choices, semantic similarity). Moreover, though there are task-responsive signatures in the nodes of the tripartite networks, concluding that cognitive control is present because cognitive control networks are active would be a reverse inference.

      Recommendation: If present, highlight components of the tasks that are known to elicit cognitive control processes and cite relevant literature. If the tasks cannot be argued to elicit cognitive control, reframe the motivation to focus on task-related attention or memory processes. If the latter, reframe the motivation for investigating the tripartite network in this context absent control.

      We appreciate the reviewer's insightful comment and recommendation. We acknowledge that our tasks do not include specific manipulations designed to elicit cognitive control processes over memory retrieval. In light of this, we have reframed our motivation and discussion to focus on the role of the tripartite network in attention and memory processes more broadly, rather than cognitive control specifically (pages 3-6).

      As noted in Response 1.1, we have revised the Introduction to emphasize the domain-general nature of these networks and their involvement in various cognitive processes, including memory. We also highlight how the salience, default mode, and frontoparietal networks contribute to different aspects of memory formation and retrieval, drawing on relevant literature.

      Our revised framing examines the salience network's role in detecting behaviorally relevant stimuli and orienting attention during encoding, the default mode network's involvement in internally-driven processes during recall, and the frontoparietal network's contribution to maintaining and manipulating information in working memory. We now present our study as an investigation into how these networks interact during different phases of memory processing, rather than focusing specifically on cognitive control. This approach aligns better with our experimental design and allows us to explore the broader applicability of the tripartite network model to memory processes. 

      This revised reframing provides a more accurate representation of our study's scope and contribution to understanding the role of large-scale brain networks in memory formation and retrieval (pages 3-6). 

      (1.3) It is currently unclear if the directed information flow from AI to DMN and FPN nodes truly arises from task-related processes such as cognitive control or if it is a function of static brain network characteristics constrained by anatomy (such as white matter connection patterns, etc.). This is a concern because the authors did not find that influences of AI on DMN or FPN are increased relative to a resting baseline (collected during the task) or that directed information flow differs in successful compared to unsuccessful retrieval. I doubt that this AI influence is 1) supporting a switch between the DMN and FPN via the SN or 2) relevant for behavior if it doesn't differ from baseline-active task or across accuracy conditions. An additional comparison that may help investigate whether this is reflective of static connectivity characteristics would be a baseline comparison during non-task rest or sleep periods.  

      Recommendation: As described in the task of the concern, analyze the PTE across the same contacts during sleep or task-free rest periods (if present in the dataset). 

      We thank the reviewer for this suggestion. We have now carried out additional analyses using resting-state baseline epochs. We found that directed information flow from the AI to both the DMN and FPN were enhanced during the encoding and recall periods compared to resting-state baseline in all four experiments. These new results have now been included in the revised Results (page 12):    

      “Enhanced information flow from the AI to the DMN and FPN during episodic memory processing, compared to resting-state baseline  

      We next examined whether directed information flow from the AI to the DMN and FPN nodes during the memory tasks differed from the resting-state baseline. Resting-state baselines were extracted immediately before the start of the task sessions and the duration of task and rest epochs were matched to ensure that differences in network dynamics could not be explained by differences in duration of the epochs. Directed information flow from the AI to both the DMN and FPN were higher during both the memory encoding and recall phases and across the four experiments, compared to baseline in all but two cases (Figures S6, S7). These findings provide strong evidence for enhanced role of AI directed information flow to the DMN and FPN during memory processing compared to the resting state.” 

      (1.4) Related to the above concern, it is also questionable how directed information flow from AI facilitates switching between FPN and DMN during both encoding and recall if high gamma activity does not significantly differ in AI versus PCC or mPFC during recall as it does during encoding. It seems erroneous to conclude that the network-level communication is happening or happening with the same effect during both task time points when these effects are decoupled in such a way from the power findings.  

      We appreciate the reviewer's insightful observation regarding the apparent discrepancy between our directed information flow findings and the high-gamma activity results. This comment highlights an important distinction in interpreting our results, and we thank the reviewer for the opportunity to address this point.

      Our findings demonstrate that directed information flow from the AI to the DMN and FPN persists during both encoding and recall, despite differences in local high-gamma activity patterns. This dissociation suggests that the network-level communication facilitated by the AI may operate independently of local activation levels in individual nodes. It is important to note that our directed connectivity analysis (using phase transfer entropy) was conducted on broadband signals (0.5-80 Hz), while the power analysis focused specifically on the high-gamma band (80-160 Hz). These different frequency ranges may capture distinct aspects of neural processing. The broadband connectivity might reflect more general, sustained network interactions, while high-gamma activity may be more sensitive to specific task demands or cognitive processes.

      The phase transfer entropy analysis captures directed interactions over extended time periods, while the high-gamma power analysis provides a more temporally precise measure of local neural activity. The persistent directed connectivity from AI during recall, despite changes in local activity, might reflect the AI's ongoing role in coordinating network interactions, even when its local activation is not significantly different from other regions.

      Rather than facilitating "switching" between FPN and DMN, as we may have previously overstated, our results suggest that the AI maintains a consistent pattern of influence on both networks across task phases. This influence might serve different functions during encoding (e.g., orienting attention to external stimuli) and recall (e.g., monitoring and evaluating retrieved information), even if local activation patterns differ.

      It is crucial to note that in the three verbal tasks, our analysis of memory recall is time-locked to word production onset. However, the precise timing of the internal recall process initiation is unknown. This limitation may affect our ability to capture the full dynamics of network interactions during recall, particularly in the early stages of memory retrieval. Interestingly, in the spatial memory task WMSM, the PCC/precuneus exhibited an earlier onset and enhanced activity compared to the AI. This task may provide a clearer window into recall processes:

      findings align with the view that DMN nodes may play a crucial role in triggering internal recall processes. However, the precise timing of internal retrieval initiation remains a challenge in the three verbal tasks, potentially limiting our ability to capture the full dynamics of regional activity, and its replicability, during early stages of recall.

      These observations highlight the need for more detailed investigation of the temporal dynamics of network interactions during memory recall. To further elucidate the relationship between directed connectivity and local activity, future studies could employ time-resolved connectivity analyses and investigate coupling between different frequency bands. This could provide a more precise understanding of how network-level communication relates to local neural dynamics across different task phases.

      We have revised the manuscript to more accurately reflect these points and avoid overstating the implications of our findings (pages 15-19). We thank the reviewer for prompting this important clarification, which we believe strengthens the interpretation and discussion of our results.

      (1.5) Missing information about the methods used for time-frequency conversion for power calculation and the power normalization/baseline-correction procedure bars a thorough evaluation of power calculation methods and results. 

      Recommendation: Include more information about how power was calculated. For example, how were time-series data converted to time-frequency (with complex wavelets, filter-hilbert, etc.)? What settings were used (frequency steps, wavelet length)? How were power values checked for outliers and normalized (decibels, Z-transform)? How was baseline correction applied (subtraction, division)?

      We have now included detailed information related to our power calculation and normalization steps as we note on page 28: “We first filtered the signals in the high-gamma (80160 Hz) frequency band (Canolty et al., 2006; Helfrich & Knight, 2016; Kai J. Miller, Weaver, & Ojemann, 2009) using sequential band-pass filters in increments of 10 Hz (i.e., 80–90 Hz, 90– 100 Hz, etc.), using a fourth order two-way zero phase lag Butterworth filter. We used these narrowband filtering processing steps to correct for the 1/f decay of power. We then calculated the amplitude (envelope) of each narrow band signal by taking the absolute value of the analytic signal obtained from the Hilbert transform (Foster, Rangarajan, Shirer, & Parvizi, 2015). Each narrow band amplitude time series was then normalized to its own mean amplitude, expressed as a percentage of the mean. Finally, we calculated the mean of the normalized narrow band amplitude time series, producing a single amplitude time series. Signals were then smoothed using 0.2s windows with 90% overlap (Kwon et al., 2021) and normalized with respect to 0.2s pre-stimulus periods by subtracting the pre-stimulus baseline from the post-stimulus signal.” 

      (1.6) If revisions to the manuscript can address concerns about directed information flow possibly being due to anatomical constraints - such as by indicating that directed information flow is not present during non-task rest or sleep - this work may convey important information about the structure and order of communication between these networks during attention to tasks in general. However, the ability of the findings to address cognitive control-specific communication and the nature of neurophysiological mechanisms of this communication - as opposed to the temporal order and structure of recruited networks - may be limited.

      We appreciate the reviewer's insightful feedback, which has led to significant improvements in our manuscript. In response, we have made the following key changes. We have shifted our focus from cognitive control to the broader roles of the tripartite network in attention and memory processes. This reframing aligns more closely with our experimental design and the nature of our tasks. We have revised the Introduction, Results, and Discussion sections to reflect this perspective, providing a more accurate representation of our study's scope and contribution. Additionally, to strengthen our findings, we have conducted new analyses comparing task periods to resting-state baselines. These analyses revealed that directed information flow from the anterior insula to both the DMN and FPN was significantly enhanced during memory encoding and recall periods compared to resting-state across all four experiments. This finding provides robust evidence for the specific involvement of these network interactions in memory processing. Please also see Response 1.2 above. 

      (1.7) Because phase-transfer entropy is presented as a "causal" analysis in this investigation (PTE), I also believe it is important to highlight for readers recent discussions surrounding the description of "causal mechanisms" in neuroscience (see "Confusion about causation" section from Ross and Bassett, 2024, Nature Neuroscience). A large proportion of neuroscientists (admittedly, myself included) use "causal" only to refer to a mechanism whose modulation or removal (with direct manipulation, such as by lesion or stimulation) is known to change or control a given outcome (such as a successful behavior). As Ross and Bassett highlight, it is debatable whether such mechanistic causality is captured by Granger "causality" (a.k.a. Granger prediction) or the parametric PTE, and the imprecise use of "causation" may be confusing. The authors could consider amending language regarding this analysis if they are concerned about bridging these definitions of causality across a wide audience. 

      We thank the reviewer for this suggestion. We would like to clarify here that we define causality in our manuscript as follows: a brain region has a causal influence on a target if knowing the past history of temporal signals in both regions improves the ability to predict the target's signal in comparison to knowing only the target's past, as defined in earlier studies (Granger, 1969; Lobier, Siebenhühner, Palva, & Matias, 2014). We have now included this clarification in the Introduction section (page 6).  

      We also agree with the reviewer that to more mechanistically establish a causal link between the neural dynamics and behavior, lesion or brain stimulation studies are necessary. We have now acknowledged this in the revised Discussion as we note: “Although our computational methods suggest causal influences, direct causal manipulations, such as targeted brain stimulation during memory tasks, are needed to establish definitive causal relationships between network nodes.” (page 19). 

      Minor additional information that would be helpful to the reader to include: 

      (1.8) How exactly was line noise (p. 24) removed? (For example, if notch filtered, how were slight offsets of the line noise from exactly 60.0Hz and harmonics identified and handled?). 

      We would like to clarify here that to filter line noise and its harmonics, we used bandstop filters at 57-63 Hz, 117-123 Hz, and 177-183 Hz. To create a band-stop filter, we used a fourth order two-way zero phase lag Butterworth filter. This information has now been included in the revised Methods (page 26). 

      (1.9) Why were the alpha and beta bands collapsed for narrowband filtering?

      Please note that we did not combine the alpha (8-12 Hz) and beta (12-30 Hz) bands for narrowband filtering, rather these two frequency bands were analyzed separately. However, we combined the delta (0.5-4 Hz) and theta (4-8 Hz) frequency bands into a combined delta-theta (0.5-8 Hz) frequency band for our analysis since previous human electrophysiology studies have not settled on a specific band of frequency (delta or theta) for memory processing. Previous human iEEG (Ekstrom et al., 2005; Ekstrom & Watrous, 2014; Engel & Fries, 2010; Gonzalez et al., 2015; Watrous, Tandon, Conner, Pieters, & Ekstrom, 2013) as well as scalp EEG and MEG studies, have shown that both the delta and theta frequency band oscillations play a prominent role for human memory encoding as well as retrieval (Backus, Schoffelen, Szebényi, Hanslmayr, & Doeller, 2016; Clouter, Shapiro, & Hanslmayr, 2017; Griffiths, Martín-Buro, Staresina, & Hanslmayr, 2021; Guderian & Düzel, 2005; Guderian, Schott, Richardson-Klavehn, & Düzel, 2009).  

      Reviewer 2:

      In this study, the authors leverage a large public dataset of intracranial EEG (the University of Pennsylvania RAM repository) to examine electrophysiologic network dynamics involving the participation of salience, frontoparietal, and default mode networks in the completion of several episodic memory tasks. They do this through a focus on the anterior insula (AI; salience network), which they hypothesize may help switch engagement between the DMN and FPN in concert with task demands. By analyzing high-gamma spectral power and phase transfer entropy (PTE; a putative measure of information "flow"), they show that the AI shows higher directed PTE towards nodes of both the DMN and FPN, during encoding and recall, across multiple tasks. They further demonstrate that high-gamma power in the PCC/precuneus is decreased relative to the AI during memory encoding. They interpret these results as evidence of "triple-network" control processes in memory tasks, governed by a key role of the AI. 

      I commend the authors on leveraging this large public dataset to help contextualize network models of brain function with electrophysiological mechanisms - a key problem in much of the fMRI literature. I also appreciate that the authors emphasized replicability across multiple memory tasks, in an effort to demonstrate conserved or fundamental mechanisms that support a diversity of cognitive processes. However, I believe that their strong claims regarding causal influences within circumscribed brain networks cannot be supported by the evidence as presented. In my efforts to clearly communicate these inadequacies, I will suggest several potential analyses for the authors to consider that might better link the data to their central hypotheses.

      We thank the reviewer for the encouraging comments and suggestions for improving the manuscript. Please see our detailed responses and clarifications below. 

      (2.1) As a general principle, the effects that the authors show - both in regards to their highgamma power analysis and PTE analysis - do not offer sufficient specificity for a reader to understand whether these are general effects that may be repeated throughout the brain, or whether they reflect unique activity to the networks/regions that are laid out in the Introduction's hypothesis. This lack of specificity manifests in several ways, and is best communicated through examples of control analyses. 

      We appreciate the reviewer's insightful comment regarding the specificity of our findings. We agree that additional analyses could provide valuable context for interpreting our results. In response, we have conducted the following additional analyses and made corresponding revisions to the manuscript:

      Following the reviewer's suggestion, we have selected the inferior frontal gyrus (IFG, BA 44) as a control region. The IFG serves as an ideal control region due to its anatomical adjacency to the AI, its involvement in a wide range of cognitive control functions including response inhibition (Cai, Ryali, Chen, Li, & Menon, 2014), and its frequent co-activation with the AI in fMRI studies. Furthermore, the IFG has been associated with controlled retrieval of memory (Badre et al., 2005; Badre & Wagner, 2007; Wagner et al., 2001), making it a compelling region for comparison. We repeated our PTE analysis using the IFG as the source region, comparing its directed influence on the DMN and FPN nodes to that of the AI.  

      Our analysis revealed a striking contrast between the AI and IFG in their patterns of directed information flow. While the AI exhibited strong causal influences on both the DMN and FPN, the IFG showed the opposite pattern. Specifically, both the DMN and FPN demonstrated higher influence on the IFG than the reverse, during both encoding and recall periods, and across all four memory experiments (Figures S4, S5). 

      These findings highlight the unique role of the AI in orchestrating large-scale network dynamics during memory processes. The AI's pattern of directed information flow stands in contrast to that of the IFG, despite their anatomical proximity and shared involvement in cognitive control processes. This dissociation underscores the specificity of the AI's function in coordinating network interactions during memory formation and retrieval. These results have now been included in our revised Results on page 11.  

      (2.2) First, the PTE analysis is focused solely on the AI's interactions with nodes of the DMN and FPN; while it makes sense to focus on this putative "switch" region, the fact that the authors report significant PTE from the AI to nodes of both networks, in encoding and retrieval, across all tasks and (crucially) also at baseline, raises questions about the meaningfulness of this statistic. One way to address this concern would be to select a control region that would be expected to have little/no directed causal influence on these networks and repeat the analysis. Alternatively (or additionally), the authors could examine the time course of PTE as it evolves throughout an encoding/retrieval interval, and relate that to the timing of behavioral events or changes in high-gamma power. This would directly address an important idea raised in their own Discussion, "the AI is wellpositioned to dynamically engage and disengage with other brain areas."  

      Please see Response 2.1 above for additional analyses related to control region.  

      We also appreciate the reviewer's suggestion regarding time-resolved PTE analysis. However, it's important to note that our current methodology does not allow for such fine-grained temporal analysis. This is due to the fact that PTE, which is an information theoretic measure and relies on constructing histograms of occurrences of singles, pairs, or triplets of instantaneous phase estimates from the phase time-series (Hillebrand et al., 2016) (Methods), requires sufficient number of cycles in the phase time-series for its reliable estimation (Lobier et al., 2014). PTE is based on estimating the time-delayed directed influences from one time-series to the other and its estimate is the most accurate when a large number of time-points (cycles) are available (Lobier et al., 2014). Since our encoding and recall epochs in the verbal recall tasks were only 1.6 seconds long, which corresponds to only 800 time-points with a 500 Hz sampling rate, we used the entire encoding and recall epochs for the most efficient estimate of PTE, rather than estimating PTE in a time-resolved manner. Please note that this is consistent with previous literature which have used ~ 225000 time-points (3 minutes of resting-state data with 1250 Hz sampling rate) for estimating PTE, for example, see (Hillebrand et al., 2016). 

      This limitation prevents us from examining how directed connectivity evolves throughout the encoding and retrieval intervals on a moment-to-moment basis. Future studies employing longer task epochs or alternative methods for time-resolved connectivity analysis could provide valuable insights into the dynamic engagement and disengagement of the AI with other brain areas based on task demands. Such analyses could potentially reveal task-specific temporal patterns in the AI's influence on DMN and FPN nodes during different phases of memory processing.

      Finally, it is crucial to note that in the three verbal tasks, our analysis of memory recall is timelocked to word production onset. However, the precise timing of the internal retrieval process initiation is unknown. This limitation may affect our ability to capture the full dynamics of network interactions during recall, particularly in the early stages of memory retrieval. Interestingly, in the spatial memory task, where this timing issue is less problematic due to the nature of the task, we observe that the PCC/precuneus shows an earlier onset of activity compared to the AI. This process is aligned with the view that DMN nodes may trigger internal recall processes, the full extent and replication of which across verbal and spatial tasks could not be examined in this study.  

      We have added a discussion of these limitations and future directions to the manuscript to provide a more nuanced interpretation of our findings and to highlight important areas for further investigation (page 19). 

      (2.3) Second, the authors state that high-gamma suppression in the PCC/precuneus relative to the AI is an anatomically specific signature that is not present in the FPN. This claim does not seem to be supported by their own evidence as presented in the Supplemental Data (Figures S2 and S3), which to my eye show clear evidence of relative suppression in the MFG and dPPC (e.g. S2a and S3a, most notably) which are notated as "significant" with green bars. I appreciate that the magnitude of this effect may be greater in the PCC/precuneus, but if this is the claim it should be supported by appropriate statistics and interpretation.  

      We thank the reviewer for raising this point. We have now directly compared the high-gamma power of the PCC/precuneus with the dPPC and MFG nodes of the FPN and we note that the suppression effects of the PCC/precuneus are stronger compared to those of the dPPC and MFG during memory encoding (Figures S8, S9). 

      (2.4) I commend the authors on emphasizing replicability, but I found their Bayes Factor (BF) analysis to be difficult to interpret and qualitatively inconsistent with the results that they show. For example, the authors state that BF analysis demonstrates "high replicability" of the gamma suppression effect in Figure 3a with that of 3c and 3d. While it does appear that significant effects exist across all three tasks, the temporal structure of high gamma signals appears markedly different between the two in ways that may be biologically meaningful. Moreover, it appears that the BF analysis did not support replicability between VFR and CATVFR, which is very surprising; these are essentially the same tasks (merely differing in the presence of word categories) and would be expected to have the highest degree of concordance, not the lowest. I would suggest the authors try to analytically or conceptually reconcile this surprising finding. 

      We appreciate the reviewer's commendation on our emphasis on replicability and thank the reviewer for the opportunity to provide clarification.

      First, we would like to clarify the nature of our BF analysis. Bayes factors are calculated as the ratio of the marginal likelihood of the replication data, given the posterior distribution estimated from the original data, and the marginal likelihood for the replication data under the null hypothesis of no effect (Ly, Etz, Marsman, & Wagenmakers, 2019). Specifically, BFs use the posterior distribution from the first experiment as a prior distribution for the replication test of the second experiment to constitute a joint multivariate distribution (i.e., the additional evidence for the alternative hypothesis given what was already observed in the original study) and this joint distribution is dependent on the similarity between the two experiments (Ly et al., 2019).  This analysis revealed that PCC/precuneus suppression, in comparison to the AI during memory encoding, observed in the VFR during memory encoding was detected in two other tasks, PALVCR, and WMSM with high BFs. In the CATVFR task, although there were short time periods of PCC/precuneus suppression (Figure 3), the effects were not strong enough like the three other tasks.  

      Regarding the high-gamma suppression effect, our BF analysis indeed supports replicability across the VFR, PALVCR, and WMSM tasks. While we agree with the reviewer that the temporal structure of high-gamma signals appears different across tasks, the BF analysis focuses on the overall presence of the suppression effect rather than its precise temporal profile. The high BFs indicate that the core finding - PCC/precuneus suppression relative to the AI during memory encoding - is replicated across these tasks, despite differences in the timing of this suppression. Moreover, at no time point did responses in the PCC/precuneus exceed that of the AI in any of the four memory encoding tasks. 

      The reason for differences in temporal profiles is not clear. While VFR and CATVFR are similar, the addition of categorical structure in CATVFR may have introduced cognitive processes that alter the temporal dynamics of regional responses. Moreover, differences in electrode placements across participants in each experiment may also have contributed to variability in the observed effects. Further studies using within-subjects experimental designs are needed to address this. 

      We have updated our Results and Discussion sections to reflect these points and to provide a more nuanced interpretation of the replicability across tasks.  

      (2.5) To aid in interpretability, it would be extremely helpful for the authors to assess acrosstask similarity in high-gamma power on a within-subject basis, which they are wellpowered to do. For example, could they report the correlation coefficient between HGP timecourses in paired-associates versus free-recall tasks, to better establish whether these effects are consistent on a within-subject basis? This idea could similarly be extended to the PTE analysis. Across-subject correlations would also be a welcome analysis that may provide readers with better-contextualized effect sizes than the output of a Bayes Factor analysis.  

      We thank the reviewer for this suggestion. However, a within-subject analysis was not possible because very few participants participated in multiple memory tasks. 

      For example, for the AI-PCC/Pr analysis, only 1 individual participated in both the VFR and PALVCR tasks (Tables S2a, S2c). Similarly, for AI-mPFC analysis, only 3 subjects participated in both the VFR and PALVCR tasks (Tables S2a, S2c).  

      Due to these small sample sizes, it was not feasible for us to assess replicability across tasks on a within-subject basis in our dataset. Therefore, for all our analysis, we have pooled electrode pairs across subjects and then subjected these to a linear mixed effects modeling framework for assessing significance and then subsequently assessing replicability of these effects using the Bayes factor (BF) framework.    

      Recommendations For The Authors: 

      (2.6) I would emphasize manuscript organization in a potential rewrite; it was very difficult to follow which analyses were attempting to show a contrast between effects versus a similarity between effects. Results were grouped by the underlying experimental conditions (e.g. encoding/recall, network identity, etc.) but may be better grouped according to the actual effects that were found. 

      We thank the reviewer for this suggestion. We considered this possibility, but we feel that the Results section is best organized in the order of the hypotheses we set out to test, starting from analyzing local brain activity using high-gamma power analysis, and then results related to analyzing brain connectivity using PTE. All these results are systematically ordered by presenting results related to encoding first and then the recall periods as they appear sequentially in our task-design, presenting the results related to the VFR task first and then demonstrating replicability of the results in the three other experiments. Results are furthermore arranged by nodes, where we first discuss results related to the DMN nodes, and then the same for the FPN nodes. This is to ensure systematic, unbiased organization of all our results for the readers to clearly follow the hypotheses, statistical analyses, and the brain regions considered. Therefore, for transparency and ethical reasons, we would respectfully like to present our results as they appear in our current manuscript, rather than presenting the results based on effect sizes. 

      However, please note that we indeed have ordered our results in the Discussion section based on actual effects, as suggested by the reviewer.  

      (2.7) The absence of a PTE effect when analyzing through the lens of successful vs. unsuccessful memory is an important limitation of the current study and a significant departure from the wealth of subsequent memory effects reported in the literature (which the authors have already done a good job citing, e.g. Burke et al. 2014 Neuroimage). I'm glad that the authors raised this in their Discussion, but it is important that the results of such an analysis actually be shown in the manuscript. 

      We thank the reviewer for this suggestion. We have now included the results related to PTE dynamics for successful vs. unsuccessful memory trials in the revised Results section as we note on page 12: 

      “Differential information flow from the AI to the DMN and FPN for successfully recalled and forgotten memory trials 

      We examined memory effects by comparing PTE between successfully recalled and forgotten memory trials. However, this analysis did not reveal differences in directed influence from the AI on the DMN and FPN or the reverse, between successfully recalled and forgotten memory trials during the encoding as well as recall periods in any of the memory experiments (all ps>0.05).”

      (2.8) I believe the claims of causality through the use of the PTE are overstated throughout the manuscript and may contribute to further confusion in the literature regarding how causality in the brain can actually be understood. See Mehler and Kording, 2018 arXiv for an excellent discussion on the topic (https://arxiv.org/abs/1812.03363). My recommendation would be to significantly tone down claims that PTE reflects causal interactions in the brain. 

      We thank the reviewer for this suggestion. We would like to clarify here that we define causality in our manuscript as follows: a brain region has a causal influence on a target if knowing the past history of temporal signals in both regions improves the ability to predict the target's signal in comparison to knowing only the target's past, as defined in earlier studies (Granger, 1969; Lobier et al., 2014). We have now included this clarification in the Introduction section (page 6).  

      We also agree with the reviewer that to more mechanistically establish a causal link between the neural dynamics and behavior, lesion or brain stimulation studies are necessary. We have now acknowledged this in the revised Discussion as we note: “Although our computational methods suggest causal influences, direct causal manipulations, such as targeted brain stimulation during memory tasks, are needed to establish definitive causal relationships between network nodes.” (page 19). 

      Finally, we have now significantly toned down our claims that PTE reflects causal interactions in the brain, in the Introduction, Results, and Discussion sections of our revised manuscript.  

      (2.9) Relatedly, it may be useful for the authors to consider a supplemental analysis that uses classic measures of inter-regional synchronization, e.g. the PLV, and compare to their PTE findings. They cite literature to suggest a metric like the PTE may be useful, but this hardly rules out the potential utility of investigating narrowband phase synchronization. 

      We thank the reviewer for this suggestion. We have now run new analyses based on PLV to examine phase synchronization between the AI and the DMN and FPN. However, we did not find a significant PLV for the interactions between the AI and DMN and FPN nodes for the different task periods compared to the resting baselines, as we note on page 13: 

      “Narrowband phase synchronization between the AI and the DMN and FPN during encoding and recall compared to resting baseline  

      We next directly compared the phase locking values (PLVs) (see Methods for details) between the AI and the PCC/precuneus and mPFC nodes of the DMN and also the dPPC and MFG nodes of the FPN for the encoding and the recall periods compared to resting baseline. However, narrowband PLV values did not significantly differ between the encoding/recall vs. rest periods, in any of the delta-theta (0.5-8 Hz), alpha (8-12 Hz), beta (12-30 Hz), gamma (30-80 Hz), and high-gamma (80-160 Hz) frequency bands. These results indicate that PTE, rather than phase synchronization, more robustly captures the AI dynamic interactions with the DMN and the FPN.” 

      Please note that phase locking measures such as the PLV or coherence do not probe directed causal influences and cannot address how one region drives another. Instead, our study examined the direction of information flow between the AI and the DMN and FPN using robust estimators of the direction of information flow. PTE assesses with the ability of one time-series to predict future values of other time-series, thus estimating the time-delayed causal influences between the two time-series, whereas PLV or coherence can only estimate “instantaneous” phase synchronization, but not predict the future time-series. 

      Additionally, please note that the directed information flow from the AI to both the DMN and FPN were enhanced during the encoding and recall periods compared to resting state across all four experiments, in a new set of analyses that we have carried out in our revised manuscript. Specifically, we have now carried out our task versus rest comparison by using resting baseline epochs before the start of the entire session of the task periods, rather than our previously used rest epochs which were in between the task periods. These new results have now been included in the revised Results as we note on page 12:  

      “Enhanced information flow from the AI to the DMN and FPN during episodic memory processing, compared to resting-state baseline

      We next examined whether directed information flow from the AI to the DMN and FPN nodes during the memory tasks differed from the resting-state baseline. Resting-state baselines were extracted immediately before the start of the task sessions and the duration of task and rest epochs were matched to ensure that differences in network dynamics could not be explained by differences in duration of the epochs. Directed information flow from the AI to both the DMN and FPN were higher during both the memory encoding and recall phases and across the four experiments, compared to baseline in all but two cases (Figures S6, S7). These findings provide strong evidence for enhanced role of AI directed information flow to the DMN and FPN during memory processing compared to the resting state.”  

      References 

      Backus, A. R., Schoffelen, J. M., Szebényi, S., Hanslmayr, S., & Doeller, C. F. (2016). Hippocampal-Prefrontal Theta Oscillations Support Memory Integration. Curr Biol, 26(4), 450-457. doi:10.1016/j.cub.2015.12.048

      Bastos, A. M., Vezoli, J., Bosman, C. A., Schoffelen, J. M., Oostenveld, R., Dowdall, J. R., . . . Fries, P. (2015). Visual areas exert feedforward and feedback influences through distinct frequency channels. Neuron, 85(2), 390-401. doi:10.1016/j.neuron.2014.12.018

      Canolty, R. T., Edwards, E., Dalal, S. S., Soltani, M., Nagarajan, S. S., Kirsch, H. E., . . . Knight, R. T. (2006). High gamma power is phase-locked to theta oscillations in human neocortex. Science, 313(5793), 1626-1628. doi:10.1126/science.1128115

      Canolty, R. T., & Knight, R. T. (2010). The functional role of cross-frequency coupling. Trends Cogn Sci, 14(11), 506-515. doi:10.1016/j.tics.2010.09.001

      Chen, L., Wassermann, D., Abrams, D. A., Kochalka, J., Gallardo-Diez, G., & Menon, V. (2019). The visual word form area (VWFA) is part of both language and attention circuitry. Nat Commun, 10(1), 5601. doi:10.1038/s41467-019-13634-z

      Clouter, A., Shapiro, K. L., & Hanslmayr, S. (2017). Theta Phase Synchronization Is the Glue that Binds Human Associative Memory. Curr Biol, 27(20), 3143-3148.e3146. doi:10.1016/j.cub.2017.09.001

      Crone, N. E., Boatman, D., Gordon, B., & Hao, L. (2001). Induced electrocorticographic gamma activity during auditory perception. Brazier Award-winning article, 2001. Clin Neurophysiol, 112(4), 565-582. doi:10.1016/s1388-2457(00)00545-9

      Daitch, A. L., & Parvizi, J. (2018). Spatial and temporal heterogeneity of neural responses in human posteromedial cortex. Proc Natl Acad Sci U S A, 115(18), 4785-4790. doi:10.1073/pnas.1721714115

      Das, A., de Los Angeles, C., & Menon, V. (2022). Electrophysiological foundations of the human default-mode network revealed by intracranial-EEG recordings during restingstate and cognition. Neuroimage, 250, 118927. doi:10.1016/j.neuroimage.2022.118927

      Das, A., & Menon, V. (2020). Spatiotemporal Integrity and Spontaneous Nonlinear Dynamic Properties of the Salience Network Revealed by Human Intracranial Electrophysiology: A Multicohort Replication. Cereb Cortex, 30(10), 5309-5321. doi:10.1093/cercor/bhaa111

      Das, A., & Menon, V. (2021). Asymmetric Frequency-Specific Feedforward and Feedback Information Flow between Hippocampus and Prefrontal Cortex during Verbal Memory Encoding and Recall. J Neurosci, 41(40), 8427-8440. doi:10.1523/jneurosci.080221.2021

      Das, A., & Menon, V. (2022). Replicable patterns of causal information flow between hippocampus and prefrontal cortex during spatial navigation and spatial-verbal memory formation. Cereb Cortex, 32(23), 5343-5361. doi:10.1093/cercor/bhac018

      Das, A., & Menon, V. (2023). Concurrent- and after-effects of medial temporal lobe stimulation on directed information flow to and from prefrontal and parietal cortices during memory formation. J Neurosci, 43(17), 3159-3175. doi:10.1523/jneurosci.1728-22.2023

      Edwards, E., Soltani, M., Deouell, L. Y., Berger, M. S., & Knight, R. T. (2005). High gamma activity in response to deviant auditory stimuli recorded directly from human cortex. J Neurophysiol, 94(6), 4269-4280. doi:10.1152/jn.00324.2005

      Ekstrom, A. D., Caplan, J. B., Ho, E., Shattuck, K., Fried, I., & Kahana, M. J. (2005). Human hippocampal theta activity during virtual navigation. Hippocampus, 15(7), 881-889. doi:10.1002/hipo.20109

      Ekstrom, A. D., & Watrous, A. J. (2014). Multifaceted roles for low-frequency oscillations in bottom-up and top-down processing during navigation and memory. Neuroimage, 85 Pt 2, 667-677. doi:10.1016/j.neuroimage.2013.06.049

      Engel, A. K., & Fries, P. (2010). Beta-band oscillations--signalling the status quo? Curr Opin Neurobiol, 20(2), 156-165. doi:10.1016/j.conb.2010.02.015

      Foster, B. L., Rangarajan, V., Shirer, W. R., & Parvizi, J. (2015). Intrinsic and task-dependent coupling of neuronal population activity in human parietal cortex. Neuron, 86(2), 578590. doi:10.1016/j.neuron.2015.03.018

      Gonzalez, A., Hutchinson, J. B., Uncapher, M. R., Chen, J., LaRocque, K. F., Foster, B. L., . . . Wagner, A. D. (2015). Electrocorticography reveals the temporal dynamics of posterior parietal cortical activity during recognition memory decisions. Proc Natl Acad Sci U S A, 112(35), 11066-11071. doi:10.1073/pnas.1510749112

      Granger, C. W. J. (1969). Investigating Causal Relations by Econometric Models and Crossspectral Methods. Econometrica, 37(3), 424-438. doi:10.2307/1912791

      Griffiths, B. J., Martín-Buro, M. C., Staresina, B. P., & Hanslmayr, S. (2021). Disentangling neocortical alpha/beta and hippocampal theta/gamma oscillations in human episodic memory formation. Neuroimage, 242, 118454. doi:10.1016/j.neuroimage.2021.118454

      Guderian, S., & Düzel, E. (2005). Induced theta oscillations mediate large-scale synchrony with mediotemporal areas during recollection in humans. Hippocampus, 15(7), 901-912. doi:10.1002/hipo.20125

      Guderian, S., Schott, B. H., Richardson-Klavehn, A., & Düzel, E. (2009). Medial temporal theta state before an event predicts episodic encoding success in humans. Proc Natl Acad Sci U S A, 106(13), 5365-5370. doi:10.1073/pnas.0900289106

      Helfrich, R. F., & Knight, R. T. (2016). Oscillatory Dynamics of Prefrontal Cognitive Control. Trends Cogn Sci, 20(12), 916-930. doi:10.1016/j.tics.2016.09.007

      Hillebrand, A., Tewarie, P., van Dellen, E., Yu, M., Carbo, E. W., Douw, L., . . . Stam, C. J. (2016). Direction of information flow in large-scale resting-state networks is frequencydependent. Proc Natl Acad Sci U S A, 113(14), 3867-3872. doi:10.1073/pnas.1515657113

      Jeffreys, H. (1998). The Theory of Probability (3rd ed.). Oxford, England: Oxford University Press.

      Kwon, H., Kronemer, S. I., Christison-Lagay, K. L., Khalaf, A., Li, J., Ding, J. Z., . . . Blumenfeld, H. (2021). Early cortical signals in visual stimulus detection. Neuroimage, 244, 118608. doi:10.1016/j.neuroimage.2021.118608

      Lachaux, J. P., George, N., Tallon-Baudry, C., Martinerie, J., Hugueville, L., Minotti, L., . . . Renault, B. (2005). The many faces of the gamma band response to complex visual stimuli. Neuroimage, 25(2), 491-501. doi:10.1016/j.neuroimage.2004.11.052

      Lobier, M., Siebenhühner, F., Palva, S., & Matias, P. J. (2014). Phase transfer entropy: A novel phase-based measure for directed connectivity in networks coupled by oscillatory interactions. Neuroimage, 85, 853-872. doi:10.1016/j.neuroimage.2013.08.056

      Ly, A., Etz, A., Marsman, M., & Wagenmakers, E. J. (2019). Replication Bayes factors from evidence updating. Behav Res Methods, 51(6), 2498-2508. doi:10.3758/s13428-0181092-x

      Mainy, N., Kahane, P., Minotti, L., Hoffmann, D., Bertrand, O., & Lachaux, J. P. (2007). Neural correlates of consolidation in working memory. Hum Brain Mapp, 28(3), 183-193. doi:10.1002/hbm.20264

      Miller, K. J., Leuthardt, E. C., Schalk, G., Rao, R. P., Anderson, N. R., Moran, D. W., . . . Ojemann, J. G. (2007). Spectral changes in cortical surface potentials during motor movement. J Neurosci, 27(9), 2424-2432. doi:10.1523/jneurosci.3886-06.2007

      Miller, K. J., Weaver, K. E., & Ojemann, J. G. (2009). Direct electrophysiological measurement of human default network areas. Proceedings of the National Academy of Sciences, 106(29), 12174-12177. doi:10.1073/pnas.0902071106

      Nuzzo, R. L. (2017). An Introduction to Bayesian Data Analysis for Correlations. Pm r, 9(12), 1278-1282. doi:10.1016/j.pmrj.2017.11.003

      Ray, S., Crone, N. E., Niebur, E., Franaszczuk, P. J., & Hsiao, S. S. (2008). Neural correlates of high-gamma oscillations (60-200 Hz) in macaque local field potentials and their potential implications in electrocorticography. J Neurosci, 28(45), 11526-11536. doi:10.1523/jneurosci.2848-08.2008

      Sederberg, P. B., Schulze-Bonhage, A., Madsen, J. R., Bromfield, E. B., Litt, B., Brandt, A., & Kahana, M. J. (2007). Gamma oscillations distinguish true from false memories. Psychol Sci, 18(11), 927-932. doi:10.1111/j.1467-9280.2007.02003.x

      Tallon-Baudry, C., Bertrand, O., Hénaff, M. A., Isnard, J., & Fischer, C. (2005). Attention modulates gamma-band oscillations differently in the human lateral occipital cortex and fusiform gyrus. Cereb Cortex, 15(5), 654-662. doi:10.1093/cercor/bhh167

      Watrous, A. J., Tandon, N., Conner, C. R., Pieters, T., & Ekstrom, A. D. (2013). Frequencyspecific network connectivity increases underlie accurate spatiotemporal memory retrieval. Nature Neuroscience, 16(3), 349-356. doi:10.1038/nn.3315

      Wetzels, R., & Wagenmakers, E. J. (2012). A default Bayesian hypothesis test for correlations and partial correlations. Psychon Bull Rev, 19(6), 1057-1064. doi:10.3758/s13423-0120295-x

    1. In this decayed hole among the mountains In the faint moonlight, the grass is singing Over the tumbled graves, about the chapel There is the empty chapel, only the wind’s home. It has no windows, and the door swings, Dry bones can harm no one. Only a cock stood on the rooftree Co co rico co co rico In a flash of lightning. Then a damp gust Bringing rain

      Decayed hole=broken city. Empty chapel=decrease in people's faith as their hope as also faded away.

    1. Reviewer #2 (Public review):

      Summary:

      In this paper, the authors investigate the role of NMDA-receptors in recurrent processing. In doing so, the authors present data from two studies, where they attempt to decode different stimulus features, namely contrast, collinearity, and illusory contours. The latter of which the authors claim relies uniquely on recurrent processing. Therefore, to test whether NMDA receptors are particularly involved in recurrent processing they administer a NMDA-antagonist to see whether the decoding of illusory contours is specifically perturbed, and leaves the decoding of other features intact. They further aim to disentangle the role of NMDA-receptors by manipulating visibility and task relevance of the decoded features

      In the first experiment, the authors decode two targets, the first was always presented clearly, the second's visibility was manipulated by presenting it after a short lag rather than a long lag (inducing attentional blink), as well as masking the target on half the trials. First, they find for target 1 clear evidence for the NMDA-receptor increasing (rather than decreasing) decoding performance of illusory contours. They move on to analyse target 2 to explore the manipulations of lag and masking. Here they find that masking reduced decoding of all three stimulus features, but only the lag reduced decoding of illusory contours. Importantly, the NMDA-antagonist improved decoding only in the unmasked, long lag condition, in the cluster analyses. However, the interaction with the lag condition was not significant, and the effect on decoding was primarily present in the later decoding time window, and not significant when exploring the peak of the decoding time window.

      The second experiment was highly similar, but got rid of the lag manipulation, and replaced it with a manipulation of task relevance. Notably, masking did not abolish the decoding of illusory contours completely, in contrast to the first experiment. More importantly, they find that the NMDA-receptor now clearly increases decoding of illusory contours, particularly when the illusory contours are not masked. No effect of task relevance is found.

      Taken together the authors state that evidence is found for NMDA-receptors role in recurrent processing.

      Strengths:

      This is an interesting study using state-of-the-art methods in combination with drug manipulation to study recurrent processing. Their analysis methods are state-of-the-art, and the question that they are trying to address is topical and interesting to a wide research audience, encompassing both researchers interested in visual perception and consciousness, as well as those interested in perturbed vision as found in psychiatric disorders.

      Weaknesses:

      The experimental design is somewhat complicated, which can make it difficult to match the authors' claims to the actual evidence that is provided. I have some reservations about the paper which are born out of a few issues.<br /> (1) The title, abstract, and introduction hide their counterintuitive finding of increased decoding, presumably as it was unexpected.<br /> (2) Their analysis choices are sometimes unclear, making it difficult to assess whether the analyses are sensible.<br /> (3) The appropriate tests for the interactions that the authors claim they found are often lacking.

      To start off, I think the reader is being a bit tricked when reading the paper. Perhaps my priors are too strong, but I assumed, just like the authors, that NMDA-receptors would disrupt recurrent processing, in line with previous work. However, due to the continuous use of the ambiguous word 'affected' rather than the more clear increased or perturbed recurrent processing, the reader is left guessing what is actually found. That's until they read the results and discussion finding that decoding is actually improved. This seems like a really big deal, and I strongly urge the authors to reword their title, abstract, and introduction to make clear they hypothesized a disruption in decoding in the illusion condition, but found the opposite, namely an increase in decoding. I want to encourage the authors that this is still a fascinating finding.

      Apologies if I have missed it, but it is not clear to me whether participants were given the drug or placebo during the localiser task. If they are given the drug this makes me question the logic of their analysis approach. How can one study the presence of a process, if their very means of detecting that process (the localiser) was disrupted in the first place? If participants were not given a drug during the localiser task, please make that clear. I'll proceed with the rest of my comments assuming the latter is the case. But if the former, please note that I am not sure how to interpret their findings in this paper.

      The main purpose of the paper is to study recurrent processing. The extent to which this study achieves this aim is completely dependent to what extent we can interpret decoding of illusory contours as uniquely capturing recurrent processing. While I am sure illusory contours rely on recurrent processing, it does not follow that decoding of illusory contours capture recurrent processing alone. Indeed, if the drug selectively manipulates recurrent processing, it's not obvious to me why the authors find the interaction with masking in experiment 2. Recurrent processing seems to still be happening in the masked condition, but is not affected by the NMDA-receptor here, so where does that leave us in interpreting the role of NMDA-receptors in recurrent processing? If the authors can not strengthen the claim that the effects are completely driven by affecting recurrent processing, I suggest that the paper will shift its focus to making claims about the encoding of illusory contours, rather than making primary claims about recurrent processing.

      An additional claim is being made with regards to the effects of the drug manipulation. The authors state that this effect is only present when the stimulus is 1) consciously accessed, and 2) attended. The evidence for claim 1 is not supported by experiment 1, as the masking manipulation did not interact in the cluster-analyses, and the analyses focussing on the peak of the timing window do not show a significant effect either. There is evidence for this claim coming from experiment 2 as masking interacts with the drug condition. Evidence for the second claim (about task relevance) is not presented, as there is no interaction with the task condition. A classical error seems to be made here, where interactions are not properly tested. Instead, the presence of a significant effect in one condition but not the other is taken as sufficient evidence for an interaction, which is not appropriate. I therefore urge the authors to dampen the claim about the importance of attending to the decoded features. Alternatively, I suggest the authors run their interactions of interest on the time-courses and conduct the appropriate cluster-based analyses.

      How were the length of the peak-timing windows established in Figure 1E? My understanding is that this forms the training-time window for the further decoding analyses, so it is important to justify why they have different lengths, and how they are determined. The same goes for the peak AUC time windows for the interaction analyses. A number of claims in the paper rely on the interactions found in these post-hoc analyses, so the 223- to 323 time window needs justification.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Review):

      Summary:

      Deletion of the hrp2 and hrp3 loci in P. falciparum poses an immediate public health threat. This manuscript provides a more complete understanding of the dynamic nature with which these deletions are generated. By delving into the likely mechanisms behind their generation, the authors also provide interesting insight into general Plasmodium biology that can inform our broader understanding of the parasite's genomic evolution.

      Strengths:

      The sub-telomeric regions of P. falciparum (where hrp2 and hrp3 are located) are notoriously difficult to study with short-read sequence data. The authors take an appropriate, targeted approach toward studying the loci of interest, which includes read-depth analysis and local haplotype reconstruction. They additionally use both long-read and short-read data to validate their major findings. There is an extensive set of supplementary plots, which helps clarify several aspects of the data.

      Weaknesses:

      In this first version, there are a few factors that hinder a full assessment of the robustness and replicability of the results.

      Reviewer #1 (Recommendations For The Authors):

      Reviewer comment: First, a number of the analyses lack basic details in the methods; for instance, one must visit the authors' personal website to find some of the tools used.

      We have extensively updated the methods to clarify which tools were used and how they were run. All code and results for the analyses have been deposited in Zenodo at https://doi.org/10.5281/zenodo.12167687.

      Reviewer comment: Second, there are several tricky methodological points that are not fully documented. Read depths are treated (and plotted) discretely as 0/1/2 without any discussion of how thresholds were used and determined.

      We have added to the methods section the full details on how read depth was handled, including rounding to the closest 1 normalized coverage for visualizations. To ensure analysis of only highly confident deleted strains, normalized coverage of 0.1 or more was round to 1 instead of 0. Samples were considered for potential genomic deletion if they had zero coverage after rounding from chromosome 8 1,375,557 to 1,387,982 for pfhrp2, chromosome 13 from 2,841,776 to 2,844,785 for pfhrp3, and from chromosome 11 1,991,347 to 2,003,328. These numbers were chosen after visual inspection of samples with any zero coverage within the genomic region of pfhrp2/3.

      Reviewer comment: For read mapping to standard vs hybrid chromosomes, there is no documentation on how assignments were made if partially ambiguous or how final sample calls were determined when some reads were discordant. There is no mention of how missing data were handled. Without this, it is difficult to know when conclusions were based on analyses that were more quantitative (for instance, using pre-determined read thresholds) or more subjective (with patterns being extracted visually).

      We have updated several parts of the methods section to explicitly state what thresholds and analysis pipelines to use, making our documentation clearer. For mapping to the hybrid vs standard chromosomes for the long reads, spanning reads across the duplicated region were required to extend 50bp upstream and downstream of the region. These regions are significantly different between chromosomes 11 and 13, so requiring spanning reads to map to these regions prevented multi-mapping reads. Reads that started within the duplicated region were allowed to map to both the hybrid and standard chromosomes for visualization in Figure 4. Importantly, for both HB3 and SD01, no reads spanned from the duplicated region into chromosome 13, showing a complete lack of reads that contained the portion of chromosome 13 that came after the duplicated region. None of the other isolates had any spanning reads across the hybrid chromosomes. Details on deletion calls were based on initial visualization of pfhrp2/3 and then on read thresholds (see above response for details).

      Reviewer comment: Third, while a new method is employed for local haplotype reconstruction (PathWeaver), the manuscript does not include details on this approach or benchmarking data with which to evaluate its performance and understand any potential artifacts.

      We have added an analysis based on biallelic SNPs to compare to the PathWeaver results, which produced similar results to help validate the PathWeaver results. PathWeaver manuscript is in preparation.

      Reviewer #2 (Public Review):

      This work investigates the mechanisms, patterns, and geographical distribution of pfhrp2 and pfhrp3 deletions in Plasmodium falciparum. Rapid diagnostic tests (RDTs) detect P. falciparum histidine-rich protein 2 (PfHRP2) and its paralog PfHRP3 located in subtelomeric regions. However, laboratory and field isolates with deletions of pfhrp2 and pfhrp3 that can escape diagnosis by RDTs are spreading in some regions of Africa. They find that pfhrp2 deletions are less common and likely occur through chromosomal breakage with subsequent telomeric healing. Pfhrp3 deletions are more common and show three distinct patterns: loss of chromosome 13 from pfhrp3 to the telomere with evidence of telomere healing at breakpoint (Asia; Pattern 13-); duplication of a chromosome 5 segment containing pfhrp1 on chromosome 13 through non-allelic homologous recombination (NAHR) (Asia; Pattern 13-5++); and the most common pattern, duplication of a chromosome 11 segment on chromosome 13 through NAHR (Americas/Africa; Pattern 13-11++). The loss of these genes impacts the sensitivity of RDTs, and knowing these patterns and geographic distribution makes it possible to make better decisions for malaria control.

      Reviewer #3 (Public Review):

      Summary:

      The study provides a detailed analysis of the chromosomal rearrangements related to the deletions of histidine-rich protein 2 (pfhrp2) and pfhrp3 genes in P. falciparum that have clinical significance since malaria rapid diagnostic tests detect these parasite proteins. A large number of publicly available short sequence reads for the whole genome of the parasite were analyzed, and data on coverage and discordant mapping allowed the authors to identify deletions, duplications, and chromosomal rearrangements related to pfhrp3 deletions. Long-read sequences showed support for the presence of a normal chromosome 11 and a hybrid 13-11 chromosome lacking pfhrp3 in some of the pfhrp3-deleted parasites. The findings support that these translocations have repeatedly occurred in natural populations. The authors discuss the implications of these findings and how they do or do not support previous hypotheses on the emergence of these deletions and the possible selective pressures involved.

      Strengths:

      The genomic regions where these genes are located are challenging to study since they are highly repetitive and paralogous and the use of long-read sequencing allowed to span the duplicated regions, giving support to the identification of the hybrid 13-11 chromosome.

      All publicly available whole-genome sequences of the malaria parasite from around the world were analysed which allowed an overview of the worldwide variability, even though this analysis is biased by the availability of sequences, as the authors recognize.

      Despite the reduced sample size, the detailed analysis of haplotypes and identification of the location of breakpoints gives support to a single origin event for the 13-5++ parasites.

      The analysis of haplotype variation across the duplicated chromosome-11 segment identified breakpoints at varied locations that support multiple translocation events in natural populations. The authors suggest these translocations may be occurring at high frequency in meiosis in natural populations but are strongly selected against in most circumstances, which remains to be tested.

      Weaknesses:

      Reviewer comment: Relying on sequence data publicly available, that were collected based on diagnostic test positivity and that are limited by sequencing availability, limits the interpretation of the occurrence and relative frequency of the deletions.

      However, we have uncovered more mechanisms than previously detected for hrp2 (involving MDR1) in SEA and South American parasites are likely detected by microscopy as RDTs were never introduced due to the presence of the deletions.

      Reviewer comment: In the discussion, caution is needed when identifying the least common and most common mechanisms and their geographical associations. The identification of only one type of deletion pattern for Pfhrp2 may be related to these biases.

      We added a section in the Discussion on the limitations of our study, which states the following, “Limitations of this study include the use of publicly available sequencing data that were collected often based on positive rapid diagnostic tests, which limits our interpretation of the occurrence and relative frequency of these deletions. This could introduce regional biases due to different diagnostic methods as well as limit the full range of deletion mechanisms, particularly pfhrp2.”

      Reviewer comment: The specific objectives of the study are not stated clearly, and it is sometimes difficult to know which findings are new to this study. Is it the first study analyzing all the worldwide available sequences? Is it the first one to do long-read sequencing to span the entire duplicated region?

      In the Introduction, we added, “The objectives of this study were to determine the pfhrp3 deletion patterns along with their geographical associations and sequence and assemble the chromosomes containing the deletions using long-read sequencing.”

      We also added in the Discussion, “To the best of our knowledge, no prior studies have performed long-read sequencing to definitively span and assemble the entire segmental duplication involved in the deletions.”

      Reviewer comment: Another aspect that should be explained in the introduction is that there was previous information about the association of the deletions to patterns found in chromosomes 5 and 11. In the short-read sequences results, it is not clear if these chromosomes were analysed because of the associations found in this study (and no associations were found to putative duplications or deletions in other chromosomes), or if they were specifically included in the analysis because of the previous information (and the other chromosomes were not analysed).

      The former is correct. Chromosomes 5 and 11 were analyzed due to the associations found in this study, not from prior information. We have added the following sentence in the Results: “As a result of our short-read analysis demonstrating these three patterns and discordant reads between the chromosomes involved, chromosomes 5, 11, and 13 were further examined. No other chromosomes had associated discordant reads or changes in read coverage. ”

      Reviewer comment: An interesting statement in the discussion is that existing pfhrp3 deletions in a low-transmission environment may provide a genetic background on which less frequent pfhrp2 deletion events can occur. Does it mean that the occurrence of pfhrp3 deletions would favor the pfhrp2 deletion events? How, and is there any evidence for that?

      We should have stated more explicitly that selection would better be able to act on the now doubly deleted parasite versus a parasite with HRP3 still intact and weakly detectable by RDTs.Since fully RDT-negative parasites require a two-hit mechanism, where both pfhrp2 and pfhrp3 need to be deleted, and since there appear to be more mechanisms and drivers for pfhrp3 deletions, this would create a population of parasites with one hit already and would only require the additional hit of pfhrp2 deletion to occur to become RDT negative. So the point in the discussion being made is not that the pfhrp3 deletion would favor pfhrp2 deletion but rather that there is a population circulating with one hit already, which would make it more likely that the less frequent pfhrp2 deletion would result in a dual deleted parasite and therefore an RDT-negative parasite. The discussion has been modified to the following to try to make this point more clear. “In the setting of RDT use in a low-transmission environment, a pfhrp2 deletion occurring in the context of an existing pfhrp3 deletion may be more strongly selected for compared to pfhrp2 deletion occurring alone still detectable by RDTs.”

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      Reviewer comment: In the text, clonal propagation is the proposed hypothesis for the presence of near-identical copies of the chromosome 11 duplicated region. Even among the parasites showing variation between chromosomes, Figure 5 shows 3 haplotype groups with multiple sample members, which is also suggestive that these are highly related parasites. In addition to confirming COI status, it would be straightforward to calculate the genome-wide relatedness between/among parasites belonging to the same haplotype group. The assumption is that they are clones or highly related. A different finding would require more thought into potential genomic artifacts driving the pattern.

      Thank you for this helpful suggestion. We confirmed the COI of each sample using THE REAL McCOIL. Six samples were not monoclonal, and we removed these samples from the downstream analysis to remove any contribution of polyclonal samples to the downstream haplotype analysis. Then, by using hmmIBD on whole-genome biallelic SNPs, we determined the whole-genome relatedness between the parasites. The haplotype groups do appear clonal though there appear to be several clonal groups within the larger groups of clusters 01 (n=28) and 03 (n=12) which combined with the variation seen within the 15.2kb region on chromosome 11/13, there appears to be different events that then lead to the same duplicated chromosome 11.

      Reviewer comment: By way of validating the PathWeaver results, it could be useful to use another comparator method on the samples that are COI=1 or 2.

      We have added an analysis based on biallelic SNPs to compare to the PathWeaver results, which produced similar results to help validate the PathWeaver results. We continued to use PathWeaver (Hathaway, in preparation), which is better able to detect variation relative to standard GATK4 analyses due to the refined local alignments from assembled haplotypes.

      Questions regarding Methods:

      Reviewer comment: Were any metrics of genome quality factored into sample selection?

      Yes, samples were removed if there was less than <5x median whole genome coverage. Additionally, several subsets of sWGA samples were removed based on visual inspection. These details have been added to the methods section.

      Reviewer comment: How were polyclonal samples treated to ensure they did not produce analysis artifacts?

      The read-depth analysis required zero coverage across the regions of pfhrp2/pfhrp3, which made it so that most of the samples analyzed were monoclonal (or polyclonal infections of only deleted strains). We have now used THE REAL McCOIL on whole genome SNPs to determine COIs. Six samples were identified as polyclonal, and we removed them for the analysis and updated the manuscript. Their removal did not significantly impact the results or conclusions.

      Reviewer comment: How was local realignment of short-read data performed? Was this step informed by the conserved, non-paralogous genomic regions, or were these only used for downstream variant analysis?

      No local realignment of short-read data was performed. The analysis was either read depth or de novo assembly from reads from specific regions. Regarding the de novo assembly, variant calls were replaced by complete local haplotypes, and a region was typed based on the haplotype called for the region.

      Reviewer comment: For read-depth estimation, what cutoffs were used to classify windows as deletion, WT, or duplication? How much variability was present in the data? The plot legends imply a continuous scale, but in reality, only 3 discrete colors are used (0, 1, 2), so these must represent the data after rounding.

      These have been added to the manuscript. See response to Reviewer #1 questions #2 and #3 above

      Reviewer comment: Similarly, what thresholds were used for mapping the long-reads? In Fig S21, it appears there is a high proportion of discordant reads.

      Long reads were mapped using minimap2 with default settings. For Figure 21, since it is from the mappings to 3D7 chromosome 11 and hybrid 3D7 13-11 chromosome, the genome from the duplicated region from the blue bar underneath is identical, so reads are expected to map to both since the genome regions are identical. The significance of this figure and Figure 4 is the number of long reads that span the whole chr11/13 duplicated region connection the 3D7 chromosome 11 and the hybrid proving that there are reads that start with chromosome 13 sequence and end with chromosome 11 sequence and the lack of reads that span from chromosome 13 into the 3D7 chromosome 13.

      Reviewer comment: The section on the mdr1 breakpoints is too vague.

      We have updated the methods section to be more explicit about how these breakpoints were determined.

      Reviewer comment: I assume that the "Homologous Genomic Structure" section of the Methods is the number analysis that was alluded to in the Results? As with other sections, this needs more information on exact methods and tools

      We have now updated the methods section to include exactly how the nucmer commands were run.

      Smaller comments:

      Reviewer comment: Introduction sub-header: "Precise *pfhrp2* and..."

      We have corrected the sub-header.

      Reviewer comment: Results (p.5) cite Table S4 instead of S3

      We have corrected this to Table S3.

      Reviewer comment: Results (p.5) "We identified 27 parasites with pfhrp2 deletion, 172 with pfhrp3 deletion, and 21 with both pfhrp2 and pfhrp3 deletions." This sentence makes it sound like they are 3 mutually exclusive categories. I'd suggest a rewording like "We identified 27 parasites with pfhrp2 deletion and 172 with pfhrp3 deletion. Of these, 21 contained both deletions."

      We have re-worded this sentence to the following: “We identified 26 parasites with pfhrp2 deletion and 168 with pfhrp3 deletion. Twenty field samples contained both deletions; 11 were found in Ethiopia, 6 in Peru, and 3 in Brazil, and all had the 13-11++ pfhrp3 deletion pattern.”

      Reviewer comment: The annotations used for the deletions differ between the text and the figures. It would be easier for the reader to harmonize the two if these matched.

      The figures have been updated to reflect the annotations of the text.

      Reviewer comment: Figure numbering does not match the order they are first referenced in the text

      The figure numbers have been updated to match the order in which they are first referenced.

      Reviewer comment: Results (p. 8) there is no Table S4

      This has been changed to Table S3.

      Reviewer comment: Results (p.8) mention a genome-wide number analysis, but I couldn't find these results. The referenced figure is for the duplicated region only.

      We have updated to point to the correct location of the nucmer results by adding a supplemental table with the results and updated to point to the correct figure.

      Reviewer comment: Discussion typo: "Here, we used publicly available short-read and long-read *short-read sequencing data* from..."

      This was not a typo, as we used publicly available PacBio long-read data and then generated new Nanopore long-read data. However, we did clarify this in the sentence.

      Reviewer #2 (Recommendations For The Authors):

      Introduction

      Reviewer comment: "(...) suggesting the genes have important infections in normal infections and their loss is selected against". The word "infections" is in place of "role", etc.

      We have changed the word accordingly.

      Results

      Reviewer comment: In the section "Pfhrp2 and pfhrp3 deletions in the global P. falciparum genomic dataset" it is mentioned the number of parasites with each deletion and where it is more common. "We identified 27 parasites with pfhrp2 deletion, 172 with pfhrp3 deletion, and 21 with both pfhrp2 and pfhrp3 deletions." and "Across all regions, pfhrp3 deletions were more common than pfhrp2 deletions; specifically, pfhrp3 deletions and pfhrp2 deletions were present in Africa in 43 and 12, Asia in 53 and 4, and South America in 76 and 11 parasites." It is not clear where the 21 parasites with both pfhrp2 and pfhrp3 deletions are located.

      We have specified the following in the Results section: “We identified 26 parasites with pfhrp2 deletion and 168 with pfhrp3 deletion. Twenty field samples contained both deletions; 11 were found in Ethiopia, 6 in Peru, and 3 in Brazil, and all had the 13-11++ pfhrp3 deletion pattern”

      Reviewer comment: "It should be noted that these numbers are not accurate measures of prevalence given that most WGS specimens have been collected based on RDT positivity." This, combined with the fact that subtelomeric regions are difficult to sequence and assembly, means these numbers are underestimated. I believe it should be more stressed in the text.

      We have added the following sentence, “Furthermore, subtelomeric regions are difficult to sequence and assemble, meaning these numbers may be significantly underestimated.”

      Reviewer comment: In the section "Pattern 13-11++ breakpoint occurs in a segmental duplication of ribosomal genes on chromosomes 11 and 13", Figures 2a and 2b should be mentioned in the text instead of just Figure 2.

      We have specified Figures 2a and 2b in the text now.

      Figures and Tables:

      Reviewer comment: Figure 2: I believe the color scale for percentage of identity is unnecessary given that the goal is to show that the paralogs are highly similar, and not that there is a significant difference between 0.99 and 0.998.

      Updated the color scale to represent the number of variants between segments rather than percent identity which ranges between 55-133 so that it represents something more discreet than 0.99 and 0.998.

      Reviewer comment: Adjust Figure 2b and the size of supplementary figure legends.Supplementary Figure 5-15: the legends are hard to read.

      All legends have been adjusted to be much more readable.

      Reviewer #3 (Recommendations For The Authors):

      Some minor suggestions:

      Reviewer comment: The order of the figures should follow the flow of the text, for example, Figure 5 appears in the text between Figure 1 and Figure 2.

      We have reordered the figures according to the order in which they appear in the text.

      Reviewer comment: Page 3 - "deleted parasites" - better to use: pfhrp2/3-deleted parasites.

      We have edited this accordingly.

      Reviewer comment: Define the acronyms the first time they are used, e.g. SEA.

      We have defined the acronyms accordingly.

      Reviewer comment: In the figures where pfmdr1 appears, indicate the correspondence to the full name of the gene that appears in the legend (multidrug resistance protein 1).

      Legends updated.

      Reviewer comment: Page 5 - Table S4 is missing.

      We apologize for our typo. There is no Table S4. We meant to refer to Table S3, which has been updated accordingly.

      Reviewer comment: Page 5 - "We identified 27 parasites with pfhrp2 deletion, 172 with pfhrp3 deletion, and 21 with both pfhrp2 and pfhrp3 deletions" - is it "and 21..." OR "from which, 21..."?

      We have reworded the sentence to the following: “We identified 26 parasites with pfhrp2 deletion and 168 with pfhrp3 deletion. Twenty field samples contained both deletions; 11 were found in Ethiopia, 6 in Peru, and 3 in Brazil, and all had the 13-11++ pfhrp3 deletion pattern.”

      Reviewer comment: Page 5 - "most WGS specimens have been collected based on RDT positivity." - explain better which tests are done - to detect pfhrp2, pfhrp3 or both?

      Co-occurrence is not detected?

      We used all publicly available WGS data that spanned over 30 studies, and the exact details of what RDTs were used are not readily available to fully answer this question. Though the exact details of RDTs are not known, this does not affect the deletion patterns found in the genomic data but does limit the ability to comment on how this affects prevalence. We have updated the manuscript to the following to be more explicit that we don’t have the full details: “It should be noted that these numbers are not accurate measures of prevalence, given that the publicly available WGS specimens utilized in this analysis come from locations and time periods that commonly used RDT positivity for collection”

      Reviewer comment: Supplementary Figure 1 - Legend for "Pattern" - what is the white?

      The “Pattern” refers to pfhrp3 deletion pattern with “white” being no pfhrp3 deletion. The annotation title has been changed to “pfhrp3- Pattern” to make this more clear and added to the text of the legend the following:”Of the 6 parasites without HRP3 deletion (marked as white in pfhrp3- Pattern column for having no pfhrp3 deletion),...”

      Reviewer comment: Supplementary Figure 8 - explain the haplotype rank. How was it obtained?

      The haplotype rank is based on the prevalence of the haplotype. To clarify this better the following has been added to the caption “Each column contains the haplotypes for that genomic region colored by the haplotype prevalence rank (more prevalent have a lower rank number, with most prevalent having rank 1) at that window/column. Colors are by frequency rank of the haplotypes (most prevalent haplotypes have rank 1 and colored red, 2nd most prevalent haplotypes are rank 2 and colored orange, and so forth. Shared colors between columns do not mean they are the same haplotype. If the column is black, there is no variation at that genomic window.”

      Reviewer comment: Figure 1 - Pattern in legend appears 11++13- but in text it is always referenced as 13-11++

      Figure legend has been updated to reflect the annotation within the text

      Reviewer comment: Page 6 - pattern 13- is which one(s) in Figure 1?

      This refers to the 13- with TARE1 sequence detected, the text has been updated to “(pattern 13-TARE1)” and the legend of Figure 1 has been updated so these statements match more closely.

      Reviewer comment: Page 7 - states "The 21 parasites with pattern 13-" and refers to Supplementary Figure 3 which presents "50 parasites with deletion pattern 13-". I believe this is pattern 13- unassociated with other rearrangements but it should be made clear in the text and legend of the supplementary figure.

      Thank you, you are correct. The manuscript has been updated in two locations for better clarity. The text has been updated to be “The 20 parasites with pattern 13-TARE1 without associated other chromosome rearrangements had deletions of the core genome averaging 19kb (range: 11-31kb). Of these 13-TARE1 deletions, 19 out of 20 had detectable TARE1 (pattern 13-TARE) adjacent to the breakpoint, consistent with telomere healing.” The Supplemental Figure 3 legend has been updated to “for the 48 parasites with pfhrp3 deletions not associated with pattern 13-11++”

      Reviewer comment: Supplementary figure 25 - "regions containing the pfhrp genes (lighter blue bars below chromosomes 11 and 13)" - the light blue bars are shown below chromosome 8 and 13; what is the difference between yellow and pink bars (telomere associates repetitive elements in the truncated legend)?

      The yellow bars are associated with the telomere-associated repetitive element 3 and the pink bars are telomere-associated repetitive element 1. To add clarity the legend has been updated to be “The yellow (TARE3) and pink (TARE1) bars on the bottom of the chromosomes represent the telomere-associated repetitive elements found at the end of chromosomes.”

      Reviewer comment: It would be helpful to have a positioning scale in the figures.

      Most plots have y-axis and x-axis with the genomic positioning labeled which can serve as a positioning scale so we opted not to add more to the figures to keep them less crowded. Other plots have regions plotted in genomic order but are all relatively positioned which prevents the usage of a positioning scale, we tried to clarify this by adding more details to the captions of these figures.

      Reviewer comment: Legend of Figure 6 - The last paragraph seems to be out of place

      We have deleted the last sentence in the legend of Figure 6 accordingly.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews: 

      Reviewer #1 (Public Review): 

      Strengths: 

      The paper clearly presents the resource, including the testing of candidate enhancers identified from various insects in Drosophila. This cross-species analysis, and the inherent suggestion that training datasets generated in flies can predict a cis-regulatory activity in distant insects, is interesting. While I can not be sure this approach will prevail in the future, for example with approaches that leverage the prediction of TF binding motifs, the SCRMShaw tool is certainly useful and worth consideration for the large community of genome scientists working on insects. 

      We thank the reviewer for the positive comments, and would just like to point out that we agree: while we cannot of course know if other methods will overtake SCRMshaw for enhancer prediction—we assume they will, at some point (although motif-based approaches have not fared as well in the past)—for now, SCRMshaw provides strong performance and is a useful part of the current toolkit.

      Weaknesses: 

      While the authors made the effort to provide access to the SCRMShaw annotations via the RedFly database, the usefulness of this resource is somewhat limited at the moment. First, it is possible to generate tables of annotated elements with coordinates, but it would be more useful to allow downloads of the 33 genome annotations in GFF (or equivalent) format, with SCRMshaw predictions appearing as a new feature. Also, I should note that unlike most species some annotations seem to have issues in the current RedFly implementation. For example, Vcar and Jcoen turn empty. 

      We have addressed these weaknesses in several ways:

      (1) We have created GFF versions of the SCRMshaw predictions and provide them standalone and also merged into the available annotation GFFs for each of the 33 species

      (2) We have made these GFF files, and also the original SCRMshaw output files, available for download in a Dryad repository linked to the publication (https://doi.org/10.5061/dryad.3j9kd51t0).

      (3) We have added the inadvertently omitted species to the REDfly/SCRMshaw database.

      We agree that the database functions are still somewhat limited, but note that database development is ongoing and we expect functionality to increase over time. In the meantime, the Dryad repository ensures that all results reported in this paper are directly available.

      Reviewer #2 (Public Review): 

      Summary: 

      … Upon identification of predicted enhancer regions, the authors perform post-processing step filtering and identify the most likely predicted enhancer candidates based on the proximity of an orthologous target gene. …

      We respectfully point out a small misunderstanding here on the part of the reviewer. We stress that putative target gene assignments and identities have no impact at all on our prediction of regulatory sequences, i.e., they are not “based on the proximity of an orthologous target gene.” Predictions are solely based on sequence-dependent SCRMshaw scores, with no regard to the nature or identities of nearby annotated features. Putative target genes are mapped to Drosophila orthologs purely as a convenience to aid in interpreting and prioritizing the predicted regulatory elements. We have added language on page 8 (lines 189ff) to make this more clear in the text.

      Weaknesses:

      This work provides predicted enhancer annotations across many insect species, with reporter gene analysis being conducted on selected regions to test the predictions. However, the code for the SCRMshaw analysis pipeline used in this work is not made available, making reproducibility of this work difficult. Additionally, while the authors claim the predicted enhancers are available within the REDfly database, the predicted enhancer coordinates are currently not downloadable as Supplementary Material or from a linked resource. 

      We have placed all the code for this paper into a GitHub repository “Asma_etal_2024_eLife” (https://github.com/HalfonLab/Asma_etal_2024_eLife) to address this concern. As described in our response to Reviewer 1, above, all results are now available in multiple formats in a linked Dryad repository in addition to the REDfly/SCRMshaw database.

      The authors do not validate or benchmark the application of SCRMshaw against other published methods, nor do they seek to apply SCRMshaw under a variety of conditions to confirm the robustness of the returned predicted enhancers across species. Since SCRMshaw relies on an established k-mer enrichment of the training loci, its performance is presumably highly sensitive to the selection of training regions as well as the statistical power of the given k-mer counts. The authors do not justify their selection of training regions by which they perform predictions. 

      Our objective in this study was not to provide proof-of-principle for the SCRMshaw method, as we have established the efficacy of the approach at this point in several previous publications. Rather, the objective here was to make use of SCRMshaw to provide an annotation resource for insect regulatory genomics. Note that the training regions we used here are the same as those we have used in earlier work. Naturally, we performed various assessments to establish that the method was working here, but we make no claims in this work about SCRMshaw’s relative efficiency compared to other methods. Some of our prior publications include assessments of the sort the reviewer references, which suggest that SCRMshaw is at least comparable to other enhancer discovery approaches. We note that benchmarking of such methods is in fact extremely complicated due to the fact that there are no established true positive/true negative data sets against which to benchmark (we have explored this in Asma et al. 2019 BMC Bioinformatics).

      While there is an attempt made to report and validate the annotated predicted enhancers using previously published data and tools, the validation lacks the depth to conclude with confidence that the predicted set of regions across each species is of high quality. In vivo, reporter assays were conducted to anecdotally confirm the validity of a few selected regions experimentally, but even these results are difficult to interpret. There is no large-scale attempt to assess the conservation of enhancer function across all annotated species. 

      We respectfully disagree that there is insufficient validation. We bring several different lines of evidence to bear suggesting that our results fall into the accuracy range—roughly 75%—established both here and in previous work. We are also clear about the fact that these are predictions only and need to be viewed as such (e.g. line 638). Although “large-scale” in vivo validation assays would certainly be both interesting and worthwhile, the necessary resources for such an assessment places it beyond our present capability.

      Lastly, it is suggested that predicted regions are derived from the shared presence of sequence features such as transcription factor binding motifs, detected through k-mer enrichment via SCRMshaw. This assumption has not been examined, although there are public motif discovery tools that would be appropriate to discover whether SCRMshaw is assigning predicted regions based on previously understood motif grammar, or due to other sequence patterns captured by k-mer count distributions. Understanding the sequence-derived nature of what drives predictions is within the scope of this work and would boost confidence in the predicted enhancers, even if it is limited to a few training examples for the sake of clarity of interpretation. 

      Again, we respectfully disagree that “this assumption has not been examined.” Although we did not undertake this analysis here, we have in the past, where we have shown that known TFBS motifs can be recovered from sets of SCRMshaw predictions (e.g., Kazemian et al. 2014 Genome Biology and Evolution). We return to this point when we address the Comments to Authors, below.

      Reviewer #3 (Public Review): 

      Weaknesses:  

      The rates of predicted true positive enhancer identification vary widely across the genomes included here based on the simulations and comparison to datasets of accessible chromatin in a manner that doesn't map neatly onto phylogenetic distance. At this point, it is unclear why these patterns may arise, although this may become more clear as regulatory annotation is undertaken for more genomes. 

      We agree that we do not see clear patterns with respect to phylogenetic distance in our results. However, we note that this initial data set is still fairly small, and not carefully phylogenetically distributed. We are hoping that, as the reviewer suggests, some of these questions become more clear as we add more genomes to our analysis. Fortunately, the list of available genomes with chromosome-level assembly is growing rapidly, and as we move ahead we should have much greater ability to choose informative species.

      Functional assessment of predicted enhancers was performed through reporter gene assays primarily in Drosophila melanogaster imaginal discs, a system amenable to transgenics. Unfortunately, this mode of canonical imaginal disc development is only representative of a subset of all holometabolous insects; therefore, it is difficult to interpret reporter gene expression in a fly imaginal disc as evidence of a true positive enhancer that would be active in its native species whose adult appendages develop differently through the larval stage (for example, Coleopteran and Lepidopteran legs). However, the reporter gene assays from other tissues do offer strong evidence of true positive enhancer detection, and constraints on transgenic experiments in other systems mean that this approach is the best available. 

      Please see an extensive discussion of this point in our response to Reviewer 3, below.

      Recommendations for the authors:

      Reviewer #2 (Recommendations For The Authors): 

      Major Concerns: 

      (1) While the GitHub source code for SCRMshaw is provided, the authors do not provide a repository of manuscriptspecific code and scripts for readers. This is a barrier to reproducibility and the code used to perform the analysis should be made available. Additionally, links to available scripts do not work, see Line 690. Post-processing scripts point to a general lab folder, but again, no specific analysis or code is sourced for the work in this specific manuscript (e.g. Line 637). 

      As noted above, we have corrected this oversight and established a specific GitHub repository for this manuscript “Asma_etal_2024_eLife” (https://github.com/HalfonLab/Asma_etal_2024_eLife). 

      (2) On lines 479-488, there is a discussion about the annotations being provided on REDfly, though no link is provided. 

      We have included a link in the text at this point (now line 515).

      Additionally, for transparency, it would be valuable to provide in Supplementary Table 1 the genomic coordinates of the original training sets in addition to their identity. 

      These coordinates have been added to Supplementary Table 1 as suggested.

      Also, it is suggested to provide genomic coordinates of the predicted enhancers for each training set across all species, perhaps with a column denoting a linked ID of one genomic coordinate in a species to another species (i.e. if there is a linked region found from D. melanogaster to J. coenia, labeling this column in both coordinate sets as blastoderm.mapping1_region1). Providing these annotations directly in the work enhances the transparency of the results. 

      We are unsure exactly what the reviewer means here by “a linked region.” It is critical to understanding our approach to recognize that the genome sequences have diverged to the point where there is no alignment of non-coding regions possible. Thus there is no way to directly “link” coordinates of a predicted enhancer from one species to those of a predicted enhancer in another species. The coordinates for each prediction are available on a per-species basis either through the database or in the files now available in the linked Dryad repository; these can be filtered for results from a specific training set. The database will allow users to select all results for a given orthologous locus, from any subset of species. More complex searches will continue to become available as we improve functionality of the database, an ongoing project in collaboration with the REDfly team.

      (3) Figure 2B: It is unclear what this figure shows. Are the No Fly Orthologs false positives, Orthology pipeline issues, or interesting biology? 

      We have clarified this in the Figure 2 legend. “No Mapped Fly Orthologs” indicates that our orthology mapping pipeline did not identify clear D. melanogaster orthologs. For any given gene, this could reflect either a true lack of a respective ortholog, or failure of our procedure to accurately identify an existing ortholog.

      (4) SCRMshaw appears to be a versatile tool, previously published in a variety of works. However, in this manuscript, there is little discussion of the sensitivity of SCRMshaw to different initial parameters, how the selection of training loci can impact outcomes, or how SCRMshaw k-mer discovery methods compare to other similar tools.

      - This paper would be strengthened by addressing this weakness. Some specific suggestions below: 

      In order to strengthen confidence that SCRMshaw is a reliable predictor of enhancer regions in other species, it is suggested that you benchmark against other k-mer-derived methods to assign enhancers, such as GSK-SVM developed by the Beer Lab in 2016  (https://www.beerlab.org/gkmsvm/, https://www.biorxiv.org/content/10.1101/2023.10.06.561128v1). 

      We have established the effectiveness of SCRMshaw as an enhancer discovery method in previous work, and the main goal of this study was to make use of the established method to annotate numerous insect genomes as a community resource. Our claim here is that SCRMshaw works well for this purpose; we do not attempt a strong claim about whether other approaches may work equally well or marginally better (although we do not believe this is the case, based on prior work). Benchmarking enhancer discovery is challenging, as we point out in Asma et al. 2019 (BMC Bioinformatics), and, while important, best left for a dedicated comprehensive study. A major problem is that there are no independent objective “truth” sets for enhancers from the various species we interrogate here. Thus, while we could also run, e.g., GSK-SVM, what criteria would we use to establish which method had better accuracy for a given species? Note that the work from Beer’s lab took advantage of the ability to match human-mouse orthologous (or syntenic) regions and available open-chromatin data to assess whether conserved enhancers were discovered, but this is not possible given the degree of divergence, limited synteny, and relative lack of additional data for the insect genomes we are annotating.

      - In Table S1, we see that 7-146 regions are used as training sets, which is a huge variety. Does an increase in training set size provide a greater "rate of return" for predicted regions? Is the opposite true? Addressing this question would allow readers to understand if they wish to use SCRMshaw, a reasonable scope for their own training region selections. 

      - Within a training set, does subsampling provide the same outcomes in terms of prediction rates? There is no exploration of how "brittle" the training sets are, and whether the generalized k-mer count distributions that are established in a training set are consistent across randomly selected subgroups. Performing this analysis would raise confidence in the method applied and the resulting annotations. 

      These are interesting and important questions, but again we feel they are beyond the scope of this particular study, which is focused primarily on using SCRMshaw and not on optimizing various search parameters. That said, this is of course something we have investigated, although as with other aspects of enhancer discovery, the absence of a true gold standard enhancer set makes evaluation difficult. We have not found a clear correlation between training set size and performance beyond the very general finding that performance appears to be best when training set size is moderate, e.g. 20-40 initial enhancers. We suspect that larger training sets often contain too many members that don’t fit the core regulatory model and thus add noise, whereas sets that are too small may not contain enough signal for best performance (although small sets can still be useful, especially if used in an iterative cycle; see Weinstein et al. 2023 PLoS Genetics). However, establishing this rigorously is highly challenging given the limitations with assessing true and false positive rates at scale.

      (5) In Figure 2C, when plotting hexMCD, IMM, pacRC, and then the merged set, it is unclear whether the scorespecific bar allows coordinate redundancy, though this is implied. What might be more useful is a revision of this plot where the hexMCD/IMM/pac-RC-specific loci are plotted, with the merged set alongside as is currently reported. This would give the reader a clearer understanding of the variability between these scoring methods and why this variability occurs. 

      We have added the breakdowns between IMM, hexMCD, and pacRC in Supplementary Table S2, and made more complete reference to this in the text (lines 682ff). Both the database and the data files in the Dryad repository allow exploration of the overlap between the different methods and contain both separate and merged (for overlap and redundancy) results.

      Additionally, there is no information in the Methods section of these three SCRMshaw scores and what they represent, even colloquially. While SCRMshaw has been applied in several papers previously, it would help with scientific clarity to describe in a sentence or two what each score is meant to represent and why one is different from another. 

      We had chosen to err on the side of brevity given prior publication of the SCRMshaw methodology, but we recognize now that we went too far in that direction. We have added more complete descriptions of the methods in both the Results (lines 164-167) and the Methods (lines 667-681) sections.

      (6) When describing results in Figure 2, an important question arises: "Is there an anti-correlation between the number of predicted regions and evolutionary distance?" This would be an expected result that could complement Figure 4's point that shared orthology across 16 species is rarer than across 10 species. Visualizing and adding this to Figure 2 or Figure 4 would be a powerful statement that would boost confidence in the returned predicted enhancers and/or orthologous regions. 

      This is an important question and one in which we are very interested. Unfortunately, we do not have sufficient data at this time to address this proper statistical rigor. As we remarked above in response to Reviewer 3, “We agree that we do not see clear patterns with respect to phylogenetic distance in our results. However, we note that this initial data set is still fairly small, and not carefully phylogenetically distributed. We are hoping that, as the reviewer suggests, some of these questions become more clear as we add more genomes to our analysis. Fortunately, the list of available genomes with chromosome-level assembly is growing rapidly, and as we move ahead we should have much greater ability to choose informative species.”

      (7) In Figure 3, the authors seek to convey that SCRMshaw predicts enhancer regions that are mapped nearby one another, across different loci widths, and that this occurrence of nearby predicted regions occurs more than a randomly selected control. This is presumably meant to validate that SCRMshaw is not providing predictions with low specificity, but rather to highlight the possibility that SCRMshaw is identifying groups of shadow enhancers. However, these plots are extremely difficult to decipher and do not strongly support the claims due to the low resolution and difficult interpretability of the boxplot interquartile distributions.

      Additionally, as the majority of predicted regions are around ~750bp, how does that address loci groups of <1000bp? This suggests that predicted regions are overlapping, and therefore cannot be meaningfully interpreted as shadow enhancers. This plot should either be moved to the supplements or reworked to more effectively convey the point that "SCRMshaw is detecting predicted regions that are proximal to one another and that this proximity is not due to chance". 

      - A suggestion to rework this plot is to change this instead to a bar plot, where the y-axis instead represents "number of predictions with at least 2 predicted regions proximal to one another" divided by "total number of predictions", separating bar color by simulated/observed values. The x-axis grouping can remain the same. Because this plot is a broad generalization of the statement you're trying to make above, knowing whether a few loci have 2 versus 4 proximal predicted enhancers doesn't enhance your point. 

      We agree with the reviewer that these are not the clearest plots, and thank them for the suggestions regarding revision. We tried many variations on visualizing these complex data, including those suggested by the reviewer, and have concluded that despite their weaknesses, these plots are still the best visualization. The main problem is that the observed data cluster heavily around zero, so that the box plots are very squat and mainly only the outlier large values are observed. The key point, however, is that the expected values almost never give values much greater than one, so that the observed outlier points are the only points seen in the upper ranges of the y-axis. This is true across the three species, across the bins of locus sizes, and across training sets (averaged into the box plots). The reviewer is correct as well about the bins where locus size is < 1000. However, inspection of the data shows that this is not a large concern, as very few data points lie in this range and we never see multiple predicted enhancers there. Thus we believe while not the prettiest of graphs, Figure 3 does effectively support the claims made in the text. In keeping with our view that it is preferable to have data in the main paper whenever possible, we choose to keep the figure in place rather than move it to the Supplement.

      - Label the species for the reader's understanding of each subplot on the plot. 

      We apologize for this oversight and have now labeled each plot with its relevant species.

      (8) SCRMshaw operates on k-mer count distributions compared to a genomic background across different species, allowing it to assign predicted regions without prior knowledge of an organism's cis-regulatory sequences. This is powerful and boosts the versatility of the method. However, understanding the cis-regulatory origins of the kinds of kmers that are driving the detection of orthologous regions across species is crucial and absolutely within the scope of the paper, particularly for the justification of the provided annotations. Is SCRMshaw making use of enriched motifs within the training region set to assign regions in other species? One would presume so, but it is necessary to show this. There are many motif discovery tools that are readily available and require little up-front knowledge and little to no use of a CLI, such as MEMESuite (https://meme-suite.org/meme/tools/meme). It is highly recommended that, even for a few training pairs that are well understood (e.g. mesoderm.mapping1, dorsal_ectoderm.mapping1), assess the motif enrichment within the original sequence set, then see whether motif enrichments are reflected in the predicted enhancers. As evolutionary distance increases between D. melanogaster and the species of interest, is the assignment of enriched motifs more sparse? Is there a loss of a key motif? These are the kinds of questions that will allow readers to understand how these annotations are assigned as well as boost confidence in their usage. 

      This is a very important point and a subject of significant interest to us. We have demonstrated in earlier work (e.g., Kazemian et al. 2014 Genome Biol. Evol.) that SCRMshaw-predicted enhancers do contain expected TFBS motifs, across multiple species—and that even an overall arrangement of sites is sometimes conserved. Thus we have previously answered, in part, the reviewer’s question. 

      What we also learned from our previous work is that filtering out relevant motifs from the noise inherent in motif-finding is both arduous and challenging. As the reviewer is no doubt aware, while using motif discovery tools is simple, interpreting the output is much less so. In response to the reviewer’s comments, we revisited this issue with data from a small sample of training sets. We can discover motifs; we can see that the motif profiles are different between different training sets; and we can observe the presence of expected motifs based on the activity profile of the enhancers (e.g., Single-minded binding sites in our mesectoderm/midline training and result data). However, to do this cleanly and with appropriate statistical rigor is beyond what we feel would be practical for this paper. We hope to return to this important question in the future when we have a larger and phylogenetically more evenly-distributed set of species, and the time and resources to address it appropriately.

      (9) Figures 5-7 need to have better descriptions. 

      We have added to the figure 6 and 7 legends in response to this comment; please note as well that there is substantial detail provided in the text. If there are specific aspects of the figures that are not clear or which lack sufficient description, we are happy to make additional changes.

      Minor Concerns 

      (1)  In Figure 1A, it is implied that "k-mer count distributions" are actually only "5-mer count distributions". However, in the published documentation of SCRMshaw, it is suggested that k-mers between 1-6 bp are involved in establishing sequence distributions. Please add a justification for the selection of these criteria. It would be helpful to understand the implications of using up to a 3-mer versus a 12-mer when assessing k-mer counts using SCRMshaw.

      We have clarified in the Figure 1 legend that this is just an example, and the k-mers of different sizes are used in the IMM method; we have also increased the description of the basic method in the Methods section. To be clear, the hexMCD sub-method is 6-mer based (5th-order Markov chain), as is pacRC, while the IMM method considers Markov chains of orders 0-5.

      (2) Control the y-axis to remove white space from Figure 2D. 

      We have amended the figure as suggested.

      Additionally, expand in the manuscript on expected results from SCRMshaw. Given training regions of 750 bp, is the expectation that you return predicted enhancers of the same length? This is not explicitly stated, only a description of outliers. 

      The scoring is not dependent on the length of the training sequences, and there is no direct expectation of predicted enhancer length. Scores are calculated on 10-bp intervals, and a peak-calling algorithm is used to determine the endpoints of each prediction based on where the scores drop below a cutoff value. Thus there is no explicit minimum prediction length beyond the smallest possible length of 10-bp. That said, the initial scoring takes place over a 500-bp sequence window (for reasons of computational efficiency), which does influence scores away from the smaller end of the possible range. We correct for this in part by reducing scores below a certain threshold to zero, to prevent multiple low-scoring regions from combining to give a low but positive score over a long interval. Indeed, we found that in the original version of SCRMshawHD (Asma et al. 2019), multiple low-scoring but above-threshold intervals would get concatenated together in broad peaks, leading to an unrealistically large average prediction length. In the version used here, described in Supplementary Figure S6, low-scoring windows are now first reset to zero and a new threshold is calculated before overlapping scores are summed. This helps to prevent the broad peak problem, and we find that it results in a median prediction length ~750 bp, more in line with expected enhancer sizes.

      Reviewer #3 (Recommendations For The Authors): 

      Line 161: Given that the SCRMshaw HD method is the basis for the pipeline, the methodology deserves at least an "in brief" recapitulation in this manuscript. 

      As we remark in our response to Reviewer 2, above, “We had chosen to err on the side of brevity given prior publication of the SCRMshaw methodology, but we recognize now that we went too far in that direction. We have added more complete descriptions of the methods in both the Results (lines 164-167) and the Methods (lines 667-681) sections.” 

      Line 219: Throughout the reporting of the results, there appeared to be a bit of inconsistency/potential typos regarding whether threshold or exact P values were reported. In lines 219, 222, 265, 696, and 811, the reported values seem to clearly be thresholds (< a standard cutoff), while in lines 291,293, 297,300, values appear to be exact but are reported as thresholds (<). 

      This is not an error but rather reflects two different types of analysis. The predictions per locus (originally lines 219, 222 etc) are evaluated using an empirical P-value based on 1000 permutations. As such, they are thresholded at 1/1000. The overlap with open chromatin regions, on the other hand, are based on a z-score with the P-values taken from a standard conversion of z-scores to P-values.

      Page 13/Table 2: At face value, it seems surprising that the overlap between Dmel SCRMshaw predictions with open chromatin is so much smaller than the overlap between predictions and open chromatin in other species, both in raw % (Tcas, D plexippus, H. himera) and fold enrichment (Tcas), given that the training sets for SCRMshaw are all derived from Dmel data. The discussion here does not touch on this aspect of the results, and the interpretation of this approach, in general, would be strengthened if the authors could comment on potential reasons why this pattern may be arising here, or at least acknowledge that this is an open question.

      There are many variables at play here, as the data are from different species, from different tissues, and from different methods. Thus we think it is difficult to read too much into the precise results from these comparisons—the main take-home is really just that there is a significant amount of overlap. In acknowledgment of this, we have slightly modified the text in this section so that it now notes (line 302ff): “These comparisons are imperfect, as the tissues used to obtain the chromatin data do not precisely correspond to the training sequences used for SCRMshaw, and the data were obtained using a variety of methods.”

      Line 318-329: The inferences from the reporter gene assay deserve a more nuanced treatment than they are given here. The important nuance that was not addressed by the discussion here is that the imaginal disc mode of development in Drosophila is not broadly representative of the development of larval/adult epithelial tissues across Holometabola; thus, inference of a true positive validation becomes complicated in cases where predicted enhancers from a species were tested and shown to drive expression in a fly imaginal disc that the native species have no direct disc counterpart to. For example, in line 388 a Tcas enhancer is reported to drive expression in the eye-antennal disc, and in lines 404 and 423 additional Tcas enhancers were reported to drive expression in the leg discs; however, Tribolium larvae do not possess antennal discs or leg discs set aside during embryogenesis in the sense that flies do - instead the homologous epithelial tissues form larval antennae and larval legs external to the body wall that are actively used at this life stage and are starkly different in morphology than an internally invaginated epithelial disc, that will directly give rise to adult tissues in subsequent molts. Is the interpretation of an expression pattern driven in a fly disc as a true positive really as straightforward as it was presented here, when in the native species the expression pattern driven by the enhancer in question would be in the context of an extremely different tissue morphology? That said, I understand and am deeply sympathetic to the constraints on the authors in performing transgenic experiments outside of the model fly; but these divergent modes of development across Holometabola deserve a mention and nuance in the interpretation here. 

      This is indeed a very important point, and we greatly appreciate Reviewer 3 pointing out this caveat when interpreting the outcomes of our cross-species reporter assay. Reviewer 3 is correct that the imaginal disc mode of adult tissue (i.e. imaginal) development found in Diptera does not represent the imaginal development across Holometabola. 

      In fact, imaginal development is quite diverse among Holometabola. For instance, larval leg and antennal cells appear to directly develop into the adult legs and antennae in Coleoptera (i.e. primordial imaginal cells function as larval appendage cells), while some cells within the larval legs and antennae are set aside during larval development specifically for adult appendages in Lepidopteran species (i.e. imaginal cells exist within the larval appendages but do not contribute to the formation of larval appendages). In contrast, an almost entire set of cells that develop into adult epithelia are set aside as imaginal discs during embryogenesis in Diptera. Furthermore, the imaginal disc mode of development appears to have evolved independently in

      Hymenoptera. Therefore, determining how imaginal primordial tissues correspond to each other among Holometabola has been a challenging task and a topic of high interest within the evo-devo and entomology communities.

      Nevertheless, despite these differences in mode of imaginal development, decades of evo-devo studies suggest that the gene regulatory networks (GRNs) operating in imaginal primordial tissues appear to be fairly well conserved among holometabolan species (for example, see Tomoyasu et al. 2009 regarding wing development and Angelini et al. 2012 regarding leg development between flies and beetles). These outcomes imply that a significant portion of the transcriptional landscape might be conserved across different modes of imaginal development. Therefore, an enhancer functioning in the Tribolium larval leg tissue (which also functions as adult leg primordium) could be active even in the leg imaginal disc of Drosophila, if the trans factors essential for the activation of the enhancer are conserved between the two imaginal tissues. 

      That being said, we fully expect there to be both false negative and false positive results in our cross-species reporter assay. We are optimistic about the biological relevance of the positive outcomes of our crossspecies reporter assay, especially when the enhancer activity recapitulates the expression of the corresponding gene in Drosophila (for example, Am_ex Fig6B and Tc_hth Fig7B). Nonetheless, the biological relevance of these enhancer activities needs to be further verified in the native species through reporter assays, enhancer knock-outs, or similar experiments.

      In recognition of the Reviewer’s important point, we added the following caveat in our Discussion (lines 549553): “Furthermore, the unique imaginal disc mode of adult epithelial development in D. melanogaster  might have prevented some enhancers of other species from working properly in D. melanogaster imaginal discs, likely producing additional false negative results. Evaluating enhancer activities in the native species will allow us to address the degree of false negatives produced by the cross-species setting.” We moreover mention this caveat in the Results section when we first introduce the reporter assays (line 342).

      Line 580: This is the first time that the weakness of the closest-gene pairing approach is mentioned. This deserves mention earlier in the manuscript, as unfortunately, this is one of the major bottlenecks to this and any other approaches to investigating enhancer function. Could the authors address this earlier, perhaps pages 7-8, and provide citations for current understanding in the field of how often closest-gene pairing approaches correctly match enhancers to target genes? 

      We have added text as suggested on p.7-8 acknowledging the shortcomings of the closest-gene approach. We also clarify at the end of that section (lines 173-181) that target gene assignments, while useful for interpretation, have no bearing on the enhancer predictions themselves (which are generated prior to the target gene assignment steps).

    1. Welcome back. In this part of the series, I'm going to be primarily covering the transport layer, which is layer 4 of the OSI model.

      I'm also going to be touching upon layer 5, which is the session layer of the OSI model, because there is some overlap for certain features, and so it's easier to cover them in one lesson.

      The transport layer runs over the top of the network layer and provides most of the functionality, which supports most of the networking, which we use day-to-day on the internet.

      The session layer runs on top of the transport layer, and many features, which you might use, are often mixed between these two layers.

      Now, as I've already mentioned, it's not generally worth the argument of deciding whether things are covered in layer 4 or layer 5, so I'd only explain both of these layers as one grouping of functionality.

      The OSI model is conceptual, after all, and many things exist between or across two different layers.

      Now, we've got a lot to cover, so let's jump in and get started.

      Before we get started with layer 4, I want to summarize the situation and limitations with layer 3.

      Now, we have a functional layer 3, which means that we can communicate between two devices, say a source and destination laptop, using a source and destination IP address.

      If both of these use public IP addresses, it doesn't matter where on the internet these devices are, layer 3 and IP routing will ensure that any packets generated and sent from the source laptop will move across any layer 2 networks between the source and destination.

      Let's say that using layer 3, the source laptop on the top generates 6 IP packets, and these are all destined for the destination laptop at the bottom right.

      The important thing to understand about layer 3 in this context is that each packet is a separate and isolated thing, and it's routed independently over the internet.

      It might be logical to assume that the packets arrive in the same state, so the same timing, the same order, and the same quality, but sadly, that's not true.

      In ideal conditions, yes, but generally, if you're communicating using only IP, then you're going to have intermittent network conditions, and that can result in a few cases where the arrival condition of packets is different than the condition when they were generated and sent.

      One of the first things which we might encounter is out-of-order arrival.

      In this case, where packet 3 arrives before packet 2, layer 3, specifically IP, provides no method to ensure the ordering of packet arrival.

      For applications which only used IP, this would mean complex logic would need to be built into the application to ensure packets could be sequenced in the same way, and this is not a trivial task.

      Because each packet is routed as an independent thing, it's possible packet 2 could have taken a slow, less efficient route, which is why it arrives later.

      This is a negative of layer 3, which can be fixed at layer 4.

      Another issue with layer 3 is that packets can just go missing.

      This can be due to network outages or network conditions, which cause temporary routing loops.

      Remember, when I talked about packet structure, I talked about the TTL field, which limited the number of hops a packet could go through.

      Well, if the number of hops exceeds this, then it will be discarded.

      With IP, there's no reliable method of ensuring packet delivery, and so it's a relatively regular occurrence that packets go missing.

      Now, network conditions can also cause delay in delivery, and for any latency-sensitive applications, this can cause significant issues.

      The key thing to keep in mind about layer 3, every packet is different.

      It's single, it's isolated.

      It's a different unit of data which is being routed across a layer 3 network using layer 2 networks as transit.

      What happens to one packet might not happen or might happen in a different way to another packet.

      Another limitation with layer 3, and this one is probably the one which has the most obvious effect, is that if you think back to the structure of IP packets, they have a source and destination field.

      They don't have anything beyond that to distinguish channels of communication.

      Packets from a source IP to a destination IP, they're all the same.

      You couldn't have two applications running on the source IP, communicating with two applications running on the destination IP, because there's no method of distinguishing between the applications.

      Any packet sent by one application would look to be the same as one sent by another.

      Think about what you're doing on your device right now.

      You might be watching a video.

      Do you have a web browser open doing something else?

      Do you have an SSH connection or email or any other application which uses the internet?

      This means multiple applications, and IP on its own offers no way to separate the packets for individual applications.

      This is something which is remedied at layer 4.

      Lastly, IP has no flow control.

      If a source device is transmitting packets faster than a destination device can receive them, then it can saturate the destination connection and cause loss of data, packets which will be dropped.

      Now with only layer 3, we wouldn't have anywhere near the flexibility required to have the internet function in the way that it does.

      For that, we need layer 4, and that's what I want to cover in this part of the lesson series.

      So what is layer 4 and how does it function?

      Let's take a look.

      So far, this is what we have network model-wise.

      We've discussed the physical layer which is layer 1 at the OSI model.

      This relates to how raw bit screen data is transmitted to or received from physical shared media.

      We've talked about layer 2 which adds identifiable devices, switches and media access control, but layer 2 ends with isolated layer 2 networks.

      In the previous part of this lesson series, I introduced layer 3 which adds IP addressing and routing, so packets can be routed from source to destination across multiple interconnected networks.

      Layer 4 builds on top of this.

      It adds two new protocols, TCP which stands for transmission control protocol and UDP which stands for user datagram protocol.

      Now both of these run on top of IP, and both of them add a collection of features depending on which one of them is used.

      Now if you've heard the term TCP/IP, that means TCP running on top of IP.

      At a high level, you would pick TCP when you want reliability, error correction and ordering of data.

      It's used for most of the important application layer protocols such as HTTP, HTTPS, SSH and so on.

      Now TCP is a connection-oriented protocol which means that you need to set up a connection between two devices and once set up, it creates a bidirectional channel of communications.

      UDP on your hand is faster because it doesn't have the TCP overhead required for the reliable delivery of data.

      This means that it's less reliable.

      Now there's a great joke about UDP.

      I will tell you about it, but you might not get it.

      Anyway, it's a good job my lessons are better than my jokes.

      In this lesson, I'm going to spend most of my time talking about TCP because it's used by more of the important protocols that you'll use day-to-day on the internet.

      But just know that both TCP and UDP, they both run on top of IP and they're used in the same way.

      They use IP as transit.

      TCP just offers a more reliable connection-oriented architecture whereas UDP is all about performance.

      So there's a simple trade-off.

      Now for this lesson series, as I talk about, I'm going to be focusing on TCP because that's what's used for most of the important upper layer protocol.

      So let's take a look at exactly how TCP works.

      TCP introduces something called segments.

      Now a segment is just another container for data like packets and frames before them.

      Segments are specific to TCP.

      Before we get started talking about the segments themselves, it's important to understand that segments are actually contained in which is known as encapsulated within IP packets.

      So let's say that we have a stream of packets.

      You know by now that these are all isolated packets.

      They're just pieces of data which are routed independently from source to destination.

      They're all treated separately.

      Well, TCP segments are placed inside packets and the packets carry the segments from their source to their destination.

      Segments don't have source or destination IP addresses because they use the IP packets for the transit from source to destination.

      This is all handled by layer 3.

      In this case, the internet protocol.

      TCP segments add additional capabilities to IP packets.

      Let's step through the structure of segments so that we can fully understand them.

      And I'm going to skip past a few attributes of segments just as I did with layer 3 because there are some parts which are less important or less situational.

      So I won't be covering either the options or padding fields within a segment.

      The first fields which I want to cover are the source and destination ports.

      In addition to the source and destination IP addresses that IP packets provide, TCP segments add source and destination ports.

      And this gives the combined TCP/IP protocol the ability to have multiple streams of conversations at the same time between two devices.

      When you open the AWS web interface, you're communicating from a port on your local machine to a port on the AWS servers, TCP port 443, which is HTTPS.

      Now because of port, you can have multiple streams of communication from your machine.

      One to AWS, one to Netflix, and one to this website where you're watching this video.

      At the other side, AWS can have multiple streams of communication to their servers.

      Each conversation is a unique combination of the source and destination IP, the source port, and the destination port.

      All of these four values together identify as a single conversation, a single communications channel.

      These two fields ought to allow the internet to function in a flexible way that it does.

      It's why SSH and HTTPS can exist on the same EC2 instance and why you can have multiple SSH connections open to the same EC2 instance if you wanted to.

      And I'll cover more on how this works as we move through this lesson.

      Now next, within the segment, is sequence.

      And the sequence number is incremented with each segment that's sent.

      And it's unique.

      It can be used for error correction if things need to be retransmitted.

      You can use to ensure that one IP pass is received and the TCP segments are pulled out.

      They can be correctly ordered.

      So the sequence number is already uniquely identifying the particular segment within a particular connection so that both sides can make observations about it.

      And the way that these observations are done is using app knowledgements.

      The app knowledgement field on the right, on the one side, can indicate that it's received up to and including a certain sequence number.

      Every segment which is transmitted needs to be acknowledged.

      Remember that TCP is a reliable protocol and so if the device is transmitting segment one, two, three, and four to another device, then the other device needs to acknowledge that it's received segment one, two, three, and four.

      And this is what the app knowledgement field is for.

      So sequence number and app knowledgement are used hand in hand.

      Next we have a field called flags and things.

      Now within a segment, there is an actual flags component which is nine bits.

      And this allows various controls over the TCP segments and the wider connection.

      Flags are used to close the connection of the synchronized sequence numbers, but there's also additional things like a data offset and some reserved space.

      So I thought this flags and things is essentially the flags plus a number of extra fields which I don't need to go into at this point in the lesson.

      Now next we've got the TCP window.

      And this is interesting.

      This defines the number of bytes that you indicate that you're willing to receive between app knowledgements.

      Once reached, the sender will pause until you acknowledge that amount of data.

      And this is how flow control is implemented.

      It lets the receiver control the rate at which the sender sends data.

      If you use a smaller window, it provides additional levels of control over how quickly you're sent data.

      Larger windows are more efficient because the header of a TCP segment takes up an amount of space and the smaller the window, the more headers are involved.

      So this window setting is quite important if you're using a TCP for practical reasons, but we don't need to go into too much detail in this lesson.

      Next we have checks on which is used for error checking.

      It means that a TCP layer is able to detect errors and can arrange for retransmission of the data as required.

      And then lastly, we have the urgent pointer.

      And this is a cool feature.

      Imagine if you have a data transfer application where 99% of data is the data being transferred and 1% is control traffic.

      So communication between the client and the server, coordinating the actual data transfer.

      While setting this field in a segment means that both sides can have separate processing.

      So the control traffic always takes priority within the communication.

      So any protocols which are latency sensitive and transfer data such as FTP and PellNet can use this field.

      Now all of these fields together are known as the TCP header.

      And the capacity of a TCP segment which remains is logically enough used for data.

      So that's a segment that are placed inside packets and transmitted by one network stack, specifically layer 4 of one network stack and received by another network stack using the layer 4 protocol.

      In this case TCP.

      Okay so this is the end of part 1 of this lesson.

      It was getting a little bit on the long side and so I wanted to add a break.

      It's an opportunity just to take a rest or grab a coffee.

      Part 2 will be continuing immediately from the end of part 1.

      So go ahead, complete video and when you're ready join me in part 2.

    1. Welcome back, this is part two of this lesson.

      We're going to continue immediately from the end of part one.

      So let's get started.

      Now the only thing that remains is just to test out this configuration.

      And to do that we're going to launch an EC2 instance into the WebA subnet.

      So click on services and just type EC2 to move across to the EC2 console.

      Now once we're on the EC2 console, just click on launch instance.

      Then you'll be taken to the launch instance console.

      Into the name box, just go ahead and type a4l-bastian.

      Scroll down and we're going to create a bastion instance using Amazon Linux.

      So click on Amazon Linux.

      In the dropdown below, go ahead and select the latest version of Amazon Linux.

      Just make sure that it does say free tier eligible on the right of this dropdown.

      Assuming that's all good, just below that make sure that in the architecture dropdown it's set to 64-bit x86.

      Moving down further still, under instance type, just make sure that this is set to a free tier eligible instance.

      It should default to T2.micro or T3.micro.

      Depending on your region, either of these could be free tier eligible.

      In my case it's T2.micro, but whatever your shows, just make sure that it's similar sized and says free tier eligible.

      Now directly below that, under key pair, just click in this box.

      You should at this point in the course have a key pair creator called a4l.

      If you do, go ahead and select that key pair in the box.

      If you don't, don't worry, you can just go ahead and click on create new key pair.

      Enter a4l into the key pair name, select RSA, and then select PEM for the private key format and click on create key pair.

      This will download the key pair to your local machine and then you can continue following along with this video.

      So select that from the dropdown.

      Directly below, under network settings click on edit.

      This instance is going to go into the animals for live VPC.

      So click on the VPC dropdown and select a4l-vpc1.

      Directly below that, click in the subnet dropdown and we want to go ahead and look for sn-web-a.

      So select the weba subnet.

      This should change both of the dropdowns below.

      So auto assign public IP and auto assign IPv6 IP to enable.

      So just make sure that both of those are set to enable.

      Directly below this, make sure that create security group is checked.

      We're going to create a new security group.

      Under security group name, just go ahead and enter a4l-bassian-sg and then put that same text in the description box directly below.

      Now all of these defaults should be good.

      Just make sure it's set to SSH, source anywhere.

      Make sure that 0.0.0.0/0 and double colon 4/0 are both present directly below source.

      Everything else looks good.

      We can accept the rest of the defaults.

      Just go ahead and click on launch instance.

      Then click on instances at the top left of the screen.

      At this point the instance is launching and we'll see that a4l-bassian is currently running.

      We'll see the status check is showing initializing.

      So we need to give this instance a few minutes to fully provision.

      So go ahead and pause this video and we're going to resume it once this instance is ready to go.

      And it has two out of two status checks.

      So our instance is now showing two out of two status checks.

      And that means everything's good and we're ready to connect.

      Now if you select the instance, you'll see in the details pane below how it has a public IP version 4 address, a private IP version 4 address, public and private IP version 4 DNS.

      And if we scroll down, you'll see lots of other information about this instance.

      Now we're only concerned with the public IP version 4 address.

      We're going to go ahead and connect to this instance this time using a local SSH client on our machine.

      So right click and then select connect.

      Now if we want to quickly connect into this instance, we can choose to use EC2 instance connect, which is a way to connect into the instance using a web console.

      Now this does need an instance with a public IP version 4 address, but we have allocated a public address.

      So if we wanted to, we can just make sure that the username is correct.

      It should be EC2-user.

      If we hit connect, it will open up a connection to this instance using a web console.

      And this is often much easier to connect to EC2 instances if you don't have access to a local SSH client, or if you just want to quickly connect to perform some administration.

      We can also connect with an SSH client.

      If we select SSH client, it gives us the commands to run in order to connect to this EC2 instance.

      So right at the bottom is an example connect command.

      So SSH, we pick the key to use and then we pick the user at and then the public IP version 4 DNS.

      So if we copy that into our clipboard and then move across to our terminal or command prompt, move into the folder where you downloaded the SSH keybearant to, in my case downloads, and paste in that command and press enter, that should connect us to the EC2 instance.

      We'll have to verify the fingerprint, so we need to verify the authenticity of this host.

      For this purpose, we can just go ahead and answer yes and press enter.

      Now if it's the first time we're connecting using a particular key, and if you're running either macOS or Linux, you might be informed that the permissions on the key are too open.

      In this case, the permissions are 0, 6, 4, 4, which are too open and we get this error.

      Now it's possible to correct that if we move back to the AWS console.

      It also gives us the command to correct these permissions.

      So CHmod, space 400, space and then the name of the key.

      So I'm going to copy that into my clipboard and move back to my terminal, paste that in and press enter, and that will correct those permissions.

      Now if I get the connection command again, so copy that into my clipboard, and this time I'll paste it in and press enter and now I will be connected to this EC2 instance.

      Now if you're doing this demonstration on Windows 10, you probably won't have to correct those permissions.

      This is something specific to macOS or Linux.

      So whenever you're connecting to EC2 instances which have a public IP version 4 address, you've always got the ability to use either EC2 instance connect or a local SSH client.

      Now the third option which is session manager, this is a way that you can connect to instances even if they don't have public IP version 4 addressing.

      And I'll be detailing this product fully later on in the course because there is some additional configuration that's required.

      Now this bastion host, it's an EC2 instance and it does fall under the free tier.

      So because it's a T2.micro or whatever type of instance you picked which falls under the free tier, you're not going to be billed for any usage of this instance in a given month.

      Now as a general rule, as you're moving through the course, if you're ever intending to take a break, then you always have the option of deleting all of the infrastructure that you've created within a specific demo lesson.

      So most of the more complex demo lessons that you'll have moving through the course, at the end of every demo lesson there will be a brief set of steps where I explain how to clean up the account and return it into the same state as it was at the start of the lesson.

      But in certain situations I might tell you that one option is not to delete the infrastructure.

      Whether you do delete it or not depends on whether you're intending to complete the next demo straight away or whether you're taking a break.

      Now in this particular case I'm going to demonstrate exactly how you can clear up this infrastructure. [background noise] [background noise] In the next demo lesson you're going to be continuing using this structure, but I'm going to demonstrate how you can automate the creation using a CloudFormation template.

      To clear up this infrastructure though, go ahead, right click on this bastion host and select terminate instance.

      You'll need to click terminate to confirm and that will terminate and delete the instance.

      You won't be charged for any further usage of that instance.

      We need to wait for that instance to fully terminate, so pause the video and wait for it to move into a terminated state and then we can continue.

      So that instance is terminated and now that that's done, we can click on services and move across to the VPC console and we're going to delete the entire Animals for Life VPC.

      And don't worry, in the next demo lesson I'll explain how we can automate the creation.

      So for now and in the course we're going to be using much more automation so that anything that you've done previously, we're going to automate the creation and focus your valuable time only on the things that you've just learned.

      So click on your VPCs.

      It should list two VPCs, the default one and the Animals for Life VPC.

      Select the Animals for Life VPC, click on Actions and then delete the VPC.

      Now this is going to delete all of the resources that are associated with this VPC.

      So the internet gateway, the route tables and all of the subnets that you've created as part of the demo lessons to this point in the course.

      So go ahead and type delete and then click delete to confirm that process and that will fully tidy up the account and return it into the same state as it was at the start of the VPC section.

      Now with that being said, this is the end of this lesson.

      You've successfully converted three subnets, so Web A, Web B and Web C to be public and you've done that by creating an internet gateway, associating that with the VPC, creating a route table, associating that with those subnets, adding two routes, pointing those routes at the internet gateway and then configuring those subnets to allocate a public IP version for address to any resources launched into those subnets.

      So that's the same set of steps that you'll need to do to make any subnets public from an IP version for perspective in future.

      So this is going to be the same tasks that you would use in larger production projects.

      Although in production, you would probably automate it and don't just show you how to do that as you move through the course.

      Now at this point, you've finished everything that you need to do in this demo lesson, so great job.

      You've actually created something that is production ready and production useful.

      Over the remainder of this section of the course, we're going to refine the design that we've got and add additional capabilities.

      So in the upcoming lessons, I'll be talking about network address translation and how that can be used to give private EC2 instances access to the internet for things like software updates.

      We'll be talking about the security of subnets using network access control lists known as knuckles and much, much more.

      But you're doing a fantastic job so far.

      This is not a trivial thing that you've implemented to this point.

      So really great job.

      But at this point, just go ahead and complete the video.

      And then when you're ready, I look forward to you joining me in the next. [no audio]

    1. Borrowing the metaphors,'museums of identity and mentality would be the ''mirrors and windows''that permit to reflect us and pass throt1gh our identity and otherness

      Mirrors are for locals and windows are for those not as familiar with the area.

    Annotators

    1. From reading the book, I learned that Cutler had the same mentality for his OS and, in fact, the system wasn’t ported to x86 until late in its development. He wanted developers to target non-x86 machines first to prevent sloppy non-portable coding, because x86 machines were already the primary desktops that developers had. In fact, even as x86 increasingly became the primary target platform for NT, the team maintained a MIPS build at all times, and having most tests passing in this platform was a requirement for launching.

      I'm reminded of the time when it was revealed about 10–15 years or so ago that when Apple switched to x86 from PowerPC, it wasn't the result of a big porting effort. They'd been maintaining portability all along—doing private builds internally that just never saw the light of day. When this came to light, the reaction was huge. People were awed.

      A few years ago, when this piece of trivia was brought back to the forefront of my mind again after having not thought about it for years, I was struck by how silly that reaction was. Of course it makes sense that they'd been maintaining portability. There was nothing stupendous about this.

      I think this is the one time when I saw and felt the effects of Apple's legendary reality distortion field firsthand. (In every other instance, I hadn't been close enough and so only perceived it from afar and only had other sources to trust that it was a real phenomenon.)

    2. This sounds ridiculous, right? Why wasn’t there a CI/CD pipeline to build the system every few hours and publish the resulting image to a place where engineers could download it? Ha, ha. Things weren’t this rosy back then. Having automated tests, reproducible builds, a CI/CD system, unattended nightly builds, or even version control… these are all a pretty new “inventions”. Quality control had to be done by hand, as did integrating the various pieces that formed the system.

      This still describes the way the semiconductor manufacturing world works.

    3. Now, more than 30 years later, NT is behind almost all desktop and laptop computers in the world.

      This is sort of an odd remark. Even excluding servers and focusing only on traditional desktop and laptop computers, Windows' dominance is as weak now as it has been at any other point in the last 30 years.

      Most desktop and laptop computers? Yeah, probably.

      "Almost all"? Surely not.

    1. Reviewer #2 (Public Review):

      Summary:

      In this work, the authors present a new Python software package, Avian Vocalization Network (AVN) aimed at facilitating the analysis of birdsong, especially the song of the zebra finch, the most common songbird model in neuroscience. The package handles some of the most common (and some more advanced) song analyses, including segmentation, syllable classification, featurization of song, calculation of tutor-pupil similarity, and age prediction, with a view toward making the entire process friendlier to experimentalists working in the field.

      For many years, Sound Analysis Pro has served as a standard in the songbird field, the first package to extensively automate songbird analysis and facilitate the computation of acoustic features that have helped define the field. More recently, the increasing popularity of Python as a language, along with the emergence of new machine learning methods, has resulted in a number of new software tools, including the vocalpy ecosystem for audio processing, TweetyNet (for segmentation), t-SNE and UMAP (for visualization), and autoencoder-based approaches for embedding.

      Strengths:

      The AVN package overlaps several of these earlier efforts, albeit with a focus on more traditional featurization that many experimentalists may find more interpretable than deep learning-based approaches. Among the strengths of the paper are its clarity in explaining the several analyses it facilitates, along with high-quality experiments across multiple public datasets collected from different research groups. As a software package, it is open source, installable via the pip Python package manager, and features high-quality documentation, as well as tutorials. For experimentalists who wish to replicate any of the analyses from the paper, the package is likely to be a useful time saver.

      Weaknesses:

      I think the potential limitations of the work are predominantly on the software end, with one or two quibbles about the methods.

      First, the software: it's important to note that the package is trying to do many things, of which it is likely to do several well and few comprehensively. Rather than a package that presents a number of new analyses or a new analysis framework, it is more a codification of recipes, some of which are reimplementations of existing work (SAP features), some of which are essentially wrappers around other work (interfacing with WhisperSeg segmentations), and some of which are new (similarity scoring). All of this has value, but in my estimation, it has less value as part of a standalone package and potentially much more as part of an ecosystem like vocalpy that is undergoing continuous development and has long-term support. While the code is well-documented, including web-based documentation for both the core package and the GUI, the latter is available only on Windows, which might limit the scope of adoption.

      That is to say, whether AVN is adopted by the field in the medium term will have much more to do with the quality of its maintenance and responsiveness to users than any particular feature, but I believe that many of the analysis recipes that the authors have carefully worked out may find their way into other code and workflows.

      Second, two notes about new analysis approaches:

      (1) The authors propose a new means of measuring tutor-pupil similarity based on first learning a latent space of syllables via a self-supervised learning (SSL) scheme and then using the earth mover's distance (EMD) to calculate transport costs between the distributions of tutors' and pupils' syllables. While to my knowledge this exact method has not previously been proposed in birdsong, I suspect it is unlikely to differ substantially from the approach of autoencoding followed by MMD used in the Goffinet et al. paper. That is, SSL, like the autoencoder, is a latent space learning approach, and EMD, like MMD, is an integral probability metric that measures discrepancies between two distributions. (Indeed, the two are very closely related: https://stats.stackexchange.com/questions/400180/earth-movers-distance-and-maximum-mean-discrepency.) Without further experiments, it is hard to tell whether these two approaches differ meaningfully. Likewise, while the authors have trained on a large corpus of syllables to define their latent space in a way that generalizes to new birds, it is unclear why such an approach would not work with other latent space learning methods.

      (2) The authors propose a new method for maturity scoring by training a model (a generalized additive model) to predict the age of the bird based on a selected subset of acoustic features. This is distinct from the "predicted age" approach of Brudner, Pearson, and Mooney, which predicts based on a latent representation rather than specific features, and the GAM nicely segregates the contribution of each. As such, this approach may be preferred by many users who appreciate its interpretability.

      In summary, my view is that this is a nice paper detailing a well-executed piece of software whose future impact will be determined by the degree of support and maintenance it receives from others over the near and medium term.

    2. Author response:

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary: 

      This paper applies methods for segmentation, annotation, and visualization of acoustic analysis to zebra finch song. The paper shows that these methods can be used to predict the stage of song development and to quantify acoustic similarity. The methods are solid and are likely to provide a useful tool for scientists aiming to label large datasets of zebra finch vocalizations. The paper has two main parts: 1) establishing a pipeline/ package for analyzing zebra finch birdsong and 2) a method for measuring song imitation. 

      Strengths: 

      It is useful to see existing methods for syllable segmentation compared to new datasets. 

      It is useful, but not surprising, that these methods can be used to predict developmental stage, which is strongly associated with syllable temporal structure. 

      It is useful to confirm that these methods can identify abnormalities in deafened and isolated songs. 

      Weaknesses: 

      For the first part, the implementation seems to be a wrapper on existing techniques. For instance, the first section talks about syllable segmentation; they made a comparison between whisperseg (Gu et al, 2024), tweetynet (Cohen et al, 2022), and amplitude thresholding. They found that whisperseg performed the best, and they included it in the pipeline. They then used whisperseg to analyze syllable duration distributions and rhythm of birds of different ages and confirmed past findings on this developmental process (e.g. Aronov et al, 2011). Next, based on the segmentation, they assign labels by performing UMAP and HDBScan on the spectrogram (nothing new; that's what people have been doing). Then, based on the labels, they claimed they developed a 'new' visualization - syntax raster ( line 180 ). That was done by Sainburg et. al. 2020 in Figure 12E and also in Cohen et al, 2020 - so the claim to have developed 'a new song syntax visualization' is confusing. The rest of the paper is about analyzing the finch data based on AVN features (which are essentially acoustic features already in the classic literature). 

      First, we would like to thank this reviewer for their kind comments and feedback on this manuscript. It is true that many of the components of this song analysis pipeline are not entirely novel in isolation. Our real contribution here is bringing them together in a way that allows other researchers to seamlessly apply automated syllable segmentation, clustering, and downstream analyses to their data. That said, our approach to training TweetyNet for syllable segmentation is novel. We trained TweetyNet to recognize vocalizations vs. silence across multiple birds, such that it can generalize to new individual birds, whereas Tweetynet had only ever been used to annotate song syllables from birds included in its training set previously. Our validation of TweetyNet and WhisperSeg in combination with UMAP and HDBSCAN clustering is also novel, providing valuable information about how these systems interact, and how reliable the completely automatically generated labels are for downstream analysis. 

      Our syntax raster visualization does resemble Figure 12E in Sainburg et al. 2020, however it differs in a few important ways, which we believe warrant its consideration as a novel visualization method. First, Sainburg et al. represent the labels across bouts in real time; their position along the x axis reflects the time at which each syllable is produced relative to the start of the bout. By contrast, our visualization considers only the index of syllables within a bout (ie. First syllable vs. second syllable etc) without consideration of the true durations of each syllable or the silent gaps between them. This makes it much easier to detect syntax patterns across bouts, as the added variability of syllable timing is removed. Considering only the sequence of syllables rather than their timing also allows us to more easily align bouts according to the first syllable of a motif, further emphasizing the presence or absence of repeating syllable sequences without interference from the more variable introductory notes at the start of a motif. Finally, instead of plotting all bouts in the order in which they were produced, our visualization orders bouts such that bouts with the same sequence of syllables will be plotted together, which again serves to emphasize the most common syllable sequences that the bird produces. These additional processing steps mean that our syntax raster plot has much starker contrast between birds with stereotyped syntax and birds with more variable syntax, as compared to the more minimally processed visualization in Sainburg et al. 2020. There doesn’t appear to be any similar visualizations in Cohen et al. 2020. 

      The second part may be something new, but there are opportunities to improve the benchmarking. It is about the pupil-tutor imitation analysis. They introduce a convolutional neural network that takes triplets as an input (each tripled is essentially 3 images stacked together such that you have (anchor, positive, negative), Anchor is a reference spectrogram from, say finch A; positive means a different spectrogram with the same label as anchor from finch A, and negative means a spectrogram not related to A or different syllable label from A. The network is then trained to produce a low-dimensional embedding by ensuring the embedding distance between anchor and positive is less than anchor and negative by a certain margin. Based on the embedding, they then made use of earth mover distance to quantify the similarity in the syllable distribution among finches. They then compared their approach performance with that of sound analysis pro (SAP) and a variant of SAP. A more natural comparison, which they didn't include, is with the VAE approach by Goffinet et al. In this paper (https://doi.org/10.7554/eLife.67855, Fig 7), they also attempted to perform an analysis on the tutor pupil song. 

      We thank the reviewer for this suggestion, and plan to include a comparison of the triplet loss embedding space to the VAE space for song similarity comparisons in the revised manuscript.

      Reviewer #2 (Public Review):

      Summary: 

      In this work, the authors present a new Python software package, Avian Vocalization Network (AVN) aimed at facilitating the analysis of birdsong, especially the song of the zebra finch, the most common songbird model in neuroscience. The package handles some of the most common (and some more advanced) song analyses, including segmentation, syllable classification, featurization of song, calculation of tutor-pupil similarity, and age prediction, with a view toward making the entire process friendlier to experimentalists working in the field. 

      For many years, Sound Analysis Pro has served as a standard in the songbird field, the first package to extensively automate songbird analysis and facilitate the computation of acoustic features that have helped define the field. More recently, the increasing popularity of Python as a language, along with the emergence of new machine learning methods, has resulted in a number of new software tools, including the vocalpy ecosystem for audio processing, TweetyNet (for segmentation), t-SNE and UMAP (for visualization), and autoencoder-based approaches for embedding. 

      Strengths: 

      The AVN package overlaps several of these earlier efforts, albeit with a focus on more traditional featurization that many experimentalists may find more interpretable than deep learning-based approaches. Among the strengths of the paper are its clarity in explaining the several analyses it facilitates, along with high-quality experiments across multiple public datasets collected from different research groups. As a software package, it is open source, installable via the pip Python package manager, and features high-quality documentation, as well as tutorials. For experimentalists who wish to replicate any of the analyses from the paper, the package is likely to be a useful time saver. 

      Weaknesses: 

      I think the potential limitations of the work are predominantly on the software end, with one or two quibbles about the methods. 

      First, the software: it's important to note that the package is trying to do many things, of which it is likely to do several well and few comprehensively. Rather than a package that presents a number of new analyses or a new analysis framework, it is more a codification of recipes, some of which are reimplementations of existing work (SAP features), some of which are essentially wrappers around other work (interfacing with WhisperSeg segmentations), and some of which are new (similarity scoring). All of this has value, but in my estimation, it has less value as part of a standalone package and potentially much more as part of an ecosystem like vocalpy that is undergoing continuous development and has long-term support. 

      We appreciate this reviewer’s comments and concerns about the structure of the AVN package and its long-term maintenance. We have considered incorporating AVN into the VocalPy ecosystem but have chosen not to for a few key reasons. (1) AVN was designed with ease of use for experimenters with limited coding experience top of mind. VocalPy provides excellent resources for researchers with some familiarity with object-oriented programming to manage and analyze their datasets; however, we believe it may be challenging for users without such experience to adopt VocalPy quickly. AVN’s ‘recipe’ approach, as you put it, is very easily accessible to new users, and allows users with intermediate coding experience to easily navigate the source code to gain a deeper understanding of the methodology. AVN also consistently outputs processed data in familiar formats (tables in .csv files which can be opened in excel), in an effort to make it more accessible to new users, something which would be challenging to reconcile with VocalPy’s emphasis on their `dataset`classes. (2) AVN and VocalPy differ in their underlying goals and philosophies when it comes to flexibility vs. standardization of analysis pipelines. VocalPy is designed to facilitate mixing-and-matching of different spectrogram generation, segmentation, annotation etc. approaches, so that researchers can design and implement their own custom analysis pipelines. This flexibility is useful in many cases. For instance, it could allow researchers who have very different noise filtering and annotation needs, like those working with field recordings versus acoustic chamber recordings, analyze their data using this platform. However, when it comes to comparisons across zebra finch research labs, this flexibility comes at the expense of direct comparison and integration of song features across research groups. This is the context in which AVN is most useful. It presents a single approach to song segmentation, labeling, and featurization that has been shown to generalize well across research groups, and which allows direct comparisons of the resulting features. AVN’s single, extensively validated, standard pipeline approach is fundamentally incompatible with VocalPy’s emphasis on flexibility. We are excited to see how VocalPy continues to evolve in the future and recognize the value that both AVN and VocalPy bring to the songbird research community, each with their own distinct strengths, weaknesses, and ideal use cases. 

      While the code is well-documented, including web-based documentation for both the core package and the GUI, the latter is available only on Windows, which might limit the scope of adoption. 

      We thank the reviewer for their kind words about AVN’s documentation. We recognize that the GUI’s exclusive availability on Windows is a limitation, and we would be happy to collaborate with other researchers and developers in the future to build a Mac compatible version, should the demand present itself. That said, the python package works on all operating systems, so non-Windows users still have the ability to use AVN that way.  

      That is to say, whether AVN is adopted by the field in the medium term will have much more to do with the quality of its maintenance and responsiveness to users than any particular feature, but I believe that many of the analysis recipes that the authors have carefully worked out may find their way into other code and workflows. 

      Second, two notes about new analysis approaches: 

      (1) The authors propose a new means of measuring tutor-pupil similarity based on first learning a latent space of syllables via a self-supervised learning (SSL) scheme and then using the earth mover's distance (EMD) to calculate transport costs between the distributions of tutors' and pupils' syllables. While to my knowledge this exact method has not previously been proposed in birdsong, I suspect it is unlikely to differ substantially from the approach of autoencoding followed by MMD used in the Goffinet et al. paper. That is, SSL, like the autoencoder, is a latent space learning approach, and EMD, like MMD, is an integral probability metric that measures discrepancies between two distributions.

      (Indeed, the two are very closely related: https://stats.stackexchange.com/questions/400180/earth-movers-distance-andmaximum-mean-discrepency.) Without further experiments, it is hard to tell whether these two approaches differ meaningfully. Likewise, while the authors have trained on a large corpus of syllables to define their latent space in a way that generalizes to new birds, it is unclear why such an approach would not work with other latent space learning methods. 

      We recognize the similarities between these approaches, and plan to include a comparison of triplet loss embeddings compared with MMD and VAE embeddings compared with MMD and EMD in the revised manuscript. Thank you for this suggestion.  

      (2) The authors propose a new method for maturity scoring by training a model (a generalized additive model) to predict the age of the bird based on a selected subset of acoustic features. This is distinct from the "predicted age" approach of Brudner, Pearson, and Mooney, which predicts based on a latent representation rather than specific features, and the GAM nicely segregates the contribution of each. As such, this approach may be preferred by many users who appreciate its interpretability. 

      In summary, my view is that this is a nice paper detailing a well-executed piece of software whose future impact will be determined by the degree of support and maintenance it receives from others over the near and medium term. 

      Reviewer #3 (Public Review):

      Summary: 

      The authors invent song and syllable discrimination tasks they use to train deep networks. These networks they then use as a basis for routine song analysis and song evaluation tasks. For the analysis, they consider both data from their own colony and from another colony the network has not seen during training. They validate the analysis scores of the network against expert human annotators, achieving a correlation of 80-90%. 

      Strengths: 

      (1) Robust Validation and Generalizability: The authors demonstrate a good performance of the AVN across various datasets, including individuals exhibiting deviant behavior. This extensive validation underscores the system's usefulness and broad applicability to zebra finch song analysis, establishing it as a potentially valuable tool for researchers in the field. 

      (2) Comprehensive and Standardized Feature Analysis: AVN integrates a comprehensive set of interpretable features commonly used in the study of bird songs. By standardizing the feature extraction method, the AVN facilitates comparative research, allowing for consistent interpretation and comparison of vocal behavior across studies. 

      (3) Automation and Ease of Use. By being fully automated, the method is straightforward to apply and should introduce barely an adoption threshold to other labs. 

      (4) Human experts were recruited to perform extensive annotations (of vocal segments and of song similarity scores). These annotations released as public datasets are potentially very valuable. 

      Weaknesses: 

      (1) Poorly motivated tasks. The approach is poorly motivated and many assumptions come across as arbitrary. For example, the authors implicitly assume that the task of birdsong comparison is best achieved by a system that optimally discriminates between typical, deaf, and isolated songs. Similarly, the authors assume that song development is best tracked using a system that optimally estimates the age of a bird given its song. My issue is that these are fake tasks since clearly, researchers will know whether a bird is an isolated or a deaf bird, and they will also know the age of a bird, so no machine learning is needed to solve these tasks. Yet, the authors imagine that solving these placeholder tasks will somehow help with measuring important aspects of vocal behavior. 

      We appreciate this reviewer’s concerns and apologize for not providing sufficiently clear rationale for the inclusion of our phenotype classifier and age regression models in the original manuscript. These tasks are not intended to be taken as a final, ultimate culmination of the AVN pipeline. Rather, we consider the carefully engineered 55-interpretable feature set to be AVN’s final output, and these analyses serve merely as examples of how that feature set can be applied. That said, each of these models do have valid experimental use cases that we believe are important and would like to bring to the attention of the reviewer.

      For one, we showed how the LDA model that can discriminate between typical, deaf, and isolate birds’ songs not only allows us to evaluate which features are most important for discriminating between these groups, but also allows comparison of the FoxP1 knock-down (FP1 KD) birds to each of these phenotypes. Based on previous work (Garcia-Oscos et al. 2021), we hypothesized that FP1 KD in these birds specifically impaired tutor song memory formation while sparing a bird’s ability to refine their own vocalizations through auditory feedback. Thus, we would expect their songs to resemble those of isolate birds, who lack a tutor song memory, but not to resemble deaf birds who lack a tutor song memory and auditory feedback of their own vocalizations to guide learning. The LDA model allowed us to make this comparison quantitatively for the first time and confirm our hypothesis that FP1 KD birds’ songs are indeed most like isolates’. In the future, as more research groups publish their birds’ AVN feature sets, we hope to be able to make even more fine-grained comparisons between different groups of birds, either using LDA or other similar interpretable classifiers. 

      The age prediction model also has valid real-world use cases. For instance, one might imagine an experimental manipulation that is hypothesized to accelerate or slow song maturation in juvenile birds. This age prediction model could be applied to the AVN feature sets of birds having undergone such a manipulation to determine whether their predicted ages systematically lead or lag their true biological ages, and which song features are most responsible for this difference. We didn’t have access to data for any such birds for inclusion in this paper, but we hope that others in the future will be able to take inspiration from our methodology and use this or a similar age regression model with AVN features in their research. We will revise the original manuscript to make this clearer. 

      Along similar lines, authors assume that a good measure of similarity is one that optimally performs repeated syllable detection (i.e. to discriminate same syllable pairs from different pairs). The authors need to explain why they think these placeholder tasks are good and why no better task can be defined that more closely captures what researchers want to measure. Note: the standard tasks for self-supervised learning are next word or masked word prediction, why are these not used here? 

      There appears to be some misunderstanding regarding our similarity scoring embedding model and our rationale for using it. We will explain it in more depth here and provide some additional explanation in the manuscript. First, we are not training a model to discriminate between same and different syllable pairs. The triplet loss network is trained to embed syllables in an 8-dimensional space such that syllables with the same label are closer together than syllables with different labels. The loss function is related to the relative distance between embeddings of syllables with the same or different labels, not the classification of syllables as same or different. This approach was chosen because it has repeatedly been shown to be a useful data compression step (Schorff et al. 2015, Thakur et al. 2019) before further downstream tasks are applied on its output, particularly in contexts where there is little data per class (syllable label). For example, Schorff et al. 2015 trained a deep convolutional neural network with triplet loss to embed images of human faces from the same individual closer together than images of different individuals in a 128-dimensional space. They then used this model to compute 128-dimensional representations of additional face images, not included in training, which were used for individual facial recognition (this is a same vs. different category classifier), and facial clustering, achieving better performance than the previous state of the art. The triplet loss function results in a model that can generate useful embeddings of previously unseen categories, like new individuals’ faces, or new zebra finches’ syllables, which can then be used in downstream analyses. This meaningful, lower dimensional space allows comparisons of distributions of syllables across birds, as in Brainard and Mets 2008, and Goffinet et al. 2021. 

      Next word and masked word prediction are indeed common self-supervised learning tasks for models working with text data, or other data with meaningful sequential organization. That is not the case for our zebra finch syllables, where every bird’s syllable sequence depends only on its tutor’s sequence, and there is no evidence for strong universal syllable sequencing rules (James et al. 2020). Rather, our embedding model is an example of a computer vision task, as it deals with sets of twodimensional images (spectrograms), not sequences of categorical variables (like text). It is also not, strictly speaking, a self-supervised learning task, as it does require syllable labels to generate the triplets. A common self-supervised approach for dimensionality reduction in a computer vision task such as this one would be to train an autoencoder to compress images to a lower dimensional space, then faithfully reconstruct them from the compressed representation.  This has been done using a variational autoencoder trained on zebra finch syllables in Goffinet et al. 2021. In keeping with the suggestions from reviewers #1 and #2, we plan to include a comparison of our triplet loss model with the Goffinet et al. VAE approach in the revised manuscript.  

      (2) The machine learning methodology lacks rigor. The aims of the machine learning pipeline are extremely vague and keep changing like a moving target. Mainly, the deep networks are trained on some tasks but then authors evaluate their performance on different, disconnected tasks. For example, they train both the birdsong comparison method (L263+) and the song similarity method (L318+) on classification tasks. However, they evaluate the former method (LDA) on classification accuracy, but the latter (8-dim embeddings) using a contrast index. In machine learning, usually, a useful task is first defined, then the system is trained on it and then tested on a held-out dataset. If the sensitivity index is important, why does it not serve as a cost function for training?

      Again, there appears to be some misunderstanding of our similarity scoring methodology. Our similarity scoring model is not trained on a classification task, but rather on an embedding task. It learns to embed spectrograms of syllables in an 8dimensional space such that syllables with the same label are closer together than syllables with different labels. We could report the loss values for this embedding task on our training and validation datasets, but these wouldn’t have any clear relevance to the downstream task of syllable distribution comparison where we are using the model’s embeddings. We report the contrast index as this has direct relevance to the actual application of the model and allows comparisons to other similarity scoring methods, something that the triplet loss values wouldn’t allow. 

      The triplet loss method was chosen because it has been shown to yield useful lowdimensional representations of data, even in cases where there is limited labeled training data (Thakur et al. 2019). While we have one of the largest manually annotated datasets of zebra finch songs, it is still quite small by industry deep learning standards, which is why we chose a method that would perform well given the size of our dataset. Training a model on a contrast index directly would be extremely computationally intensive and require many more pairs of birds with known relationships than we currently have access to. It could be an interesting approach to take in the future, but one that would be unlikely to perform well with a dataset size typical to songbird research. 

      Also, usually, in solid machine learning work, diverse methods are compared against each other to identify their relative strengths. The paper contains almost none of this, e.g. authors examined only one clustering method (HDBSCAN). 

      We did compare multiple methods for syllable segmentation (WhisperSeg,  TweetyNet, and Amplitude thresholding) as this hadn’t been done previously. We chose not to perform extensive comparison of different clustering methods as Sainburg et al. 2020 already did so and we felt no need to reduplicate this effort. We encourage this reviewer to refer to Sainburg et al.’s excellent work for comparisons of multiple clustering methods applied to zebra finch song syllables.  

      (3) Performance issues. The authors want to 'simplify large-scale behavioral analysis' but it seems they want to do that at a high cost. (Gu et al 2023) achieved syllable scores above 0.99 for adults, which is much larger than the average score of 0.88 achieved here (L121). Similarly, the syllable scores in (Cohen et al 2022) are above 94% (their error rates are below 6%, albeit in Bengalese finches, not zebra finches), which is also better than here. Why is the performance of AVN so low? The low scores of AVN argue in favor of some human labeling and training on each bird. 

      Firstly, the syllable error rate scores reported in Cohen et al. 2022 are calculated very differently than the F1 scores we report here and are based on a model trained with data from the same bird as was used in testing, unlike our more general segmentation approach where the model was tested on different birds than were used in testing. Thus, the scores reported in Cohen et al. and the F1 scores that we report cannot be compared. 

      The discrepancy between the F1seg scores reported in Gu et al. 2023 and the segmentation F1 scores that we report are likely due to differences in the underlying datasets. Our UTSW recordings tend to have higher levels of both stationary and nonstationary background noise, which make segmentation more challenging. The recordings from Rockefeller were less contaminated by background noise, and they resulted in slightly higher F1 scores. That said, we believe that the primary factor accounting for this difference in scores with Gu et al. 2023 is the granularity of our ‘ground truth’ syllable segments. In our case, if there was ever any ambiguity as to whether vocal elements should be segmented into two short syllables with a very short gap between them or merged into a single longer syllable, we chose to split them. WhisperSeg had a strong tendency to merge the vocal elements in ambiguous cases such as these. This results in a higher rate of false negative syllable onset detections, reflected in the low recall scores achieved by WhisperSeg (see supplemental figure 2b), but still very high precision scores (supplemental figure 2a). While WhisperSeg did frequently merge these syllables in a way that differed from our ground truth segmentation, it did so consistently, meaning it had little impact on downstream measures of syntax entropy (Fig 3c) or syllable duration entropy (supplemental figure 7a). It is for that reason that, despite a lower F1 score, we still consider AVN’s automatically generated annotations to be sufficiently accurate for downstream analyses. 

      Should researchers require a higher degree of accuracy and precision with their annotations (for example, to detect very subtle changes in song before and after an acute manipulation) and be willing to dedicate the time and resources to manually labeling a subset of recordings from each of their birds, we suggest they turn toward one of the existing tools for supervised song annotation, such as TweetyNet.  

      (4) Texas bias. It is true that comparability across datasets is enhanced when everyone uses the same code. However, the authors' proposal essentially is to replace the bias between labs with a bias towards birds in Texas. The comparison with Rockefeller birds is nice, but it amounts to merely N=1. If birds in Japanese or European labs have evolved different song repertoires, the AVN might not capture the associated song features in these labs well. 

      We appreciate the reviewer’s concern about a bias toward birds from the UTSW colony. However, this paper shows that despite training (for the similarity scoring) and hyperparameter fitting (for the HDBSCAN clustering) on the UTSW birds, AVN performs as well if not better on birds from Rockefeller than from UTSW. To our knowledge, there are no publicly available datasets of annotated zebra finch songs from labs in Europe or in Asia but we would be happy to validate AVN on such datasets, should they become available. Furthermore, there is no evidence to suggest that there is dramatic drift in zebra finch vocal repertoire between continents which would necessitate such additional validation. While we didn’t have manual annotations for this dataset (which would allow validation of our segmentation and labeling methods), we did apply AVN to recordings share with us by the Wada lab in Japan, where visual inspection of the resulting annotations suggested comparable accuracy to the UTSW and Rockefeller datasets.  

      (5) The paper lacks an analysis of the balance between labor requirement, generalizability, and optimal performance. For tasks such as segmentation and labeling, fine-tuning for each new dataset could potentially enhance the model's accuracy and performance without compromising comparability. E.g. How many hours does it take to annotate hundred song motifs? How much would the performance of AVN increase if the network were to be retrained on these? The paper should be written in more neutral terms, letting researchers reach their own conclusions about how much manual labor they want to put into their data. 

      With standardization and ease of use in mind, we designed AVN specifically to perform fully automated syllable annotation and downstream feature calculations. We believe that we have demonstrated in this manuscript that our fully automated approach is sufficiently reliable for downstream analyses across multiple zebra finch colonies. That said, if researchers require an even higher degree of annotation precision and accuracy, they can turn toward one of the existing methods for supervised song annotation, such as TweetyNet. Incorporating human annotations for each bird processed by AVN is likely to improve its performance, but this would require significant changes to AVN’s methodology and is outside the scope of our current efforts.  

      (6) Full automation may not be everyone's wish. For example, given the highly stereotyped zebra finch songs, it is conceivable that some syllables are consistently mis-segmented or misclassified. Researchers may want to be able to correct such errors, which essentially amounts to fine-tuning AVN. Conceivably, researchers may want to retrain a network like the AVN on their own birds, to obtain a more fine-grained discriminative method. 

      Other methods exist for supervised or human-in-the-loop annotation of zebra finch songs, such as TweetyNet and DAN (Alam et al. 2023). We invite researchers who require a higher degree of accuracy than AVN can provide to explore these alternative approaches for song annotation. Incorporating human annotations for each individual bird being analyzed using AVN was never the goal of our pipeline, would require significant changes to AVN’s design, and is outside the scope of this manuscript.  

      (7) The analysis is restricted to song syllables and fails to include calls. No rationale is given for the omission of calls. Also, it is not clear how the analysis deals with repeated syllables in a motif, whether they are treated as two-syllable types or one. 

      It is true that we don’t currently have any dedicated features to describe calls. This could be a useful addition to AVN in the future. 

      What a human expert inspecting a spectrogram would typically call ‘repeated syllables’ in a bout are almost always assigned the same syllable label by the UMAP+HDBSCAN clustering. The syntax analysis module includes features examining the rate of syllable repetitions across syllable types. See https://avn.readthedocs.io/en/latest/syntax_analysis_demo.html#SyllableRepetitions

      (8) It seems not all human annotations have been released and the instruction sets given to experts (how to segment syllables and score songs) are not disclosed. It may well be that the differences in performance between (Gu et al 2023) and (Cohen et al 2022) are due to differences in segmentation tasks, which is why these tasks given to experts need to be clearly spelled out. Also, the downloadable files contain merely labels but no identifier of the expert. The data should be released in such a way that lets other labs adopt their labeling method and cross-check their own labeling accuracy. 

      All human annotations used in this manuscript have indeed been released as part of the accompanying dataset. Syllable annotations are not provided for all pupils and tutors used to validate the similarity scoring, as annotations are not necessary for similarity comparisons. We will expand our description of our annotation guidelines in the methods section of the revised manuscript. All the annotations were generated by one of two annotators. The second annotator always consulted with the first annotator in cases of ambiguous syllable segmentation or labeling, to ensure that they had consistent annotation styles. Unfortunately, we haven’t retained records about which birds were annotated by which of the two annotators, so we cannot share this information along with the dataset. The data is currently available in a format that should allow other research groups to use our annotations either to train their own annotation systems or check the performance of their existing systems on our annotations.  

      (9) The failure modes are not described. What segmentation errors did they encounter, and what syllable classification errors? It is important to describe the errors to be expected when using the method. 

      As we discussed in our response to this reviewer’s point (3), WhisperSeg has a tendency to merge syllables when the gap between them is very short, which explains its lower recall score compared to its precision on our dataset (supplementary figure 2). In rare cases, WhisperSeg also fails to recognize syllables entirely, again impacting its precision score. TweetyNet hardly ever completely ignores syllables, but it does tend to occasionally merge syllables together or over-segment them. Whereas WhisperSeg does this very consistently for the same syllable types within the same bird, TweetyNet merges or splits syllables more inconsistently. This inconsistent merging and splitting has a larger effect on syllable labeling, as manifested in the lower clustering v-measure scores we obtain with TweetyNet compared to WhisperSeg segmentations. TweetyNet also has much lower precision than WhisperSeg, largely because TweetyNet often recognizes background noises (like wing flaps or hopping) as syllables whereas WhisperSeg hardly ever segments nonvocal sounds. 

      Many errors in syllable labeling stem from differences in syllable segmentation. For example, if two syllables with labels ‘a’ and ‘b’ in the manual annotation are sometimes segmented as two syllables, but sometimes merged into a single syllable, the clustering is likely to find 3 different syllable types; one corresponding to ‘a’, one corresponding to ‘b’ and one corresponding to ‘ab’ merged. Because of how we align syllables across segmentation schemes for the v-measure calculation, this will look like syllable ‘b’ always has a consistent cluster label, but syllable ‘a’ can carry two different cluster labels, depending on the segmentation. In certain cases, even in the absence of segmentation errors, a group of syllables bearing the same manual annotation label may be split into 2 or 3 clusters (it is extremely rare for a single manual annotation group to be split into more than 3 clusters). In these cases, it is difficult to conclusively say whether the clustering represents an error, or if it actually captured some meaningful systematic difference between syllables that was missed by the annotator. Finally, sometimes rare syllable types with their own distinct labels in the manual annotation are merged into a single cluster. Most labeling errors can be explained by this kind of merging or splitting of groups relative to the manual annotation, not to occasional mis-classifications of one manual label type as another. 

      For examples of these types of errors, we encourage this reviewer and readers to refer to the example confusion matrices in figure 2f and supplemental figure 4b&e. We will also expand our discussion of these different types of errors in the revised manuscript. 

      (10) Usage of Different Dimensionality Reduction Methods: The pipeline uses two different dimensionality reduction techniques for labeling and similarity comparison - both based on the understanding of the distribution of data in lower-dimensional spaces. However, the reasons for choosing different methods for different tasks are not articulated, nor is there a comparison of their efficacy. 

      We apologize for not making this distinction sufficiently clear in the manuscript and will add additional explanation to the main text to make the reasoning more apparent. We chose to use UMAP for syllable labeling because it is a common embedding methodology to precede hierarchical clustering and has been shown to result in reliable syllable labels for birdsong in the past (Sainburg et al. 2020). However, it is not appropriate for similarity scoring, because comparing EMD scores between birds requires that all the birds’ syllable distributions exist within the same shared embedding space. This can be achieved by using the same triplet loss-trained neural network model to embed syllables from all birds. This cannot be achieved with UMAP because all birds whose scores are being compared would need to be embedded in the same UMAP space, as distances between points cannot be compared across UMAPs. In practice, this would mean that every time a new tutor-pupil pair needs to be scored, their syllables would need to be added to a matrix with all previously compared birds’ syllables, a new UMAP would need to be computed, and new EMD scores between all bird pairs would need to be calculated using their new UMAP embeddings. This is very computationally expensive and quickly becomes unfeasible without dedicated high power computing infrastructure. It also means that similarity scores couldn’t be compared across papers without recomputing everything each time, whereas EMD scores obtained with triplet loss embeddings can be compared, provided they use the same trained model (which we provide as part of AVN) to embed their syllables in a common latent space.  

      (11) Reproducibility: are the measurements reproducible? Systems like UMAP always find a new embedding given some fixed input, so the output tends to fluctuate. 

      There is indeed a stochastic element to UMAP embeddings which will result in different embeddings and therefore different syllable labels across repeated runs with the same input. Anecdotally, we observed that v-measures scores were quite consistent within birds across repeated runs of the UMAP, but we will add an additional supplementary figure to the revised manuscript showing this.

    1. nside was a SamSpade labyrinth of bad linoleum, but the offices had giant windows and a swankmailing address (“World Trade Center”), and my father had what used to be thethrill of urban real estate:

      Main Idea of Section: Heller's father highlighted San Francisco's changing urban scene by renting reasonably priced office space at the Ferry Building, which reflected the post-earthquake transition as underutilized places became sought-after real estate.

    Annotators

    1. Reviewer #3 (Public review):

      Summary:

      The authors aimed to investigate the multifaceted roles of the Inferior Colliculus (IC) in auditory and cognitive processes in monkeys. Through extracellular recordings during a sound duration-based novelty detection task, the authors observed a "climbing effect" in neuronal firing rates, suggesting an enhanced response during sensory prediction. Observations of reward prediction errors within the IC further highlight its complex integration in both auditory and reward processing. Additionally, the study indicated IC neuronal activities could be involved in decision-making processes.

      Strengths:

      This study has the potential to significantly impact the field by challenging the traditional view of the IC as merely an auditory relay station and proposing a more integrative role in cognitive processing. The results provide valuable insights into the complex roles of the IC, particularly in sensory and cognitive integration, and could inspire further research into the cognitive functions of the IC.

      Weaknesses:

      Major Comments:

      (1) Structural Clarity and Logic Flow:<br /> The manuscript investigates three intriguing functions of IC neurons: sensory prediction, reward prediction, and cognitive decision-making, each of which is a compelling topic. However, the logical flow of the manuscript is not clearly presented and needs to be well recognized. For instance, Figure 3 should be merged into Figure 2 to present population responses to the order of sounds, thereby focusing on sensory prediction. Given the current arrangement of results and figures, the title could be more aptly phrased as "Beyond Auditory Relay: Dissecting the Inferior Colliculus's Role in Sensory Prediction, Reward Prediction, and Cognitive Decision-Making."

      (2) Clarification of Data Analysis:<br /> Key information regarding data analysis is dispersed throughout the results section, which can lead to confusion. Providing a more detailed and cohesive explanation of the experimental design would significantly enhance the interpretation of the findings. For instance, including a detailed timeline and reward information for the behavioral paradigms shown in Figures 1C and D would offer crucial context for the study. More importantly, clearly presenting the analysis temporal windows and providing comprehensive statistical analysis details would greatly improve reader comprehension.

      (3) Reward Prediction Analysis:<br /> The conclusion regarding the IC's role in reward prediction is underdeveloped. While the manuscript presents evidence that IC neurons can encode reward prediction, this is only demonstrated with two example neurons in Figure 6. A more comprehensive analysis of the relationship between IC neuronal activity and reward prediction is necessary. Providing population-level data would significantly strengthen the findings concerning the IC's complex functionalities. Additionally, the discussion of reward prediction in lines 437-445, which describes IC neuron responses in control experiments, does not sufficiently demonstrate that IC neurons can encode reward expectations. It would be valuable to include the responses of IC neurons during trials with incorrect key presses or no key presses to better illustrate this point.

    1. AbstractBackground Next-generation sequencing for assaying microbial communities has become a standard technique in recent years. However, the initial investment required into in-silico analytics is still quite significant, especially for facilities not focused on bioinformatics. With the rapid decline in costs and growing adoption of sequencing-based methods in a number of fields, validated, fully automated, reproducible and yet flexible pipelines will play a greater role in various scientific fields in the future.Results We present RiboSnake, a validated, automated, reproducible QIIME2-based analysis pipeline implemented in Snakemake for the computational analysis of 16S rRNA gene amplicon sequencing data. The pipeline comes with pre-packaged validated parameter sets, optimized for different sample types. The sets range from complex environmental samples to patient data. The configuration packages can be easily adapted and shared, requiring minimal user input.Conclusion RiboSnake is a new alternative for researchers employing 16S rRNA gene amplicon sequencing and looking for a customizable and yet user-friendly pipeline for microbiome analysis with in-vitro validated settings. The complete analysis generated with a fully automated pipeline based on validated parameter sets for different sample types is a significant improvement to existing methods. The workflow repository can be found on GitHub (https://github.com/IKIM-Essen/RiboSnake).

      This work has been published in GigaByte Journal under a CC-BY 4.0 license (https://doi.org/10.46471/gigabyte.132), and has published the reviews under the same license. These are as follows.

      Reviewer 1. Michael Hall

      Is installation/deployment sufficiently outlined in the paper and documentation, and does it proceed as outlined?

      Unable to test. The README states "If you want to test the RiboSnake functions yourself, you can use the same data used for the CI/CD tests." A worked example of how I can do this would be appreciated so I can test the workflow.

      Is there enough clear information in the documentation to install, run and test this tool, including information on where to seek help if required?

      The Usage instructions say to create a new repository using ribosnake as a template, but ribosnake is not a template repository (see https://docs.github.com/en/repositories/creating-and-managing-repositories/creating-a-repository-from-a-template). The README states "If you want to test the RiboSnake functions yourself, you can use the same data used for the CI/CD tests." A worked example of how I can do this would be appreciated so I can test the workflow.

      Have any claims of performance been sufficiently tested and compared to other commonly-used packages?

      Not applicable.

      Is automated testing used or are there manual steps described so that the functionality of the software can be verified?

      Yes, though as mentioned above, the README states "If you want to test the RiboSnake functions yourself, you can use the same data used for the CI/CD tests." A worked example of how I can do this would be appreciated so I can test the workflow.

      Additional Comments:

      The Introduction could be make far more concise, there's a lot of repetition.

      The installation command in figure 1 is three commands, not two as stated in the text (third-last paragraph Introduction), and is slightly misleading from an installation point of view as it assumes conda and snakemake are installed. Though it is mentioned later in the text (p5) that snakemake and conda require manual installation.

      The in-text citation for Greengenes2 is just [?] - maybe a latex issue?

      The last paragraph of the 'Features and Implementations' section was mostly already stated earlier in the manuscript.

      Make the colouring consistent between fig 2a-c and 2d as well as the vertical ordering to make for easier comparison. For example, in figures 2a-c Enterococcus (grey) is on the bottom, whereas in fig 2d it is red and in the middle. Colour legends should also be added to Figures 3-5 to match Fig 2.

      A small table should be added showing the comparison of RiboSnake and the original publication for the top 10 most abundant phyla for the Atacama soil dataset and their abundances (see last paragraph of 'Usage and Findings'.

      Reviewer 2. Yong-Xin Liu and Salsabeel Yousuf

      The manuscript presented by the authors describes a comprehensive study on the “RiboSnake pipeline” for 16S rRNA gene microbiome analysis, which is a user-friendly, robust, and multipurpose. RiboSnake, a validated, automated, reproducible QIIME2-based analysis pipeline implemented in Snakemake, offers parallel processing for efficient analysis of large datasets in both environmental and medical research contexts. Further demonstrating its effectiveness, this pipeline effectively analyzes human-associated microbiomes and environmental samples like wastewater and soil, thus expanding the scope of analysis for 16S rRNA data. The overall computational pipeline is useful and results are sound, validated through rigorous testing on MOCK communities and real-world datasets. However, there are some issues for improvement in the manuscript.

      Major comments: 1. In the clinical data section the author mentions rectal swabs were used from a published study [31]. While the source is referenced, it would be helpful to know if any information was provided in the referenced study regarding the collection methods or storage conditions for the rectal swabs. 2. The text mentions using cotton swabs pre-moistened with TE buffer + 0.5% Tween 20. While cotton swabs are common, are there any considerations for using different swab materials depending on the target analytes or sampling surface (e.g., flocked swabs for better epithelial cell collection)? 3. Does RiboSnake require user intervention during any steps, or is it fully automated? 4. The author mentions that contamination filtering parameters should be adjusted based on the sample type. How can users determine the appropriate filtering parameters for their specific samples? Are there guidelines for users to know how much adjustment is needed for specific scenarios? 5. The default abundance threshold for filtering low-frequency reads is chosen based on Nearing et al. [44]. Please discuss the rationale behind using a single threshold for all sample types? Would it be beneficial to allow users to define this threshold based on their data characteristics? 6. Would you like to explain the limitation of RiboSnake, such as specific types of samples it may not be suitable for or potential biases introduced by certain functionalities? 7. The manuscript mentions various visualization tools used throughout the pipeline (QIIME2, qurro). Please clarify which types of data are visualized with each tool, and how users can access or customize these visualizations? 8. To strengthen the manuscript's impact, consider discussing the specific novelty of RiboSnake compared to existing 16S rRNA gene microbiome analysis pipelines. Would you be able to elaborate on the unique features or functionalities of RiboSnake that address limitations of current methods? 9. EasyAmplicon is recently published pipeline and easy using in windows, mac and linux system,

      Minor comments: 1. Reference is missing in this sentence. “The default is the SILVA database [47]. Greengenes2 [? ] can be used alternatively”. 2. The author should careful about the lowercase and upper case throughout the manuscript. Please check the following for references:  ..the 2017 published Atacama Soil data set with samples taken fromthe Atacama desert was used [32] as well as samples collected fromsoil under switchgrass published in [33].  based on an Euclidean beta diversitymetric, shows that the positive controls, as well as the samples taken from subjects 1 and 3 (S1 and S3), cluster together.  A wide range of diversity analysis parameters are available in QIIME2 and its associated tools. These include the Shannon diversity index to measure richness, the Pielou index tomeasure evenness, or perform standard correlation analysis using Pearson or Spearman indices, among others. 3. In the introduction part this sentences “However, while these methods enable 16S rRNA analysis with minimal user interaction…” needs attention for clarity. Consider separating it into two sentences to emphasize the limitations of existing pipelines compared to the described methods’. Alternatively, using contrasting words like "in contrast" could highlight these differences. 4. More detail in attached PDF.

      https://gigabyte-review.rivervalleytechnologies.comdownload-api-file?ZmlsZV9wYXRoPXVwbG9hZHMvZ3gvVFIvNTM5L2d4LVRSLTE3MTY5Nzk4MTktcmV2aXNlZC5wZGY=

      Re-review: The author's response has been fully addressed my concerns. The quality of the paper has apparently improved. I agree with the publication of this article.

    1. Reviewer #2 (Public review):

      Deciphering the metabolic alterations characterizing the prediabetes-diabetes spectrum could provide early time windows for targeted preventive measures to extend precision medicine while avoiding disproportionate healthcare costs. The authors identified a panel of 9 circulating metabolites combined with basic clinical variables that significantly improved the prediction from prediabetes to diabetes. These findings provided insights into the integration of these metabolites into clinical and public health practice.

      Comments on the revised version:

      Congratulations to the authors. I have no more comments.

    2. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Review):

      Using the UK Biobank, this study assessed the value of nuclear magnetic resonance measured metabolites as predictors of progression to diabetes. The authors identified a panel of 9 circulating metabolites that improved the ability in risk prediction of progression from prediabetes to diabetes. In general, this is a well-performed study, and the findings may provide a new approach to identifying those at high risk of developing diabetes. I have some comments that may improve the importance of this study.

      We deeply appreciate the reviewer's invaluable time dedicated to the review of this manuscript and the insightful comments to enhance its overall quality.

      (1) It is unclear why the authors only considered the top 20 variables in the metabolite selection and why they did not set a wider threshold.

      Thank you for the comment. We set the top 20 variables in the metabolite selection balancing the performance of the final diabetes risk prediction model and the clinical applicability due to measurement costs. We have added this explanation in the “Methods” section.

      “We chose the intersection set of the top 20 most important variables selected by the three machine learning models, after balancing the performance of the final diabetes risk prediction model and the clinical applicability associated with measurement costs of metabolites.”

      (2) The methods section would benefit from a more detailed exposition of how parameter tuning was conducted and the range of parameters explored during the training of the RSF model.

      According to the reviewer’s suggestion, we have added a more detailed description of parameters tunning and the range of parameters explored during the training of the RSF model in the “Method S3” section in the Supplementary material.

      “The RSF model was fitted using the “randomForestSRC” package and the grid search method was used for hyperparameter tuning. Specifically, the grid search method was used to tune hyperparameters among the RSF model, through minimizing out-of-sample or out-of-bag error1. Each tree in the RSF is constructed from a random sample of the data, typically a bootstrap sample or 63.2% of the sample size (as in the present study). Consequently, not all observations are used to construct each tree. The observations that are not used in the construction of a tree are referred to as out-of-bag observations. In an RSF model, each tree is built from a different sample of the original data, so each observation is “out-of-bag” for some of the trees. The prediction for an observation can then be obtained using only those trees for which the observation was not used for the construction. A classification for each observation is obtained in this way and the error rate can be estimated from these predictions. The resulting error rate is referred to as the out-of-bag error. Through calculating the out-of-bag error in each iteration, the best hyperparameters were finally determined.

      The hyperparameters to be tuned and range of grid search in the present study were below: number of trees (50-1000, by 50), number of variables to possibly split at each node (3-6, by 1), and minimum size of terminal node (1-20, by 1)2.”

      (3) It is hard to understand the meaning of the decision curve analysis and the clinical implications behind the net benefit, which are required to clarify the application values of models.

      Thank you for the comment. We have added more description and discussion about the decision curve analysis in the “Methods” and “Discussion” sections.

      “Furthermore, we used decision curve analysis (DCA) to assess the clinical usefulness of prediction model-based guidance for prediabetes management, which calculates a clinical “net benefit” for one or more prediction models in comparison to default strategies of treating all or no patients3.”

      “Most importantly, a model with good discrimination does not necessarily have high clinical value. Hence, DCA was used to compare the clinical utility of the model before and after adding the metabolites, and this showed a higher net benefit for the latter than the basic model, suggesting the addition of the metabolites increased the clinical value of prediction, i.e., the potential benefit of guiding management in individuals with prediabetes3,4. These results provided novel evidence supporting the value of metabolic biomarkers in risk prediction and stratification for the progression from prediabetes to diabetes.”

      (4) Notably, the NMR platform utilized within the UK Biobank primarily focused on lipid species. This limitation should be discussed in the manuscript to provide context for interpreting the results and acknowledge the potential bias from the measuring platform.

      Thank you for the comment. We acknowledged this limitation that NMR platform within the UK Biobank primarily focused on lipid species and the potential bias from the measuring platform and have added this in “Discussion” section.

      “Third, the Nightingale metabolomics platform primarily focused on lipids and lipoprotein sub-fractions, and thus the predictive value of other metabolites in the progression from prediabetes to diabetes warranted further research using an untargeted metabolomics approach.”

      (5) The manuscript should explain the potential influence of non-fasting status on the findings, particularly concerning lipoprotein particles and composition. There should be a detailed discussion of how non-fasting status may impact the measurement and the findings.

      According to the reviewer’s suggestion, we have added more details to explain the potential influence of non-fasting status on our findings in the “Discussion” section.

      “Additionally, the use of non-fasting blood samples might increase inter-individual variation in metabolic biomarker concentrations, however, fasting duration has been reported to account for only a small proportion of variation in plasma metabolic biomarker concentrations5. Therefore, we believe the impact of non-fasting samples on our findings would be minor.”

      (6) Cross-platform standardization is an issue in metabolism, and further descriptions of quality control are recommended.

      Thank you for the comment. We have added more description of quality control in the “Method S1” section in the Supplementary material.

      “Metabolic biomarker profiling by Nightingale Health’s NMR platform provides consistent results over time and across spectrometers. Furthermore, the sample preparation is minimal in the Nightingale Health’s metabolic biomarker platform, circumventing all extraction steps. These aspects result in highly repeatable biomarker measurements. Pre-specified quality metrics were agreed between UK Biobank and Nightingale Health to ensure consistent results across the samples, and pilot measurements were conducted. Nightingale Health performed real-time monitoring of the measurement consistency within and between spectrometers throughout the UK Biobank samples. Two control samples provided by Nightingale Health were included in each 96-well plate for tracking the consistency across multiple spectrometers. Furthermore, two blind duplicate samples provided by the UK Biobank were included in each well plate, with the position information unlocked only after results delivery. Coefficient of variation (CV) targets across the metabolic biomarker profile were pre-specified for both Nightingale Health’s internal control samples and UK Biobank’s blind duplicates. The targets were met for each consecutively measured batch of ~25,000 samples. For the majority of the metabolic biomarkers, the CVs were below 5% (https://biobank.ndph.ox.ac.uk/showcase/refer.cgi?id=3000). Further, the distributions of measured biomarkers from 5 sample batches indicated absence of batch effects (https://biobank.ctsu.ox.ac.uk/ukb/ukb/docs/nmrm_app1).”

      Reviewer #2 (Public Review):<br /> Deciphering the metabolic alterations characterizing the prediabetes-diabetes spectrum could provide early time windows for targeted preventive measures to extend precision medicine while avoiding disproportionate healthcare costs. The authors identified a panel of 9 circulating metabolites combined with basic clinical variables that significantly improved the prediction from prediabetes to diabetes. These findings provided insights into the integration of these metabolites into clinical and public health practice. However, the interpretation of these findings should take account of the following limitations.

      We appreciate the reviewer’s positive comments and encouragement.

      (1) First, the causal relationship between identified metabolites and diabetes or prediabetes deserves to be further examined particularly when the prediabetic status was partially defined. Some metabolites might be the results of prediabetes rather than the casual factors for progression to diabetes.

      Thank you for your insightful comments. We agree with you that the panel of metabolites in this study might not be the causal factor for progression from prediabetes to diabetes, which needs further validation in experimental studies. We have added this limitation in the “Discussion” section.

      “Fifth, we could not draw any conclusion about the causality between the identified metabolites and the risk for progression to diabetes due to the observational nature, which remained to be validated in further experimental studies.”

      (2) The blood samples were taken at random (not all in a non-fasting state) and so the findings were subjected to greater variability. This should be discussed in the limitations.

      According to the reviewer’s suggestion, we have added more details to explain the potential influence of non-fasting status on our findings in the “Discussion” section.

      “Additionally, the use of non-fasting blood samples might increase inter-individual variation in metabolic biomarker concentrations, however, fasting duration has been reported to account for only a small proportion of variation in plasma metabolic biomarker concentrations5. Therefore, we believe the impact of non-fasting samples on our findings would be minor.”

      (3) The strength of NMR in metabolic profiling compared to other techniques (i.e., mass spectrometry [MS], another commonly used metabolic profiling method) could be added in the Discussion section.

      According to the reviewer’s suggestion, we have added the strength of NMR in metabolic profiling compared to other techniques in the “Discussion” section.

      “Circulating metabolites were quantified via NMR-based metabolome profiling within the UK Biobank, which offers metabolite qualification with relatively lower costs and better reproducibility6.”

      (4) Fourth, the applied platform focuses mostly on lipid species which may be a limitation as well.

      Thank you for the comment. We acknowledged this limitation that NMR platform within the UK Biobank primarily focused on lipid species and the potential bias from the measuring platform and have added this in the “Discussion” section.

      “Third, the Nightingale metabolomics platform primarily focused on lipids and lipoprotein sub-fractions, and thus the predictive value of other metabolites in the progression from prediabetes to diabetes warranted further research using an untargeted metabolomics approach.”

      (5) It is a very large group with pre-diabetes, but the results only apply to prediabetes and not to the general population. This should be clear, although the authors have also validated the predictive value of these metabolites in the general population.

      Thank you for the comment. We agree with you that the results only apply to prediabetes and not to the general population, though they also showed potential predictive value among participants with normoglycemia. We have accordingly modified the relevant expressions in the “Conclusion” section to restrict these findings to participants with prediabetes.

      “In this large prospective study among individuals with prediabetes, we detected a panel of circulating metabolites that were associated with an increased risk of progressing to diabetes.”

      Recommendations for the Authors:

      Thank you for providing the valuable feedback and the time you have dedicated to our work.

      (1) In the first paragraph of the Discussion section, please include the specific names of the metabolites selected from machine learning methods.

      Thank you for your comment and we have added accordingly in the first paragraph of the “Discussion” section.

      “More importantly, our findings suggested that adding the selected metabolites (i.e., cholesteryl esters in large HDL, cholesteryl esters in medium VLDL, triglycerides in very large VLDL, average diameter for LDL particles, triglycerides in IDL, glycine, tyrosine, glucose, and docosahexaenoic acid) could significantly improve the risk prediction of progression from prediabetes to diabetes beyond the conventional clinical variables.”

      (2) To enhance the readability and simplicity of the paper, the description of covariate collection in the methods section should be streamlined, with detailed information provided in the supplementary materials.

      Thank you for your suggestion and we have moved details about covariates collection to the “Supplementary method S2” to enhance the readability and simplicity of the paper.

      “Information on covariates was collected through a self-completed touchscreen questionnaire or verbal interview at baseline, including age, sex, ethnicity, Townsend deprivation index, household income, education, employment status, smoking status, moderate alcohol, physical activity, healthy diet score, healthy sleep score, family history of diabetes, history of cardiovascular disease (CVD), history of hypertension, history of dyslipidemia, history of chronic lung diseases (CLD), and history of cancer.

      Physical measurements included systolic (SBP) and diastolic blood pressure (DBP), height, weight, waist circumference (WC), and hip circumference (HC). Body mass index (BMI) was calculated as weight in kilograms divided by the square of height in meters (kg/m²). Missing covariates were imputed by the median value for continuous variables and a missing indicator for categorical variables. More details about covariates collection can be found in Method S2.”

      3. Title for Table 2, using Cox proportional hazards prediction models is not common. You may consider the title "Performance of Cox proportional hazards regression models in prediction of progression of prediabetes to diabetes".

      Thank you for your suggestion and we have revised it accordingly.

      4. Figure 3, did the authors consider competing risk to compute cumulative incidence function?

      Thank you for your comment. We did not consider competing risk from death when plotting the cumulative hazard curves. However, following your suggestion, we have included an additional cumulative hazard plot after considering the competing

      References

      (1) Janitza S, Hornung R. On the overestimation of random forest's out-of-bag error. PLoS One. 2018;13(8):e0201904.

      (2) Tian D, Yan HJ, Huang H, et al. Machine Learning-Based Prognostic Model for Patients After Lung Transplantation. JAMA Netw Open. 2023;6(5):e2312022.

      (3) Vickers AJ, van Calster B, Steyerberg EW. A simple, step-by-step guide to interpreting decision curve analysis. Diagn Progn Res. 2019;3:18.

      (4) Li J, Xi F, Yu W, Sun C, Wang X. Real-Time Prediction of Sepsis in Critical Trauma Patients: Machine Learning-Based Modeling Study. JMIR Form Res. 2023;7:e42452.

      (5) Li-Gao R, Hughes DA, le Cessie S, et al. Assessment of reproducibility and biological variability of fasting and postprandial plasma metabolite concentrations using 1H NMR spectroscopy. PLoS One. 2019;14(6):e0218549.

      (6) Geng T-T, Chen J-X, Lu Q, et al. Nuclear Magnetic Resonance–Based Metabolomics and Risk of CKD. American Journal of Kidney Diseases. 2023.

    1. Reviewer #3 (Public Review):

      Summary:<br /> The authors introduce a tool that employs thermal cameras to automatically detect urine and feces deposits in rodents. The detection process involves a heuristic to identify potential thermal regions of interest, followed by a transformer network-based classifier to differentiate between urine, feces, and background noise. The tool's effectiveness is demonstrated through experiments analyzing social preference, stress response, and temporal dynamics of deposits, revealing differences between male and female mice.

      Strengths:<br /> The method effectively automates the identification of deposits<br /> The application of the tool in various behavioral tests demonstrates its robustness and versatility.<br /> The results highlight notable differences in behavior between male and female mice

      Weaknesses:<br /> The definition of 'start' and 'end' periods for statistical analysis is arbitrary. A robustness check with varying time windows would strengthen the conclusions.<br /> The paper could better address the generalizability of the tool to different experimental setups, environments, and potentially other species.<br /> The results are based on tests of individual animals, and there is no discussion of how this method could be generalized to experiments tracking multiple animals simultaneously in the same arena (e.g., pair or collective behavior tests, where multiple animals may deposit urine or feces).

    2. Author response:

      We want to thank the reviewers for their constructive feedback.

      General

      The recall values of our method range between 78.6% for all urine cases to 83.3% for feces (and not between 70-80%, as stated by reviewer #2), with a mean precision of 85.6%. This is rather similar to other machine learning-based methods commonly used for the analysis of complicated behavioral readouts. For example, in the paper presenting DeepSqueak for analysis of mouse ultrasonic vocalizations (Coffey et al. DeepSqueak: a deep learning-based system for detection and analysis of ultrasonic vocalizations. Neuropsychopharmacol. 44, 859–868 (2019). https://doi.org/10.1038/s41386-018-0303-6), the recall values reported for both DeepSqueak, Mupet and Ultravox (Fig. 2c, f) are very similar to our method.

      We have analyzed and reported all the types of errors made by our methods, which are mostly technical. For example, depositions that overlap the mouse blob for too long till getting cold will be associated with the mouse and therefore will not be detected (“miss” events). These technical errors are not supposed to create a bias for a specific biological condition and, hence, shouldn’t interfere with the use of our method. A video showing all of the mistakes made by our algorithm on the test set was submitted (Figure 2-video 1).

      Below we will to relate to specific points and describe our plan to revise the manuscript accordingly.

      Detection accuracy

      a. It should be noted that when large urine spots are considered, our algorithm got 100% correct classification (Figure 2, supplement 1, panel b). However, small urine deposits are very similar to feces in their appearance in the thermal picture. In fact,  if the feces are not shifted, discrimination can be quite challenging even for human annotators. To demonstrate the accuracy of the proposed method relative to human annotators, we plan to compare its results with the accuracy of a second human annotator.

      b. As part of the revision, we plan to test general machine learning-based object detectors such as faster-RCNN or YOLO (as suggested by Reviewer 2) and compare them with our method.

      c. To check if our method may introduce bias to the results, we plan to check if the errors are distributed evenly across time, space, and genders.

      Design choices

      (A) The preliminary detection algorithm has several significant parameters. These are:

      a. Minimal temperature rise for detection: 1.1°C rise during 5 sec.

      b. Size limits of the detection: 2 - 900 pixels.

      c. Minimal cooldown during 40 sec: 1.1°C and at least half the rise.

      d. Minimal time between detections in the same location: 30 sec.

      We chose to use low thresholds for the preliminary detection to allow detection of very small urinations and to minimize the number of “miss” events, relying on the classifier to robustly reject false alarms. Indeed, we achieved a low rate of miss events: 5 miss events for the entire test set (1 miss event per ~90 minutes of video). We attribute these 5 “miss” events to partial occlusion of the detection by the mouse.

      To adjust the preliminary detection parameters to a new environment, one will need to calibrate these parameters in their own setup. Mainly, the size of the detection depends on the resolution of the video, and the cooldown rate might be affected by the material of the floor, as well as the room temperature.

      We plan to explore the robustness of these parameters in our setup and report the influence on the accuracy of the preliminary algorithm.

      (B) We chose to feed the classifier with 71 seconds of videos (11 seconds before the event and 60 seconds after it) as we wanted the classifier to be able to capture the moment of the deposition, the cooldown process, as well as urine smearing or feces shifting which might give an additional clue for the classification. In the revised paper we plan to report accuracy when using a shorter video for classification.

      Generability

      a. In the revised version, we plan to report the accuracy of the method used on a different strain of mice (C57), with a different arena color (white arena instead of black).

      Statistics

      a. In the revised paper, we will explain why we chose each time window for analysis. Also, we will report statistics for different time windows, as suggested by Reviewer 3.

      b. Unlike reviewer #2, we don’t think that the small difference in recall rate between urine and feces (78.6% vs. 83.3%, respectively) creates a bias between them. Moreover, we don’t compare the urine rate to the feces rate.

      c. In the revised manuscript we will explicitly report the precision scores, although they also appear in our manuscript in Fig. 2- Supplement 1b.

  3. Aug 2024
    1. Reviewer #1 (Public Review):

      The authors are attempting to use the internal workings of a language hierarchy model, comprising phonemes, syllables, words, phrases, and sentences, as regressors to predict EEG recorded during listening to speech. They also use standard acoustic features as regressors, such as the overall envelope and the envelopes in log-spaced frequency bands. This is valuable and timely research, including the attempt to show differences between normal-hearing and hearing-impaired people in these regards.

      I will start with a couple of broader questions/points, and then focus my comments on three aspects of this study: The HM-LSTM language model and its usage, the time windows of relevant EEG analysis, and the usage of ridge regression.

      Firstly, as far as I can tell, the OSF repository of code, data, and stimuli is not accessible without requesting access. This needs to be changed so that reviewers and anybody who wants or needs to can access these materials.

      What is the quantification of model fit? Does it mean that you generate predicted EEG time series from deconvolved TRFs, and then give the R2 coefficient of determination between the actual EEG and predicted EEG constructed from the convolution of TRFs and regressors? Whether or not this is exactly right, it should be made more explicit.

      About the HM-LSTM:

      • In the Methods paragraph about the HM-LSTM, a lot more detail is necessary to understand how you are using this model. Firstly, what do you mean that you "extended" it, and what was that procedure? And generally, this is the model that produces most of the "features", or regressors, whichever word we like, for the TRF deconvolution and EEG prediction, correct? A lot more detail is necessary then, about what form these regressors take, and some example plots of the regressors alongside the sentences.<br /> • Generally, it is necessary to know what these regressors look like compared to other similar language-related TRF and EEG/MEG prediction studies. Usually, in the case of e.g. Lalor lab papers or Simon lab papers, these regressors take the form of single-sample event markers, surrounded by zeros elsewhere. For example, a phoneme regressor might have a sample up at the onset of each phoneme, and a word onset regressor might have a sample up at the onset of each word, with zeros elsewhere in the regressor. A phoneme surprisal regressor might have a sample up at each phoneme onset, with the value of that sample corresponding to the rarity of that phoneme in common speech. Etc. Are these regressors like that? Or do they code for these 5 linguistic levels in some other way? Either way, much more description and plotting is necessary in order to compare the results here to others in the literature.<br /> • You say that the 5 regressors that are taken from the trained model's hidden layers do not have much correlation with each other. However, the highest correlations are between syllable and sentence (0.22), and syllable and word (0.17). It is necessary to give some reason and interpretation of these numbers. One would think the highest correlation might be between syllable and phoneme, but this one is almost zero. Why would the syllable and sentence regressors have such a relatively high correlation with each other, and what form do those regressors take such that this is the case?<br /> • If these regressors are something like the time series of zeros along with single sample event markers as described above, with the event marker samples indicating the onset of the relevant thing, then one would think e.g. the syllable regressor would be a subset of the phoneme regressor because the onset of every syllable is a phoneme. And the onset of every word is a syllable, etc.

      For the time windows of analysis:

      • I am very confused, because sometimes the times are relative to "sentence onset", which would mean the beginning of sentences, and sometimes they are relative to "sentence offset", which would mean the end of sentences. It seems to vary which is mentioned. Did you use sentence onsets, offsets, or both, and what is the motivation?<br /> • If you used onsets, then the results at negative times would not seem to mean anything, because that would be during silence unless the stimulus sentences were all back to back with no gaps, which would also make that difficult to interpret.<br /> • If you used offsets, then the results at positive times would not seem to mean anything, because that would be during silence after the sentence is done. Unless you want to interpret those as important brain activity after the stimuli are done, in which case a detailed discussion of this is warranted.<br /> • For the plots in the figures where the time windows and their regression outcomes are shown, it needs to be explicitly stated every time whether those time windows are relative to sentence onset, offset, or something else.<br /> • Whether the running correlations are relative to sentence onset or offset, the fact that you can have numbers outside of the time of the sentence (negative times for onset, or positive times for offset) is highly confusing. Why would the regressors have values outside of the sentence, meaning before or after the sentence/utterance? In order to get the running correlations, you presumably had the regressor convolved with the TRF/impulse response to get the predicted EEG first. In order to get running correlation values outside the sentence to correlate with the EEG, you would have to have regressor values at those time points, correct? How does this work?<br /> • In general, it seems arbitrary to choose sentence onset or offset, especially if the comparison is the correlation between predicted and actual EEG over the course of a sentence, with each regressor. What is going on with these correlations during the middle of the sentences, for example? In ridge regression TRF techniques for EEG/MEG, the relevant measure is often the overall correlation between the predicted and actual, calculated over a longer period of time, maybe the entire experiment. Here, you have calculated a running comparison between predicted and actual, and thus the time windows you choose to actually analyze can seem highly cherry-picked, because this means that most of the data is not actually analyzed.<br /> • In figures 5 and 6, some of the time window portions that are highlighted as significant between the two lines have the lines intersecting. This looks like, even though you have found that the two lines are significantly different during that period of time, the difference between those lines is not of a constant sign, even during that short period. For instance, in figure 5, for the syllable feature, the period of 0 - 200 ms is significantly different between the two populations, correct? But between 0 and 50, normal-hearing are higher, between 50 and 150, hearing-impaired are higher, and between 150 and 200, normal-hearing are higher again, correct? But somehow they still end up significantly different overall between 0 and 200 ms. More explanation of occurrences like these is needed.

      Using ridge regression:

      • What software package(s) and procedure(s) were specifically done to accomplish this? If this is ridge regression and not just ordinary least squares, then there was at least one non-zero regularization parameter in the process. What was it, how did it figure in the modeling and analysis, etc.?<br /> • It sounds like the regressors are the hidden layer activations, which you reduced from 2,048 to 150 non-acoustic, or linguistic, regressors, per linguistic level, correct? So you have 150 regressors, for each of 5 linguistic levels. These regressors collectively contribute to the deconvolution and EEG prediction from the resulting TRFs, correct? This sounds like a lot of overfitting. How much correlation is there from one of these 150 regressors to the next? Elsewhere, it sounds like you end up with only one regressor for each of the 5 linguistic levels. So these aspects need to be clarified.<br /> • For these regressors, you are comparing the "regression outcomes" for different conditions; "regression outcomes" are the R2 between predicted and actual EEG, which is the coefficient of determination, correct? If this is R2, how is it that you have some negative numbers in some of the plots? R2 should be only positive, between 0 and 1.

    1. ¡Hola! Fossil es un sistema de control de versiones distribuido (DVCS) que también ofrece seguimiento de errores, wiki, foro, alertas por correo electrónico, chat y servicios de notas técnicas. Viene con una interfaz web integrada, simplificando el seguimiento de proyectos y promoviendo la conciencia de la situación. Fossil admite sistemas operativos Linux, BSD, Mac, y Windows. Es una herramienta útil para mantener el control de versiones de software y colaborar en proyectos.

      Ni la pregunta ni la respuesta capturan el contexto.

    1. Reviewer #3 (Public Review):

      Summary:

      This paper aims to investigate how the human brain represents different forms of value and uncertainty that participate in active inference within a free-energy framework, in a two-stage decision task involving contextual information sampling, and choices between safe and risky rewards, which promotes shifting between exploration and exploitation. They examine neural correlates by recording EEG and comparing activity in the first vs second half of trials and between trials in which subjects did and did not sample contextual information, and perform a regression with free-energy-related regressors against data "mapped to source space."

      Strengths:

      This two-stage paradigm is cleverly designed to incorporate several important processes of learning, exploration/exploitation and information sampling that pertain to active inference. Although scalp/brain regions showing sensitivity to the active-inference related quantities do not necessary suggest what role they play, they are illuminating and useful as candidate regions for further investigation. The aims are ambitious, and the methodologies impressive. The paper lays out an extensive introduction to the free energy principle and active inference to make the findings accessible to a broad readership.

      Weaknesses:<br /> In its revised form the paper is complete in providing the important details. Though not a serious weakness, it is important to note that the high lower-cutoff of 1 Hz in the bandpass filter, included to reduce the impact of EEG noise, would remove from the EEG any sustained, iteratively updated representation that evolves with learning across trials, or choice-related processes that unfold slowly over the course of the 2-second task windows.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:<br /> I really enjoyed this manuscript from Torsekar et al on "Contrasting responses to aridity by

      different-sized decomposers cause similar decomposition rates across a precipitation gradient". The authors aimed to examine how climate interacts with decomposers of different size categories to influence litter decomposition. They proposed a new hypothesis: "The opposing climatic dependencies of macrofauna and that of microorganisms and mesofauna should lead to similar overall decomposition rates across precipitation gradients".

      This study emphasizes the importance as well as the contribution of different groups of organisms (micro, meso, macro, and whole community) across different seasons (summer with the following characteristics: hot with no precipitation, and winter with the following characteristics: cooler and wetter winter) along a precipitation gradient. The authors made use of 1050 litter baskets with different mesh sizes to capture decomposers contribution. They proposed a new hypothesis that was aiming to understand the "dryland decomposition conundrum". They combined their decomposition experiment with the sampling of decomposers by using pittfall traps across both experiment seasons. This study was carried out in Israel and based on a single litter species that is native to all seven sites. The authors found that microorganism contribution dominated in winter while macrofauna decomposition dominated the overall decomposition in summer. These seasonality differences combined with the differences in different decomposers groups fluctuation along precipitation resulted in similar overall decomposition rates across sites.<br /> I believe this manuscript has a potential to advance our knowledge on litter decomposition.

      Strengths:

      Well design study with combination of different approaches (methods) and consideration of seasonality to generalize pattern.

      The study expands to current understanding of litter decomposition and interaction between factors affecting the process (here climate and decomposers).

      Weaknesses:

      The study was only based on a single litter species.

      We now discuss the advantages and limitations of this approach in the methods and devote a completely new paragraph to this important point in the discussion (lines 394-401).

      Reviewer #2 (Public Review):

      Summary: Torsekar et al. use a leaf litter decomposition experiment across seasons, and in an aridity gradient, to provide a careful test of the role of different-sized soil invertebrates in shaping the rates of leaf litter decomposition. The authors found that large-sized invertebrates are more active in the summer and small-sized invertebrates in the winter. The summed effects of all invets then translated into similar levels of decomposition across seasons. The system breaks down in hyper-arid sites.

      Strengths: This is a well-written manuscript that provides a complete statistical analysis of a nice dataset. The authors provide a complete discussion of their results in the current literature.

      Weaknesses:

      I have only three minor comments. Please standardize the color across ALL figures (use the same color always for the same thing, and be friendly to color-blind people).

      Thank you for this important suggestion. We have now changed all figures to standardize all colors and chose a more color-blind friendly pallete.

      Fig 1 may benefit from separating the orange line (micro and meso) into two lines that reflect your experimental setup and results. I would mention the dryland decomposition conundrum earlier in the Introduction.

      We based our novel hypotheses on a thorough literature search. Accordingly, decomposition is expected to be positively associated with moisture, regardless of the decomposer body size. Our contribution to theory was to suggest that macro-detritivores may respond very differently to climatic conditions and dominate litter decomposition in warm arid-lands (we listed the reasons in the text). Consequently, we did not distinguish between microorganisms and mesofauna. We assumed that both groups inhabit the litter substrate and have limited adaptation to dry conditions. Our results provide strong evidence that this presumption is likely wrong and that mesofauna respond to climate very differently from micro-decomposers. Yet, we cannot use hindsight understanding to improve our original hypothesis. We now emphasize this important point at the discussion as important future direction. 

      Although we are very appreciative and pleased with the reviewer enthusiasm to highlight the importance of our work as a possible solution to the longstanding dryland decomposition conundrum, we decided not to move it to the introduction. This is because we think that our work is not centred on resolving the DDC but provides more general principles that may lead to a paradigm shift in the way ecologists study nutrient cycling across ecosystems.

      And the manuscript is full of minor grammatical errors. Some careful reading and fixing of all these minor mistakes here and there would be needed.

      We apologize and did our best to find and fix those mistakes

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      I really enjoyed this manuscript from Torsekar et al on "Contrasting responses to aridity by different-sized decomposers cause similar decomposition rates across a precipitation gradient". The authors aimed to examine how climate interacts with decomposers of different size categories to influence litter decomposition. They proposed a new hypothesis: "The opposing climatic dependencies of macrofauna and that of microorganisms and mesofauna should lead to similar overall decomposition rates across precipitation gradients".

      This study emphasizes the importance as well as the contribution of different groups of organisms (micro, meso, macro, and whole community) across different seasons (summer with the following characteristics: hot with no precipitation, and winter with the following characteristics: cooler and wetter winter) along a precipitation gradient. The authors made use of 1050 litter baskets with different mesh sizes to capture decomposers contribution. They proposed a new hypothesis that was aiming to understand the "dryland decomposition conundrum". They combined their decomposition experiment with the sampling of decomposers by using pitfall traps across both experiment seasons. This study was carried out in Israel and based on a single litter species that is native to all seven sites. The authors found that microorganism contribution dominated in winter while macrofauna decomposition dominated the overall decomposition in summer. These seasonality differences combined with the differences in different decomposers groups fluctuation along precipitation resulted in similar overall decomposition rates across sites.

      I believe this manuscript has the potential to advance our knowledge on litter decomposition. Below i provide my general and specific comments.

      General comments:

      (1) Study in general is well designed and well thought beforehand,

      (2) Study aims to expand the current understanding of the dryland decomposition conundrum

      (3) The should put a caveat to the fact they only use one litter species and call for examining litter mixture in the same gradient.

      (4) Please check the way you reduce the random effects from your initial model, I have provided a better way to do so in my specific comments

      (5) For Figure 1, authors can check my comment on this and see if they could revise the figure.

      Thank you for the positive feedback and your valuable comments. We have tried to best address all comments and suggestions for improvement and clarification

      Specific comments

      Line # 57 Please write "Theory suggests" instead of "Theory suggest"

      We changed the text as suggested

      Line # 70, please write "Indeed, handful evidence shows" instead of "Indeed, handful evidence show"

      We changed the text as suggested

      Figure 1: I like this conceptual framework. I have a silly question, why is it that the slopes of the whole community at the beginning (between Hyperarid and Arid) is the same as the Macro fauna, I would think the slope should be higher as this is adding up right? and also the same goes for the decomposition of whole community later on. For me this should reflect the adding or summing up (if i am right) then the authors should think about how this could be reflected in the figure.

      We agree with your interpretation that the whole community decomposition reflects the addition by constituent decomposers. The slope of the whole community decomposition between hyper-arid and arid is slightly higher than the one of macro decomposition to reflect the additive effect of macro with meso+micro decomposition. We have now changed the figure slightly to make this point more visible (Line 106).

      Line # 111 Please make "Methods" bold as well to be consistent with others headings.

      We changed the formatting as suggested

      Line #125 and in other lines as well please replace "X" by "x" to denote multiplication.

      We changed the formatting as suggested

      Table 1 Please add "*" to climate like this "Climate*" so that the end note of the table could make sense

      Thank you for this suggestion. We have now added the asterisk referring to the note below the Table.

      Figure 2, please consider putting at line #133, mean annual precipitation (MAP), as such for line # 135 You can directly says The precipitation map ....

      We made both changes as suggested.

      Line # 138 I would not use the different units for the same values. I do understand that you want to emphasize the accuracy but i would write instead 3 +- 0.001 g

      We changed the units as suggested.

      Line # 145, how is the litter basket customized to rest at 1 cm above ground level?

      We have now clarified –that we cut-open windows one centimeter above the cage floor. The cages were positioned on the soil (line 144).

      Lines # 181-183, I like the approach of checking the necessity of having the random effects. However, it has been reported that likelihood ratio test (LRT) are not really reliable to test for random effects. I will suggest you rather use permutations instead. I think the function is confint(MODEL) you need to specify the number of permutation the higher the better but you should start with 99 first and see how the results look like if promising then you can even go to 9999. But it will need computation power and and time.

      Thank you for the suggestion. We now used a simulation-based exact test, instead of a LRT, to examine the random effect, as recommended by the authors from the “lme4” package. As recommended, we used 9999 simulations. The simulation test yielded a similar result to those originally reported (see lines 181-183).

      Line # 187, 188, 188, please do not use capital letter to start mesofauna, macrofauna and whole-community

      We changed the formatting as suggested

      Line # 205 Please add the version number of R in the text.

      We now included the version number as suggested.

      Line # 209-211, could you please check whether "then" is the word you want to use or "than"

      Our bad- we indeed meant “than” and have made the appropriate changes.

      Line # 227 and in other places as well please provide the second degree of freedom of the F test.

      Thank you for this important comment. We have now added the second degree of freedom to the relevant results (lines 229, 232).

      Figure 3 and Figure 4 show some results that are negative, can you please explain what might be the reasons behind this?

      We now explain this important point in the figures’ captions.

      Figure 5 Please add label to the x-axis.

      Thank you-we have now included a label.

      Line # 357, the sentence "... meso-decomposition, like microbial decomposition,...", I don't understand which criteria authors used to classify microbial decomposition as "meso-decomposition"?

      We now remove this potential cause of confusion by using the term ‘meso-decomposition’ to distinguish from microbial decomposition (Line 366).

      Line # 380 Kindly put "per se" in italic.

      We changed the formatting as suggested

      References

      The references format are not consistent. For example for the same journal (say Trends in Ecology and Evolution) the authors sometimes wrote the full name like at line # 36 (and also realize that "vol" should not be written as such) but wrote the abbreviations at line #42

      Our bad- we apologize and carefully checked all references to make sure the style is consistent.

    1. 1. 项目目录结构及介绍 这个开源项目,位于 GitHub,旨在实现Windows系统的自动主题切换功能,以适应不同的光线条件。下面简要介绍其主要目录结构: Windows Auto Dark Mode.sln - Visual Studio解决方案文件,是项目的核心入口,包含了所有相关组件。 AutoDarkModeApp - 应用程序主体代码所在目录。 AutoDarkModeComms, AutoDarkModeLib, AutoDarkModeShell, AutoDarkModeSvc, AutoDarkModeUpdater - 这些目录分别对应于应用程序的不同模块,如通信、库、壳层逻辑、服务和更新器等。 IThemeManager2Bridge, ThemeDll - 关键功能实现相关的DLL和接口管理代码。 Wiki, README.md, LICENSE, privacy.md - 文档部分,包括项目说明、许可证和隐私政策,以及用户指南。 adm-updater-rs, gitignore, gitmodules - 版本控制和更新脚本相关文件。

      -[] good #share

    1. Author response:

      We are grateful to the reviewers for recognizing the importance of our work and for their helpful suggestions. We will revise our manuscript in the revised version. However, we’d like to provide provisional responses now to answer the key questions and comments from the reviewers.

      (1) Both reviewers asked why we chose 24-120 hpf to measure the apoptotic rates. We chose this time window based on the following two reasons: 1) Previous studies showed that although the motor neuron death time windows vary in chick (E5-E10), mouse (E11.5-E15.5), rat (E15-E18) and human (11-25 weeks of gestation), the common feature of these time windows is that they are all the developmental periods when motor neurons contact with muscle cells. The contact between zebrafish motor neurons and muscle cells occurs before 72 hpf, which is included in our observation time window. 2) Zebrafish complete hatching during 48-72 hpf, and most organs form before 72 hpf. More importantly, zebrafish start swimming around 72 hpf, indicating that motor neurons are fully functional.

      Thus, we are confident that this 24-120 hpf time window covers the time window during which motor neurons undergo programmed cell death during zebrafish early development. We frequently used “early development” in this manuscript to describe our observation. However, we missed “early” in our title. We will add “early” in the title in the revised version.

      (2) Both reviewers also asked about the neurogenesis of motor neurons. Previous studies have shown that the production of spinal cord motor neurons largely ceases before 48 hpf and then the motor neurons remain largely constant until adulthood. Our observation time window covers the major motor neuron production process. Therefore, we believe that neurogenesis will not affect our data and conclusions.

      (3) Both reviewers questioned the specificity of using the mnx1 promoter to label motor neurons. The mnx1 promoter has been widely used to label motor neurons in transgenic zebrafish. Previous studies have shown that most of the cells labeled in the mnx1 transgenic zebrafish are motor neurons. In this study, we observed that the neuronal cells in our sensor zebrafish formed green cell bodies inside of the spinal cord and extended to the muscle region, which is an important morphological feature of the motor neurons. Furthermore, a few of those green cell bodies turned into blue apoptotic bodies inside the spinal cord and changed to blue axons in the muscle regions at the same time, which strongly suggests that those apoptotic neurons are not interneurons. Although the mnx1 promoter might have labeled some interneurons, this will not affect our major finding that only a small portion of motor neurons died during zebrafish early development.

      (4) Reviewer 2 is concerned that the estimated 50% of motor neuron death was in limb-innervating motor neurons but not in body wall-innervating motor neurons. The death of motor neurons in limb-innervating motor neurons has been extensively studied in chicks and rodents, as it is easy to undergo operations such as amputation. However, previous studies have shown this dramatic motor neuron death does not only occur in limb-innervating motor neurons but also occurs in other spinal cord motor neurons. In our manuscript, we studied the naturally occurring motor neuron death in the whole spinal cord during the early stage of zebrafish development.

      (5) Reviewer 2 mentioned that we ignored the death of an identified motor neuron. Our study was to examine the overall motor neuron apoptosis rather than a specific type of motor neuron death, so we did not emphasize the death of VaP motor neurons. We agree that the dead motor neurons observed in our manuscript contain VaP motor neurons. However, there were also other types of dead motor neurons observed in our study. The reasons are as follows: 1) VaP primary motor neurons die before 36 hpf, but our study found motor neuron cells died after 36 hpf and even at 84 hpf. 2) The position of the VaP motor neuron is together with that of the CaP motor neuron, that is, at the caudal region of the motor neuron cluster. Although it’s rare, we did observe the death of motor neurons in the rostral region of the motor neuron cluster. 3) There is only one or zero VaP motor neuron in each hemisegment. Although our data showed that usually one motor neuron died in each hemisegment, we did observe that sometimes more than one motor neuron died in the motor neuron cluster. We will include this information in the revised manuscript.

      (6) For the morpholinos, we did not confirm the downregulation of the target genes. These morpholino-related data are a minor part of our manuscript and shall not affect our major findings. Thus, we didn’t think we missed “important” controls. We will perform experiments to confirm the efficiency of the morpholinos or remove these morpholino-related data from the revised version.

    1. The image of someone having three different windows that each portray a misleading landscape made me reconsider the different conditions that one is put under and how it could influence decisions. I see it being that if you see the bright greenery versus the image of the broken fire hydrant, it would create vastly contrasting mental modes that would contribute to your thinking.

    1. Reviewer #1 (Public Review):

      Summary:

      In their paper, Hou and co-workers explored the use of a FRET sensor for endogenous g-sec activity in vivo in the mouse brain. They used AAV to deliver the sensor to the brain for neuron specific expression and applied NIR in cranial windows to assess FRET activity; optimizing as well an imaging and segmentation protocol. In brief they observe clustered g-sec activity in neighboring cells arguing for a cell non-autonomous regulation of endogenous g-sec activity in vivo.

      Strengths:

      Mone.

      Weaknesses:

      Overall the authors provide a very limited data set and in fact only a proof of concept that their sensor can be applied in vivo. This is not really a research paper, but a technical note. With respect to their observation of clustered activity, they now provide an overview image, next to zoomed details. However, from these images one cannot conclude 'by eye' any clustering event. This aligns with the very low r values. All neurons in the field show variable activity and a clustering is not really evident from these examples. Even within a cluster, there is variability. The authors now confirm that expression levels are indeed variable but are independent from the ratio measurements. Further, they controlled for specificity by including DAPT treatments, but opposite to their own in vitro data (in primary neurons) the ratios increased. The authors argue that both distance and orientation can either decrease or increase ratios and that the use of this biosensor should be explored model-by-model. This doesn't really confer high confidence and may hinder other groups in using this sensor reliably.

      Secondly, there is still no physiological relevance for this observation. The experiments are performed in wild-type mice, but it would be more relevant to compare this with a fadPSEN1 KI or a PSEN1cKO model to investigate the contribution of a gain of toxic function or LOF to the claimed cell non-autonomous activations. The authors acknowledge this shortcoming but argue that this is for a follow-up study.

      For instance, they only monitor activity in cell bodies, and miss all info on g-sec activity in neurites and synapses: what is the relevance of the cell body associated g-sec and can it be used as a proxy for neuronal g-sec activity? If cells 'communicate' g-sec activities, I would expect to see hot spots of activity at synapses between neurons.

      Without some more validation and physiologically relevant studies, it remains a single observation and rather a technical note paper, instead of a true research paper.

    1. Late-replicating regions show higher under-representation in both non-subtelomeric and subtelomeric regions (N.S. p > 0.05; ***p < 0.0001, Wilcoxon rank test). All 200-bp windows of measured under-replication split into bins based on their replication timing (data from51). Colour-code corresponds to the proximity to telomeres (1:100 kb). ~0.1% of 200 bp regions have under-representation less than −20%; for visualization we plot them at -20%

      After release from G1, subtelomeric regions were more under-replicated than regions further away from telomeric regions.

    2. Under-representation values for all 200 bp windows throughout the genome, with significantly underrepresented genomic regions colored green.

      As you move further from the telomere, the underreplicated DNA decreases.

    1. Configuring the client for use in dual-boot (Windows / Linux) situations When dual booting Windows and Linux, depending on the Windows OneDrive account configuration, the 'Files On-Demand' option may be enabled when running OneDrive within your Windows environment.

      Mount, rather than rsync for offline use when traveling.

    1. Reviewer #3 (Public Review):

      This study by Guan and co-workers focuses on a model neuronal lineage in the developing Drosophila nervous system, revealing interesting aspects about: a) the generation of supernumerary cells, later destined for apoptosis; and, b) new insights into the mechanisms that regulate this process. The two RNA-binding proteins, Imp and Syp, are shown to be expressed in temporally largely complementary patterns, their expression defining early vs later born neurons in this lineage, and thus also regulating the apoptotic elimination. Moreover, neuronal 'fate' transcription factors that are downstream of Imp and signatures of early-born neurons, can also be sufficient to convert later born cells to an earlier 'fate', including survival. The authors provide solid evidence for most of their statements, including the temporal windows during which the early and the later-born motoneurons are generated by this model lineage, how this relates to patterns of cell death by apoptosis and that mis-expression of early-born transcription factors in later-born cells can be sufficient to block apoptosis (part of, and perhaps indicative of the late-born identity). Other studies have previously outlined analogous, mutually antagonistic roles for Imp and Syp during nervous system development in Drosophila, in different parts and at different stages, with which the working model of this study aligns. Overall, this study adds to and extends current working models and evidence on the developmental mechanisms that underlie temporal cell fate decisions.

    2. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Review):

      This study addresses the temporal patterning of a specific Drosophila CNS neuroblast lineage, focusing on its larval development. They find that a temporal cascade, involving the Imp and Syb genes changes the fate of one daughter cell/branch, from glioblast (GB) to programmed cell death (PCD), as well as gates the decommissioning of the NB at the end of neurogenesis.

      I believe there are some inaccuracies in this summary. We address temporal patterning during larval and pupal stages until the adult stage. The Imp and Syp genes change the fate of one daughter cell/branch from survival to programmed cell death (PCD). The change from glioblast (GB) to PCD, which occurs at an early time point, is not addressed here. The main point of the paper is missing:

      • Last-born MNs undergo apoptosis due to their failure to express a functional TF code, and this code is post-transcriptionally regulated by the opposite expression of Imp and Syp in immature MNs.

      Reviewer #2 (Public Review):

      Summary:

      Guan and colleagues address the question of how a single neuroblast produces a defined number of progeny, and what influences its decommissioning. The focus of the experiments are two well-studied RNA-binding proteins: Imp and Syp. The Authors find that these factors play an important role in determining the number of neurons in their preferred model system of VNC motor neurons coming from a single lineage (LinA/15) by separate functions taking place at specific stages of development of this lineage: influencing the life-span of the LinA neuroblast to control its timely decommissioning and functioning in the Late-born post-mitotic neurons to influence cell death after the appropriate number of progeny is generated. The post-mitotic role of Imp/Syp in regulating programmed-cell death (PCD) is also correlated with a specific code of key transcription factors that are suspected to influence neuronal identity, linking the fate of neuronal survival with its specification. This paper addresses a wide scope of phenotypes related to the same factors, thus providing an intriguing demonstration of how the nervous system is constructed by context-specific changes in key developmental regulators.

      The bulk of conclusions drawn by the authors are supported by careful experimental evidence, and the findings are a useful addition to an important topic in developmental neuroscience.

      I cannot summarize better the paper.

      Strengths:

      A major strength is the use of a genetic labeling tool that allows the authors to specifically analyze and manipulate one neuronal lineage. This allows for simultaneous study of both the progenitors and post-mitotic progeny. As a result the paper conveys a lot of useful information for this particular neuronal lineage. Furthermore addressing the association of cell fate specification, taking advantage of this lab's extensive prior work in the system, with developmentally-regulated programmed celldeath is an important contribution to the field.

      Beyond Imp/Syp, additional characterization of this model system is provided in characterizing a previously unrecognized death of a hemilineage in early-born neurons.

      Thanks!

      Weaknesses:

      The main observations that distinguish this study from others that have investigated Imp/Syp in the fly nervous system is the role played in late-born post-mitotic neurons to regulate programmed cell death. This is an important and plausible (based on the presented findings) newly discovered role for these proteins. However the precision of experiments is not particularly strong, which limits the authors claims. The genetic strategy used to manipulate Imp/Syp or the TF code appears to be done throughout the entire lineage, or all neuronal progeny, and not restricted to only the late born cells. Can the authors rule out survival of the early born hemi-lineage normally fated to die? Therefore statements such as this: 

      To further investigate this possibility, we used the MARCM technique to change the TF code of lastborn MNs without affecting the expression of Imp and Syp should be qualified to specify that the result is obtained by misexpressing these factors throughout the entire lineage.

      We agree that our genetic manipulations affect the entire lineage or all neuronal progeny. We do not have genetic tools to gain such precision. We have changed our descriptions to specify the entire lineage or all neuronal progeny. As the reviewer raised, we were also concerned about the possibility that the overexpression of Imp or knockdown of Syp could induce the survival of the early-born hemilineage. We have two experiments that rule out this possibility:

      (1) In late LL3 larvae, Imp OE or syp MARCM clones do not change the number of cells in LL3 larvae (see Guan et al., 2022), indicating that the hemilineage that died by PCD is not affected. If Imp or Syp played a role in the survival of the hemilineage, we would see at least a 50% increase in the number of MNs at this stage.

      (2) The MARCM experiment using the VGlut driver to overexpress P35 or Imp allows us to manipulate only elav+ VGlut+ neurons. The hemilineage removed by PCD is elav- VGlut- and is not affected by this experiment. Consequently, the increase in MNs in adults with genetic manipulation can only be the result of the survival of the other hemilineage (elav+, VGlut+). Moreover, this experiment shows an increase in the number of neurons in the adult but not in LL3, demonstrating that the hemilineage (elav- VGlut-) is still removed by PCD with this genetic manipulation.

      The authors make an observation that differs from other systems in which Imp/Syp have been studied: that the expression of the two proteins appears to be independent and not influenced by cross-regulation. However there is a lack of investigation as to what effect this may have on how Imp/Syp regulate temporal identity. A key implication of the previously observed cross-regulation in the fly mushroom body is that the ratio of Imp/Syp could change over the life of the NB which would permit different neuronal identities. Without cross-regulation, do the authors still observe a gradient in the expression pattern of time? Because the data is presented with Imp and Syp stained in different brain samples, and without quantification across different stages, this is unclear. The authors use the term 'gradient' but changes in levels of these factors are not evident from the presented data.

      We have now quantified the transcriptional activity of Imp and Syp in the NB over time using smFISH. We have also quantified the relative expression of Imp and Syp protein in the NB over time by co-immunostaining. Additionally, we quantified the relative expression of Imp and Syp protein in postmitotic neurons as a function of their birth order in late LL3 larvae. All these data show an opposite temporal gradient of Imp and Syp in the NB and an opposite spatial gradient in immature neurons according to their birth order (Figure. 4). How these gradients are established in our system remains to be elucidated. 

      Reviewer #3 (Public Review):

      This study by Guan and co-workers focuses on a model neuronal lineage in the developing Drosophila nervous system, revealing interesting aspects about: a) the generation of supernumerary cells, later destined for apoptosis; and, b) new insights into the mechanisms that regulate this process. The two RNA-binding proteins, Imp and Syp, are shown to be expressed in temporally largely complementary patterns, their expression defining early vs later born neurons in this lineage, and thus also regulating the apoptotic elimination. Moreover, neuronal 'fate' transcription factors that are downstream of Imp and signatures of early-born neurons, can also be sufficient to convert later born cells to an earlier 'fate', including survival.

      The authors provide solid evidence for most of their statements, including the temporal windows during which the early and the later-born motoneurons are generated by this model lineage, how this relates to patterns of cell death by apoptosis and that mis-expression of early-born transcription factors in later-born cells can be sufficient to block apoptosis (part of, and perhaps indicative of the late-born identity).

      Other studies have previously outlined analogous, mutually antagonistic roles for Imp and Syp during nervous system development in Drosophila, in different parts and at different stages, with which the working model of this study aligns.

      Overall, this study adds to and extends current working models and evidence on the developmental mechanisms that underlie temporal cell fate decisions.

      I cannot summarize better the paper.

      Reviewer #1 (Recommendations For The Authors):

      While this is an interesting topic, I raised two issues in my original review.

      (1) Against the backdrop of numerous previous studies linking many developmental regulators, including tTFs, to programmed cell death in the developing CNS, which in several cases have involved identifying key PCD genes and decoding the molecular regulatory interplay between regulators and PCD genes, this study does not provide any new insight into the regulation of developmental PCD in the CNS.

      The authors have not added any new data to address this shortcoming.

      I agree with the reviewer that we did not attempt to link Imp/Syp with the temporal transcription factor (tTF) cascade or spatial selectors such as Hox genes. However, this decision was intentional as our primary focus was on studying immature MNs. It is worth noting that the decommissioning of NBs by autophagic cell death or terminal differentiation, which is mediated by Imp/Syp in other lineages, has not been correlated with tTFs or spatial selectors. Although we have not directly examined the involvement of the hb + sv > kr > pdm > cas > cas-svp > Grh cascade in the decommissioning of the Lin A neuroblast, our preliminary data indicate that Hb, Sv, Pdm, and Cas are not expressed in the Lin A NB, while Grh is consistently expressed in the NB (Wenyue et al., 2022). Thus, it is less likely that this particular tTF cascade is not implicated in Lin A neuroblast decommissioning. In contrast, spatial selectors, such as the Hox gene Antp, play an opposing role compared to HOX transcription factors in abdominal NBs. In the Lin A lineage, Antp promotes survival (Baek, Enriquez, & Mann, 2013). Here, to avoid repeating what has already been described in the literature, we focused on the role of Imp/Syp in postmitotic neurons and revealed that the precise elimination of MNs is linked to the control of TFs expressed in the MNs.

      (2) I raised the issue that it is unclear if Imp/Syp acts in the NB, and/or in IMC/GMC, and/or in the daughter cells generated from these.

      I agree with the reviewer's concern regarding the unclear function of Imp/Syp, i.e., whether it acts in the NB, IMC/GMC, or daughter cells. To address this, one possible approach would be to attempt rescuing Imp and Syp mutants by transgenic expression in specific cell types, such as NBs, IMC/GMC, or GB/daughter cells. However, we have not conducted such experiments as we were skeptical about the outcome. Previous published work has used drivers expressed in NBs, IMC/GMC, or postmitotic neurons to decipher the function of a gene in a specific cell type. But the results of these experiments must be taken with caution. Using NB/GMC drivers to study gene function can lead to effects not only in the NB but also in its progeny, including GMC or postmitotic neurons, due to the perdurance and stability of the Gal4 and UAS-gene expression system. For instance, dpn-Gal4 UASGFP not only labels the NB but also many of its progeny, even if Dpn is only expressed in NBs. And elav-Gal4 is expressed in the NB and GMCs.

      However, our overexpression of Imp in immature neurons using Vglut demonstrates that Imp promotes cell survival through an autonomous function in these neurons. This driver is only expressed in postmitotic neurons (elav+) and not in the NB, IMC/GMC, or in the hemilineage eliminated by cell death (elav-vglut-).

      Reviewer #2 (Recommendations For The Authors):

      Oddly knockdown of Imp in the neuroblast (Fig. 5D) only led to death at 8h APF, when Imp is no longer expressed. Do the authors have an explanation as to how the stem cell can survive until this point? A discussion would be helpful.

      The simple explanation is the efficiency of RNAi. The imp-/- MARCM clones (Guan et al., 2022) lead to a stronger reduction of MNs in LL3.

      A simple experiment I would recommend is to repeat the antibody stainings of staged larvae/pupae (Fig. 4) having the anti-Imp/Syp antibodies in the same brain sample, and perhaps a quantification of the ratio in the NB. Given the species in which the ABs were raised seem compatible, this should be feasible. As it stands now, there is no indication of whether the ratio of Imp vs Syp change over time.

      We have now quantified the transcriptional activity of Imp and Syp in the NB over time. We have also quantified the relative expression of Imp and Syp proteins in the NB over time and quantified the relative expression of Imp and Syp proteins in postmitotic neurons as a function of their birth in late LL3 larvae. How these gradients are established in our system still remains to be 

      Minor errors/suggestions:

      Fig 4. Time legend at the top goes A, B, C, E, F (no D). So it doesn't match the panels below

      Yes, we have made the corrections.

      Sentence repeated in Intro:

      The process of terminating NB neurogenesis through autophagic cell death or terminal differentiation is commonly referred to as decommissioning.

      Yes, corrections have been made.

      IN FIGURE 1 THEY SAY 'TYPE IB' AND IN FIGURE 2 THEY SAY 'TYPE 1B'

      We have changed it to type 1b.

      In Fig2A-It's hard to see lack of Elav and Fig2G-It's hard to see presence of Dcp1. Panels could be adjusted to emphasize these results

      We have increased the size of the panels and made two separate panels where only the elav and Dcp1 signals are present.

      Observations that the result is equivalent in all thoracic segments is expected, since all legs need the same number of neurons. This is nice to have but can be in the supplement.

      Overall the figure number seems excessive, especially considering much of the results included(particularly the NB results) are findings consistent with previous papers and some is characterization of the system that does not fit well with the main focus regarding Imp/Syp (i.e death of one hemi-lineage:

      Figure 5 and 6 can be joined as one.

      We have combined Figures 5 and 6, showing only the T1 segments.

      There is some discrepancy between graphs Fig7F and K: At LL3 the number of neurons is different for the control in 7F and the count in K

      Yes, because the genetic backgrounds are not the same and we are not counting the same type of cells. In 7F, we are counting the elav+ and VGlut+ cells, whereas in Figure 7K, we are counting all the elav+ in Lin A, including those elav+ VGlut-. VGlut expression arrives a bit later after elav+, which is why we have fewer elav+ cells in 7F. In other words, VGlut MARCM clones do not label all Lin A elav+ cells. I have clarified this in the figure.

      Reviewer #3 (Recommendations For The Authors):

      Main comment: on the notion of Imp and Syp gradients:

      p. 5, related to figure 4 - there are clearly distinct windows for predominantly (if not exclusively) Imp, and later, Syp expression in lineage 15, with a phase of co-expression.

      However, based on the data shown, it is unclear whether these windows represent gradients, as repeatedly stated. If the notion of gradients is derived from other studies, on other lineages, then this would be good to clarify. Alternatively, the idea of temporally opposing gradients of Imp and Syp would need to be demonstrated for this lineage.

      For example, a more accurate way to describe this study's data is given on p.7 "In conclusion, our findings demonstrate that the opposite expression pattern of Imp and Syp in postmitotic neurons precisely shapes the size of Lin A/15 lineage by controlling the pattern of PCD in immature MNs (Fig. 8)."

      We have now quantified the transcriptional activity of Imp and Syp in the NB over time. We have also quantified the relative expression of Imp and Syp proteins in the NB over time. We have also quantified the relative expression of Imp and Syp proteins in postmitotic neurons as a function of their birth in late LL3 larvae. How these gradients are established in our system still remains to be identified.

      Minor points:

      p.6, related to figure 7: Are numbers of EDU- early born and EDU+, late born, MNs expressed as means in the main text? As written, it suggests absence of any variability, which one would expect and which is shown in Fig.7 data.

      Yes, we have added averages in the text.

      Methods: the author name 'Lacin' has been mis-spelled

      Sorry about that, it's been corrected.

    1. Author response:

      The following is the authors’ response to the previous reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      The authors investigate the role of chirping in a species of weakly electric fish. They subject the fish to various scenarios and correlate the production of chirps with many different factors. They find major correlations between the background beat signals (continuously present during any social interactions) or some aspects of social and environmental conditions with the propensity to produce different types of chirps. By analyzing more specifically different aspects of these correlations they conclude that chirping patterns are related to navigation purposes and the need to localize the source of the beat signal (i.e. the location of the conspecific).

      The study provides a wealth of interesting observations of behavior and much of this data constitutes a useful dataset to document the patterns of social interactions in these fish. Some data, in particular the high propensity to chirp in cluttered environments, raises interesting questions. Their main hypothesis is a useful addition to the debate on the function of these chirps and is worth considering and exploring further.

      After the initial reviewers' comments, the authors performed a welcome revision of the way the results are presented. Overall the study has been improved by the revision. However, one piece of new data is perplexing to me. The new Figure 7 presents the results of a model analysis of the strength of the EI caused by a second fish to localize when the focal fish is chirping. From my understanding of this type of model, EOD frequency is not a parameter in the model since it evaluates the strength of the field at a given point in time. Therefore the only thing that matters is the phase relationship and strength of the EOD. Assuming that the second fish's EOD is kept constant and the phases relationship is also the same, the only difference during a chirp that could affect the result of the calculation is the potential decrease in EOD amplitude during the chirp. It is indeed logical that if the focal fish decreased its EOD amplitude the target fish's EOD becomes relatively stronger. Where things are harder to understand is why the different types of chirps (e.g. type 1 vs type 2) lead to the same increase in signal even though they are typically associated with different levels of amplitude modulations. Also, it is hard to imagine that a type 2 chirps that is barely associated with any decrease in EOD amplitude (0-10% maybe), would cause doubling of the EI strength. There might be something I don't understand but the authors should provide a lot more details on how this result is obtained and convince us that it makes sense.

      We thank the author for the comments and we agree that the approach could have been better detailed. As anticipated by the Reviewer, the Boundary Element Method (BEM) model can be used simply to calculate the electric field and electric image at a specific point in time (instantaneously), regardless of EOD frequency. However, our model allows for the concatenation of consecutive instants and thus is able to render an entire sequence of electric fields - and resulting electric images - incorporating realistic EOD characteristics such as shape, duration, and frequencies (see Pedraja et al., 2014).

      Chirp-triggered EIs were modeled using real chirps produced by interacting fish. Each chirp was thus associated to its duration and peak parameters, as well as the fish positional information (distance and angle). 

      However, since we did not know the beat phase at which chirps were produced, we computed electric images for each fish position and chirp scenario by simulating various phases (here referred to the initial offset of the two EODs, set at 4 phases, equally spaced). These are intended as phases of the sender EOD and simply refer to the initial OFFSET between the two interacting EODs. However, since our simulations were run over a time window of 500 msec, all phases are likely to be covered, with a different temporal order relative to the chirp (always centered within the 500 msec).

      The simulation was run maintaining consistent timing for both chirp and non-chirp conditions, across approximately 800 body nodes. At each node, the current flow was calculated from the peak-to-peak of the EOD sum (i.e. the point-to-point of the difference between the beat positive and negative envelopes). Analyzing the EIs over this fixed time window enables us to assess the unitary changes of current flow induced by chirps over units of time (ΔI/Δt). From this, we can calculate a cumulative sum of current flow changes - expressed as delta(EI) and use it to show the effect of the chirps on the spatiotemporal EI (Figure 7C).

      One can express this cumulative change mapped onto the fish body (keeping the 800 points separated, as in Figure 7C) or further sum the current changes to obtain a single total (as shown in Figure 7D).

      One can check this by considering that a sum for example of a set of 500/800 points - judging from the size of the blue areas in C not all 800 points have a detectable change - each valued 0.1-to-0.3 mA/s, one could get circa 100 mA/s, which is what is shown in D. (is this what is happening ?)

      We do not know why chirps of different types triggered similar effects. It is possible that, since EI measurements are pooled over several chirps produced at different angles and distances, in case of a lower amount of chirps considered for a given type (as in the case of rises, very low) these measurements may not highlight more marked differences among types. In a publication we are currently working on, we are considering a larger dataset to better assess these results.

      The methods section has been edited to clarify the approach (not yet).

      Reviewer #2 (Public Review):

      Studying Apteronotus leptorhynchus (the weakly electric brown ghost knifefish), the authors provide evidence that 'chirps' (brief modulations in the frequency and amplitude of the ongoing electric signal) function in active sensing (specifically homeoactive sensing) rather than communication. Chirping is a behavior that has been well studied, including numerous studies on the sensory coding of chirps and the neural mechanisms for chirp generation.

      Chirps are largely thought to function in communication behavior, so this alternative function is a very exciting possibility that could have a great impact on the field.

      We thank the Reviewer for the extensive and constructive comments. We would like to add that, while it is true that many detailed studies have been published on the anatomy and physiology of the circuits implicated in the production and modulation of “electric chirps”, most of this  research assumed, and focused exclusively on, their possible role in communication.  In addition, most behavioral studies did the same and a meta-analysis of the existing literature on chirping allows to trace back the communication idea mainly to two studies: Hagedorn and Heiligenberg, 1985 (“Court and spark: electric signals in the courtship and mating of gymnotoid fish”) and Hopkins, 1974 (“Electric Communication: Functions in the Social Behavior of Eigenmannia Virescens”), among the main sources. Importantly, in these studies only contextual observations have been made (no playback experiment or other attempts to analyze more quantitatively the correlation of chirping with other behaviors).

      The authors do provide convincing evidence that chirps may function in homeoactive sensing. However, their evidence arguing against a role for chirps in communication is not as strong, and fails to sufficiently consider the evidence from a large body of existing research. Ultimately, the manuscript presents very interesting data that is sure to stimulate discussion and follow-up studies, but it suffers from dismissing evidence in support of, or consistent with, a communicative function for chirps.

      Although the tone of some statements present in our earlier draft may suggest otherwise, through our revisions, we have made an effort to clarify that we do not intend to dismiss a function of chirps in communication, we only intend to debate and discuss valid alternative hypothesis, advanced from reasonable considerations.

      Before writing this manuscript, we have attempted to survey  literally all the existing literature on chirps (including studies focused on behavior, peripheral sensory physiology as well as brain physiology). Although it is not unlikely that some studies have eluded our attention, an effort for a comprehensive review was made. Based on this survey we realized that none of the studies provided a clear  and  unambiguous piece of evidence to support the communication hypothesis (we refer here to the weak points highlighted in the discussion and mentioned in the previous comment). Which in fact does not come without its weak points and contradictions (see later comments).

      It follows a summary of the mentions made to the communication theory in the different section of the manuscript including several edits we have applied in response to the Reviewer’s concern:

      In the abstract we clearly state that we are considering an alternative that is only hypothetically complementary, not for sure.  Nonetheless, we have identified a couple of instances that could sound dismissive of the “communication hypothesis” in the following section.

      In the introduction we write in fact about the possibility of interference between communication signals and conspecific electrolocation cues, as they are both detected as beat perturbations. We did not mean to use “Interference” here as “reciprocal canceling”, rather we intended it as “partial or more or less conspicuous overlap” in the responses triggered in electroreceptors.

      Hoping to convey a clearer message, we have edited the related statement and changed it to “both types of information are likely to overlap and interact in highly variable ways”.

      We have also removed the statement: “According to this idea, beats and chirps are not only detected through the same input channel, but also used for the same purpose.” as at this point in the manuscript it may be too strong.

      In the results section we do not include statements that might be seen as dismissive of the communication hypothesis but only statements in support of the “probing with chirps” idea (which is the central hypothesis of the study).

      In the discussion paragraphs we elaborate on why the current functional view is either flawed or incomplete (first paragraph “existing functional hypotheses''). Namely: 1)  multiple triggering factors implied in chirp responses covary and need to be disentangled (example DF/ sex), 2) findings on brown ghosts and a few other gymnotiforms have been used to advance the hypothesis of “communication through chirps'' in all weakly electric fish (including pulse species). 3) social encounters - in which chirps are recorded - imply also other behaviors (such as probing) which have not been considered so far. This point is related to the first one on covariates. 4) most studies referring to big chirps as courtship chirps were not done in reproductive animals (added now)  and 5) no causal evidence has been provided so far to justify a role of chirps in social communication.

      We are discussing these points as challenges to the communication hypothesis, not to dismiss the hypothesis, but rather to motivate future studies addressing these challenges.

      We do not want to appear dismissive of the communication hypothesis and had therefore previously edited the manuscript to avoid the impression of exclusivity of the probing hypothesis. We have now gone over the manuscript once more and edited several sentences. Nevertheless, we want to point out again that - despite the large consensus - the communication hypothesis has, until now, never been investigated with the kind of rigor applied here.

      The authors do acknowledge that chirps could function as both a communication and homeactive sensing signal, but it seems clear they wish to argue against the former and for the latter, and the evidence is not yet there to support this.

      In both rounds of revision we have made an effort to convey a more inclusive interpretation of our findings. We tried our best to express our ideas as hypothetical, not as proof that communication through chirps does not exist. The aim of this study is to propose an alternative view, and this cannot be done without underlining the weak points of an existing hypothesis while providing and supporting reasonable arguments in favor of the alternative we advance. The actual evidence for a role of chirping in communication is much less strong than appears from the pure number of articles that have discussed chirps in this context.

      Regarding the weak evidence against communication, here we can list a few additional important points related to the proposed interpretations of chirp function (more specific than those made earlier):

      (1) A formally sound assessment of signal value/meaning - as typically done in animal communication studies should involve: 

      a) the isolation of a naturally occurring signal and determination of the context in which it is produced 

      b) the artificial replication of the signal

      c) the observation that such mimic is capable of triggering reliable and stereotyped responses in a group of individuals (identified by sex and/or species) under the same conditions (conditioned, unconditioned, state-dependent, etc.). As discussed for instance in Bradbury and Vehrencamp, 2011; Laidre and Johnstone, 2013; Wyatt, 2015; Rutz et al., 2023.

      This approach has so far not been applied to weakly electric fish. The initial purpose of the present study was in fact to conduct this type of validation.

      (2) The hypothesis of chirps used for DF-sign discrimination - for “social purposes” - although plausible in the face of theoretical considerations,  does not seem to be reasonable in practice, when one considers emission rates of 150 chirps per minute. We do find a strong correlation of chirp type with DF, which is often very abrupt and sudden (as if the fish were tracking beat frequency to guess its value) but the consideration made above on chirp rates seems to discourage this interpretation.

      (3) The hypothesis of chirp-patterning (i.e. chirping may have meaning based on the sequence of chirps of different types, a bit like syllables in birdsongs) - assessed by only one study conducted in our group - has not been enough substantiated by replication. We have surveyed all possible combinations of chirps produced by interacting pairs in different behavioral conditions using different value for chirp sequence size: 2, 3,... ,8 chirps (both considering the sender alone as well as sender+receiver together). In all cases we found no evidence for  a context dependent “modulation” of chirp types (i.e. no specific chirp type sequence in specific contexts).

      (4) The hypothesized role of “large chirps” as courtship signals could be easily criticized by noting the symmetrical distribution of these events around  a DF of 0 Hz . Although one could argue about a failure to discriminate DF-sign, to explain this well known pattern. However, we know from Walter Heiligenberg’s work and physiological considerations that such task can be solved easily through t-units and … in principle even just by motion (which would change the EOD phase in frequency dependent ways, thus potentially revealing the DF sign).

      Overall, these considerations made us think that certainly chirping occurs in a social context, but it is the meaning of this behavior that remains elusive.  We noticed that environmental factors are also strongly implied … we then formulate an alternative hypothesis to explain chirping but we do so  without dismissing the communication idea.

      All this seems to us just a careful way to critically discuss our results and those of other studies, without considering the issue resolved.

      In the introduction, the authors state, "Since both chirps and positional parameters (such as size, orientation or motion) can only be detected as perturbations of the beat, and via the same electroreceptors, the inputs relaying both types of information are inevitably interfering." I disagree with this statement, which seems to be a key assumption. Both of these features certainly modulate the activity of electroreceptors, but that does not mean those modulations are ambiguous as to their source. You do not know whether the two types of modulations can be unambiguously decoded from electroreceptor afferent population activity.

      We thank the Reviewer for noting this imprecision. We have addressed the Reviewer’s concern in another reply (see above).

      My biggest issue with this manuscript is that it is much too strong in dismissing evidence that chirping correlates with context. In your behavioral observations, you found sex differences in chirping as well as differences between freely interacting and physically separated fish. Chirps tended to occur in close proximity to another fish. Your model of chirp variability found that environmental experience, social experience, and beat frequency (DF) are the most important factors explaining chirp variability. Are these not all considered behavioral or social context? Beat frequency (DF) in particular is heavily downplayed as being a part of "context" but it is a crucial part of the context, as it provides information about the identity of the fish you're interacting with. The authors show quite convincingly that the types of chirps produced do not vary with these contexts, but chirp rates do.

      We believe the “perceived claim” may be an issue of unclear writing. We have now tried to better clarify that “context” affects chirp rates, but it does not affect chirp types as much (except when beat frequency is high).  

      We have edited two statements possibly susceptible to misinterpretation: 

      (1) In the results: “It also indicates that chirp parameters such as duration and FM do not seem to be associated with any particular context in a meaningful way, other than being affected by beat frequency.”

      (2) In the discussion: the statement

      “Recordings from interacting fish pairs confirmed the absence of any significant correlation between chirp type choice and behavioral context (Figure S2) although the variance of chirp parameters appears to be significantly affected by this factor (Figure 2). This may suggest that the effect of behavioral context is mainly detectable in the number of chirps produced (Figure S1), rather than the type (Figure S2).”

      has been changed to:

      “Recordings from interacting fish pairs confirmed the absence of any significant correlation between chirp type choice and behavioral context, except for those cases characterized by higher beat frequencies  (Figure S2). This suggests that the effect of behavioral context highlighted in our factor analysis (Figure 2) is mainly due to the number of chirps produced (Figure S1), rather than their type (Figure S2).”

      Eventually, in the results we emphasize the relatively higher impact of previously unexplored factors on chirp variance: “The plot of individual chirps (Figure 2C) shows the presence of clustering around different categorical variables and it reveals that experience levels or swimming conditions are important factors affecting chirp distribution (note for instance the large central “breeding” cluster in which fish are divided and the smaller ones in which fish are free). Sender or receiver identity does not individuate any clear clustering relative to either sex (see the overlap of male_s/male_r and female_s/female_r) or social status (dominant/subordinate). Chirps labeled based on tank experience (i.e. resident vs intruder) are instead clearly separated.”.

      Further, in your playback experiments, fish responded differently to small vs. large DFs, males chirped more than females, type 2 chirps became more frequent throughout a playback, and rises tended to occur at the end of a playback. These are all examples of context-dependent behavior.

      We do note that male brown ghosts chirp more than females. But we do also say - and show in figure 8 - that males move more in proximity to and around conspecifics. We do acknowledge that chirp time-course may be different during playbacks in a type-dependent manner. But how this can support the communication hypothesis - or other alternatives - is unclear. This result could equally imply the use of different chirp types for different probing needs. Since we cannot be sure about either, we do not want to put too much emphasis to it. Eventually, the fact that “context” (here meant broadly to define different experimental situations in which social but also physical and environmental parameters are altered) affects chirping is undeniable: cluttered and non-cluttered environments do represent different contexts which differently affect chirping in conspicuous ways.

      In the results, the authors state, "Overall, the majority of chirps were produced by male subjects, in comparable amounts regardless of environmental experience (resident, intruder or equal; Figure S1A,C), social status (dominant or subordinate; Figure S1B) or social experience (novel or experienced; Figure S1D)." This is not what is shown in Figure S1. S1A shows clear differences between resident vs. intruder males, S1B shows clear differences between dominant vs. subordinate males, and S1D shows clear differences between naïve and experienced males. The analysis shown in Figure 2 would seem to support this. Indeed, the authors state, "Overall, this analysis indicated that environmental and social experience, together with beat frequency (DF) are the most important factors explaining chirp variability."

      The Reviewer is right in pointing at this imprecise reference and we are grateful for spotting this incongruence. The writing refers probably to an earlier version of the figure in which data were grouped and analyzed differently. We now edited the text and changed it to: “Overall, the majority of chirps were produced by male subjects, at rates that seemed  affected by environmental experience (resident, intruder or equal; Figure S1A,C), social status (dominant or subordinate; Figure S1B) and social experience (novel or experienced; Figure S1D).”

      The choice of chirp type varied widely between individuals but was relatively consistent within individuals across trials of the same experiment. The authors interpret this to mean that chirping does not vary with internal state, but is it not likely that the internal states of individuals are stable under stable conditions, and that individuals may differ in these internal states across the same conditions? Stable differences in communication signals between individuals are frequently interpreted as reflecting differences between those individuals in certain characteristics, which are being communicated by these signals.

      It seems here we have been unclear in the writing: while it is true that behavioral states are stable and can imply stable chirp patterning (if the two are related), since chirp types vary abruptly and in a reliable DF-dependent manner, different types of chirps are unlikely to be matched to different internal states following the same temporal order in such a reliable way (similarly repeated through consecutive trials).

      This would imply the occurrence of different internal states in rapid sequence, reliably triggered by repeated EOD ramps, regardless of whether the playback is 20 sec long or 180 sec long.

      We have edited this paragraph to better explain this: “The reliability by which the chirping response adapts to both the rate and direction of beat frequency is variable across individuals but rather stable across trials (relative to a given subject), further suggesting that chirp type variations may not reflect changes in internal states or in the animal motivation to specific behavioral displays (which are presumably subject to less abrupt variations and stereotypical patterning based on DF).”

      I am not convinced of the conclusion drawn by the analysis of chirp transitions. The transition matrices show plenty of 1-2 and 2-1 transitions occurring.

      The only groups in which 1-2 and 2-1 transitions are as frequent as 1-1 and 2-2 (being 1 and 2 the numerical IDs of the two interacting fish) are F-F pairs. This is a result of the fact that in females chirp rates are so low that within-fish-correlations end up being as low as between-fish-correlations. We believe the impression of the Reviewer could be due to the fact that these are normalized maps (see legend of Figure 5A-B).

      Further, the cross-correlation analysis only shows that chirp timing between individuals is not phase-locked at these small timescales. It is entirely possible that chirp rates are correlated between interacting individuals, even if their precise timing is not.

      We agree with the Reviewer, this is a possibility. To address this point, we did edit the results section to acknowledge that what we see may be related to the time window chosen (i.e. 4 sec):

      “More importantly, they show that - at least in the social conditions analyzed here and within small-sized time windows - chirp time series produced by different fish during paired interactions are consistently independent of each other.”

      Further, it is not clear to me how "transitions" were defined. The methods do not make this clear, and it is not clear to me how you can have zero chirp transitions between two individuals when those two individuals are both generating chirps throughout an interaction.

      We thank the Reviewer for bringing up this unclear point. We have now clarified how transitions were calculated in the method section: “The number of chirp transitions present in each recording (dataset used for Figures 1, 2, 5) was measured by searching in a string array containing the 4 chirp types per fish pair, all their possible pairwise permutations (i.e. all possible permutations of 4+4=8 elements are: 1-1, 1-2, 1-3 … 7-6, 7-7, 7-8; considering the following legend 1 = fish1 type 1, 2 = fish 1 type 2, 3 = fish1 type 3 … 6 = fish2 type 2, 7 = fish2 type 3 and 8 = fish2 rise).”.

      Zero transitions are possible if two fish (or groups of fish) do not produce chirps of all types. Only transitions of produced types can be counted.

      In the results, "Although all chirp types were used during aggressive interactions, these seemed to be rather less frequent in the immediate surround of the chirps (Figure 6A)." A lack of precise temporal correlation on short timescales does not mean there is no association between the two behaviors. An increased rate of chirping during aggression is still a correlation between the two behaviors, even if chirps and specific aggressive behaviors are not tightly time-locked.

      The Reviewer is right in pointing out the limited temporal scaling of our observations/analysis. We have now edited the last paragraph of the results related to figure 6 to include the possibility mentioned by the Reviewer: “The significantly higher extent of chirping during swimming and locomotion, consistently confirmed by 4 different approaches (PSTH, TM, CN, MDS), suggests that - although chirp-behavior correlations may exist at time-scales larger than those here considered - chirping may be linked more strongly with scanning and environmental exploration than with a particular motivational state, thus confirming findings from our playback experiments.”

      The Reviewer here remarks an important point, yet, due to space limitations, we have considered only a sub-second scale. Most playback experiments in weakly electric fish implied the use of EOD mimics for a few tens of seconds - to avoid habituation in the fish behavioral responses -  while inter-chirp intervals usually range between a few hundreds of milliseconds to seconds (depending on how often a fish would chirp). This suggested to us that a 4 second time window may not be a bad choice to start with.

      In summary, it is simply too strong to say that chirping does not correlate with context, or to claim that there is convincing evidence arguing against a communication function of chirps. Importantly, however, this does not detract from your exciting and well-supported hypothesis that chirping functions in homeoactive sensing. A given EOD behavior could serve both communication and homeoactive sensing. I actually suspect this is quite common in electric fish (both gymnotiforms and mormyrids), and perhaps in other actively sensing species such as echolocating animals. The two are not mutually exclusive.

      We agree with the Reviewer that context - broadly speaking - does affect chirping (as we mentioned above). We hope we have improved the writing and clarified that we do not dismiss communication functions of chirping, but we do lean towards electrolocation based on the considerations above made and our results.

      We do conclude the manuscript remarking that communication and electrolocation are not mutually exclusive: ”probing cues could function simultaneously as proximity signals to signal presence, deter approaches, or coordinate behaviors like spawning, if properly timed (Henninger et al., 2018).” (see the conclusion paragraph of the discussion) .

      Therein, we further add “These findings aim to stir the pot and initiate a discussion on possible alternative functions of chirps beyond their presumed communication role.”.

      With this, we hope we’ve made it clear how we intend our manuscript to be read.

      Reviewer #3 (Public Review):

      Summary:

      This important paper provides the best-to-date characterization of chirping in weakly electric fish using a large number of variables. These include environment (free vs divided fish, with or without clutter), breeding state, gender, intruder vs resident, social status, locomotion state and social and environmental experience, without and with playback experiments. It applies state-of-the-art methods for reducing the dimensionality of the data and finding patterns of correlation between different kinds of variables (factor analysis, K-means). The strength of the evidence, collated from a large number of trials with many controls, leads to the conclusion that the traditionally assumed communication function of chirps may be secondary to its role in environmental assessment and exploration that takes social context into account. Based on their extensive analyses, the authors suggest that chirps are mainly used as probes that help detect beats caused by other fish and as well as objects.

      Strengths:

      The work is based on completely novel recordings using interaction chambers. The amount of new data and associated analyses is simply staggering, and yet, well organized in presentation. The study further evaluates the electric field strength around a fish (via modelling with the boundary element method) and how its decay parallels the chirp rate, thereby relating the above variables to electric field geometry.

      The main conclusions are that the lack of any significant behavioural correlates for chirping, and the lack of temporal patterning in chirp time series, cast doubt on a primary communication goal for most chirps. Rather, the key determinants of chirping are the difference frequency between two interacting conspecifics as well as individual subjects' environmental and social experience. The paper concludes that there is a lack of evidence for stereotyped temporal patterning of chirp time series, as well as of sender-receiver chirp transitions beyond the known increase in chirp frequency during an interaction.

      These conclusions by themselves will be very useful to the field. They will also allow scientists working on other "communication" systems to perhaps reconsider and expand the goals of the probes used in those senses. A lot of data are summarized in this paper, with thorough referencing to past work.

      The alternative hypotheses that arise from the work are that chirps are mainly used as environmental probes for better beat detection and processing and object localization, and in this sense are self-directed signals. This led to their prediction that environmental complexity ("clutter") should increase chirp rate, which is fact was revealed by their new experiments. The authors also argue that waveform EODs have less power across high spatial frequencies compared to pulse-type fish, with a resulting relatively impoverished power of resolution. Chirping in wave-type fish could temporarily compensate for the lower frequency resolution while still being able to resolve EOD perturbations with a good temporal definition (which pulse-type fish lack due to low pulse rates).

      The authors also advance the interesting idea that the sinusoidal frequency modulations caused by chirps are the electric fish's solution to the minute (and undetectable by neural wetware) echo-delays available to it, due to the propagation of electric fields at the speed of light in water. The paper provides a number of experimental avenues to pursue in order to validate the non-communication role of chirps.

      We thank the reviewer for the kind assessment.

      Weaknesses:

      My main criticism is that the alternative putative role for chirps as probe signals that optimize beat detection could be better developed. The paper could be clearer as to what that means precisely, especially since beating - and therefore detection of some aspects of beating due to the proximity of a conspecific - most often precedes chirping. One meaning the authors suggest, tentatively, is that the chirps could enhance electrosensory responses to the beat, for example by causing beat phase shifts that remediate blind spots in the electric field of view.

      We agree with the Reviewer that a better and more detailed explanation of how beat processing for conspecific electrolocation may be positively affected by chirps would be important to provide. We are currently working on a follow-up manuscript in which we intend to include these aspects. For space limitations and readability we had to discard from the current manuscript a lot of results that could further clarify these issues.

      A second criticism is that the study links the beat detection to underwater object localization. The paper does not significantly develop that line of thought given their data - the authors tread carefully here given the speculative aspect of this link. It is certainly possible that the image on the fish's body of an object in the environment will be slightly modified by introducing a chirp on the waveform, as this may enhance certain heterogeneities of the object in relation to its environment. The thrust of this argument derives mainly from the notion of Fourier analysis with pulse type fish EOD waveforms (see above, and radar theory more generally), where higher temporal frequencies in the beat waveform induced by the chirp will enable a better spatial resolution of objects. It remains to be seen whether experiments can show this to be significant.

      Perhaps the Reviewer refers to the last discussion paragraph before the conclusions in which we mention the performance of pulse or wave-type EODs in electrolocation (referring here to ideas illustrated in a recent review by Crampton, 2019). We added to this paragraph a statement which could better clarify that we do not propose that chirping could enhance object electrolocation. What we mean is that, in a context in which object electrolocation occurs through wave-type EODs - given the generally lower performance of such narrow-band signals in resolving the spatial features of any object, even a 3D electric field  - chirping could improve beat detection during social encounters by increasing the amount of information obtained by the fish.

      The edited paragraph now reads: “While broadband pulse signals may be useful to capture highly complex environments rich in foliage, roots and other structures common in vegetation featuring the more superficial habitats in which pulse-type fish live, wave-type EODs may be a better choice in the relatively simpler river-bed environments in which many wave-type fish live (e.g., the benthic zone of deep river channels; Crampton, 2019). In this case, achieving a good spatial resolution is critical during social encounters, especially considering the limited utility of visual cues in these low-light conditions. In such habitats, social encounters may “electrically” be less “abrupt”, but spatially less “conspicuous” or blurred (as a 3D electric field may be). In such a scenario, chirps could serve as a means to supplement the spatial information acquired via the beat, accentuating these cues during periods of reduced resolution.”

      Recommendations for the authors:

      Reviewer #3 (Recommendations For The Authors):

      None, my points in the original review have been properly addressed in this resubmission.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Review):  

      Summary: 

      In this study, Nandy and colleagues examine neural and behavioral correlates of perceptual variability in monkeys performing a visual change detection task. They used a laminar probe to record from area V4 while two macaque monkeys detected a small change in stimulus orientation that occurred at a random time in one of two locations, focusing their analysis on stimulus conditions where the animal was equally likely to detect (hit) or not-detect (miss) a briefly presented orientation change (target). They discovered two behavioral measures that are significantly different between hit and miss trials - pupil size tends to be slightly larger on hits vs. misses, and monkeys are more likely to miss the target on trials in which they made a microsaccade shortly before target onset. They also examined multiple measures of neural activity across the cortical layers and found some measures that are significantly different between hits and misses. 

      Strengths: 

      Overall the study is well executed and the analyses are appropriate (though multiple issues do need to be addressed). 

      We thank the reviewer for their enthusiasm and their constructive comments which we address below.

      Weaknesses: 

      My main concern with this study is that with the exception of the pre-target microsaccades, the physiological and behavioral correlates of perceptual variability (differences between hits and misses) appear to be very weak and disconnected. Some of these measures rely on complex analyses that are not hypothesis-driven and where statistical significance is difficult to assess. The more intuitive analysis of the predictive power of trial outcomes based on the behavioral and neural measures is only discussed at the end of the paper. This analysis shows that some of the significant measures have no predictive power, while others cannot be examined using the predictive power analysis because these measures cannot be estimated in single trials. Given these weak and disconnected effects, my overall sense is that the current results do not significantly advance our understanding of the neural basis of perceptual variability. 

      Reviewer #1 (Recommendations For The Authors): 

      (1) Most of the effects are very small. For example, the difference in pupil size between hits and misses is ~0.08 z-score units. The differences in firing rates between hits and misses are in the order of 1-2% of normalized firing rates. While these effects may be significant, their contribution to perceptual variability could be negligible, as suggested by the analysis of predictive power at the end of the result section. On a related note, it would be useful to mention the analysis of predictive power earlier in the paper. The finding that some of the measures do not have significant predictive power w/r to behavioral outcome raises questions regarding their importance. Finally, it would strengthen the paper if the authors could come up with methods to assess the predictive power of the PPC and interlaminar SSC. Without such analyses, it is difficult to assess the importance of these measures. 

      We expect that relatively small differences in early to intermediate sensory areas could cumulatively result in large differences in higher areas and contribute to the binary distinction between hits and misses. We certainly do not claim that these results completely explain state-dependent differences that determine the outcome of these trials. Instead, we have focused on neural signatures at the level of the V4 columnar microcircuit that might ultimately contribute to the variability in perception.

      We would like to emphasize that, based on the reviewer’s recommendation, we have now analyzed our results separately for each animal (see below). The consistency and significance of our findings across both animals give us confidence that what we have reported here are important neural signatures underlying perceptual variability at threshold.

      We would also like to note that SSC and PPC are now part of the standard toolkit of systems neuroscience and have been employed in numerous studies to our knowledge. While all measures come with their set of caveats and limitations, these two measures provide a frequency-resolved metric of the relationship between two temporal processes (point or continuous), which we believe provide insights into the interlaminar flow of information that we report here.

      Unfortunately, limitations in the GLM method and the reliability of these analyses with limited data make it impossible for these two measures to be included. The GLM requires all variables to be defined for each trial in the input. SSC and PPC can be undefined at low firing rates and require a substantial amount of data to be reliably calculated. While we did consider imputing data or estimating SSC and PPC using multiple trials, we ultimately did not pursue this idea as the purpose of the GLM was to use simultaneous measurements from single trials. 

      (2) What is the actual predictive power of the GLM model (i.e., what is the accuracy of predicting whether a given held-out trial will lead to a hit or a miss)? How much of this predictive power is accounted for by the effect of microsaccades? 

      As the GLM is not a decoder, it does not classify whether a given left out trial will be a hit or a miss. However, the GLM was highly predictive compared to a constant model. This information has been added to Table 3. The deviance of the GLM with and without microsaccades as a variable was not significantly different (p >0.9).  

      (3) The role of stimulus contrast is not explained clearly. Are all the analyses and figures restricted to a single contrast level? Was the contrast the same on both sides? If multiple contrasts are used, could contrast account for some of the observed neural-behavioral covariations? 

      All of the analyses include stimuli of all tested contrast levels. Stimulus contrasts were the same at both locations (attended and unattended). We have added a more detailed description of the contrast in hit and miss trials (Lines 289-296 and reproduced that here: 

      “Non-target stimulus contrasts were slightly different between hits and misses (mean:

      33.1% in hits, 34.0% in misses, permutation test, 𝑝 = 0.02), but the contrast of the target was higher in hits compared to misses (mean: 38.7% in hits, 27.7% in misses, permutation test, 𝑝 = 1.6 𝑒 − 31). Firing rates were normalized by contrast in Figure 3. In all other figures, we considered only non-target stimuli, which had very minor differences in contrast (<1%) across hits and misses. While we cannot completely rule out any other effects of stimulus contrast, the normalization in Figure 3 and minor differences for non-target stimuli should minimize them.”

      (4) Do the animals make false alarms (i.e., report seeing a target in non-target epochs)?

      If not, then it is not clear that the animals are performing near their perceptual threshold. If the false-alarm rate is non-zero, it should be reported and analyzed for neural/behavioral correlates. Does the logistic regression fit allow for a false alarm rate? More generally, it would be useful to see a summary of behavioral performance, such as distribution of thresholds, lower and upper asymptotes, and detection rates on foil trials vs. matched target trials. 

      The logistic regression does allow for a false alarm rate. We have reported additional behavioral parameters in Figure 1-figure supplement 3A-G.  

      (5) As far as I can tell, all the analyses in the paper are done on data combined across the two animals. Given that these effects are weak and that the analyses are complex, it is important to demonstrate for each analysis/figure that the results hold for each animal separately before combining the data across animals. This can be done in supplementary figures. 

      We have updated the paper to include all main results plotted separately for each animal as supplementary figures. 

      - Figure 2-figure supplement 2

      - Figure 3-figure supplement 1

      - Figure 3-figure supplement 2

      - Figure 4-figure supplement 1

      - Figure 5-figure supplement 2

      - Figure 7-figure supplement 1

      All the results except for the canonical correlation analysis were present, consistent, and significant when we analyzed them in each monkey independently.

      (6) The selection of the temporal interval used for the various analyses appears somewhat post hoc and is not explained clearly. Some analyses are restricted to the period immediately before or during target onset (e.g., 400 ms before target onset for analysis of the effect of microsaccade, 60 ms before stimulus onset for the analysis of the effect of neural variability). Other analyses are done on non-target rather than target stimuli. What is the justification for selecting these particular periods for these analyses? The differences in firing rates between hits and misses are restricted to the target epoch and are not present in the non-target epochs. Given these results, it seems important to compare the effects in target and non-target epochs in other analyses as well.

      Restricting the analysis of the Fano Factor to 60 ms before non-target onset seems odd. Given that the duration of the interval between stimulus presentations is random, how could this pre-stimulus effect be time-locked to target onset? 

      We selected a 200ms time window during the pre-stimulus or stimulus-evoked period for almost all our analyses. The results relating to microsaccade occurrence were robust to narrower time windows more consistent with the other pre-stimulus windows we used, but we chose to use the 400ms window to capture a larger fraction of trials with microsaccades. 

      Only the Fano factor time window was selected post-hoc based on the traces in Figure 4A, and the result is robust across animals (new Figure 4-figure supplement 1). The inter-stimulus intervals are random, and we do not believe the neural variability is timelocked to upcoming stimuli, but that lower variability in this pre-stimulus window is characteristic of hits. 

      We believe that the consistency of our results across both animals provides further evidence that our time window selection was appropriate. 

      We are interested in the extent to which these effects would remain consistent when applied only to target stimuli. However, restricting our analyses to only target stimuli substantially reduces the amount of neural data available for analysis. We plan to explore target stimulus representation more thoroughly in future studies.   

      (7) Can the measured neural response be used to discriminate between target and nontarget stimuli? If so, is the discriminability between target and non-target higher in hits vs. misses? 

      Thank you for raising this interesting point. We performed this analysis and find that target stimuli are more discriminable from non-targets in hits compared to misses. This has been added as a new Figure 3A.  

      (8) How many trials were performed per session? Did miss probability tend to increase over time over the session? If so, could this slow change in hit probability account for some of the observed neural and behavioral correlations with perceptual decisions? 

      Monkeys initiated a median of 905 trials (range of 651 to 1086). This has been added to the manuscript (Line 106). Approximately 1/8 of those trials were at perceptual threshold. Hit probability at threshold does not change substantially over the course of the session. We now report this in new Figure 1- figure Supplement 3I (error bars show standard deviation). 

      (9) Did miss probability depend on the time of the change within the trial? If so, do any of the behavioral/neural metrics share a similar within-trial time course? 

      Change times were not significantly different across hit and miss trials (p=0.15, Wilcoxon rank sum test). We now report this in new Figure 1-figure supplement 3H.

      (10) "Deep layer neurons exhibit reduced low-frequency phase-locking in hit trials than in misses (Figure 5B), suggesting an improvement in pooled signal-to-noise among this neural population." - why does this metric suggest improved SNR? Is there any evidence for improved SNR in the data? Why just in deep layers? 

      Thank you for raising this question. We agree this statement is not fully supported by the data and have removed it.  

      (11) I may have missed this but what were the sizes of the Gabor stimuli? 

      This has been added to the methods section (Line 454). The Gaussian halfwidth was 2 degrees.  

      Reviewer #2 (Public Review):  

      In this manuscript, the authors conducted a study in which they measured eye movements, pupil diameter, and neural activity in V4 in monkeys engaged in a visual attention task. The task required the monkeys to report changes in the orientation of Gabors' visual stimuli. The authors manipulated the difficulty of the trials by varying the degree of orientation change and focused their analysis on trials of intermediate difficulty where the monkeys' hit rate was approximately 50%. Their key findings include the following: 1) Hit trials were preceded by larger pupil diameter, reflecting higher arousal, and by more stable eye positions; 2) V4 neurons exhibit larger visual responses in hit trials; 3) Superficial and deep layers exhibited greater coherence in hit trials during both the pre-target stimulus period and the non-target stimulus presentation period. These findings have useful implications for the field, and the experiments and analyses presented in this manuscript validly support the authors' claims. 

      Strengths: 

      The experiments were well-designed and executed with meticulous control. The analyses of both behavioural and electrophysiological data align with the standards in the field. 

      We thank the reviewer for their enthusiasm about our study and their constructive comments which we address below.

      Weaknesses: 

      Many of the findings appear to be incremental compared to previous literature, including the authors' own work. While incremental findings are not necessarily a problem, the manuscript lacks clear statements about the extent to which the dataset, analysis, and findings overlap with the authors' prior research. For example, one of the main findings, which suggests that V4 neurons exhibit larger visual responses in hit trials (as shown in Fig. 3), appears to have been previously reported in their 2017 paper. Additionally, it seems that the entire Fig1-S1 may have been reused from the 2017 paper. These overlaps should have been explicitly acknowledged and correctly referenced. 

      While the raw data used in this paper overlaps entirely with Nandy et al. (2017), all the analyses and findings in this manuscript are new and have not been previously reported. Figure 1-figure supplement 1 is modified and reproduced from that paper only to allow readers to understand the recording methods used to collect the data without needing to go back to the previous paper. We have added an explicit acknowledgment of this to the figure caption.

      Previous studies have demonstrated that attention leads to decorrelation in V4 population activity. The authors should have discussed how and why the high coherence across layers observed in the current study can coexist with this decorrelation. 

      We have updated the discussion section (Lines 347-351) to further elaborate on this interpretation. 

      Furthermore, the manuscript does not explore potentially interesting aspects of the dataset. For instance, the authors could have investigated instances where monkeys made 'false' reports, such as executing saccades towards visual stimuli when no orientation change occurred. It would be valuable to provide the fraction of the monkeys' responses in a session, including false reports and correct rejections in catch trials, to allow for a broader analysis that considers the perceptual component of neural activity over pure sensory responses. 

      We appreciate this feedback. While we agree these are interesting directions, we decided to limit the scope of this study to only focus on trials at threshold with an orientation change, and are considering these directions for future studies. 

      Reviewer #2 (Recommendations For The Authors): 

      • Figure Design: Since eLife does not impose space limitations, it is advisable for the authors to avoid using very small font sizes. Consistency in font size throughout the figures is recommended. Some figures are challenging to discern, for example, the mean+-sem in Fig. 2B, and the alpha values of green and purple colours for superficial/deep layers are too high, making them too transparent or pale. 

      We have increased the size of some small fonts and improved font size consistency throughout the figures. We have changed the layer colors to improve legibility. 

      • Line 119: trail, 

      This has been fixed.

    1. we're told they're not supposed to do is trespass in the capital

      Exactly, it is illegal to trespass in the capitol. I've been in the capitol, and you're only allowed there under very controlled circumstances, when you're with a tour, or when you have official business and an appointment with congressional staff. You have to go through security, etc. So, the protestors who broke in were trespassing and breaking the law. To argue that after the doors and windows were beaten down that they were "invited in" by security guards that were scared shitless at that point is absolute BS.

    1. 完全シャットダウンの実行

      何をやってもダメだったのにこれで回復した。 自分は一般的にはPCの知識がある方だと思っていたけど、これは知らなかった。 本当に感謝。

  4. openblobbdgeo.z1.web.core.windows.net openblobbdgeo.z1.web.core.windows.net
    1. Table of contents

      vi behöver även lägga till (som attachment eller som en del av resultatet) available discharged power" for all cases, at the end of the discharge season

    2. 2.3.6 Recommendations It is hard to achieve this design condition, either one must have 10 m of insulation downwards or 11.5 m sideways (which is apparently impossible), either the heat leaks in the center and melts the snow, or it leaks in the surroundings (since we can’t insulate them).

      se tidigare kommentar. Om vi vill lämna kvar denna text så behöver vi en gång till påminna läsare om källan för detta påstående

    3. The efficiency varies considerably between zones, ranging from 38% for the peripheral zone and 83% for the innermost zone.

      se tidigare kommentar. Är det samma resultat än Ruden? Kommentera gärna.

    4. To achieve optimal storage, wells should be parallel. Geological structures will influence well paths, and different drilling methods need testing to control deviations effectively.

      se tidigare resultat. If we want to keep the text, make sure to reformulate it so that it is clear that this is Rudens recommendation, but even better, just remain the reader about the source of this text.

    5. Ruden worked during 2022 on several aspects related to the Furuset HT-BTES Ven (2022a) Ruden, Zengaffinen-Morris, and Ven (2022), Bastesen and Ven (2022), Zengaffinen-Morris and Ven (2022), Ven (2022b), Ven (2022c) and Ven (2022d), culminating in a global report Zengaffinen-Morris, Meakins, and Ruden (2022).

      kan vi förtydliga sekvensen mellan dessa? det är lite otydligt.

      Vi kan annars upprepa här att sekvensen mellan dessa är oklar.

    6. but it’s unclear if 91 or 392 BHEs are used.

      stor skillnad mellan dessa två case. Är vi säkra på att det inte anges? Vänligen dubbelkolla. Below we write 392. I guess thats the one?

      Om man tittar på siffrorna i output, efterliknader siffrorna ett stort lager med 392? eller med 91? vi kan isf lägga till en kommentar där vi skriver vad vi tror de har simulerat.

    7. The Furuset LT-DHN will feature an annual heat exchange of 20 GWh, a pipe length between 5-10 km, and will serve over 2,000 flats and commercial buildings with temperatures ranging from 70 ∘C (supply) to 50 ∘C (return). This network will connect to Oslo’s High Temperature District Heating Network (HT-DHN) via a 3 km transmission pipe, leveraging excess heat from the main DHN. The HT-DHN spans 400 km and delivers approximately 200 TWh per year at temperatures between 85 ∘C and 120 ∘C

      lägg till en referens till denna text.

    8. Reviewer

      det saknas mellanrubrik för UA (project responsible) enligt tidigare instruktion. Ni behöver inte fixa det denna gång men mallen måste innehålla det framöver.

    1. For the first time, there was someone else with him: A squat white woman with a plastic name-tag and the kind of squarish perm you can only get in Southern beauty salons with faded glamor-shots in the windows. The boy trailed behind her looking thin and pressed, like a flower crushed between dictionary pages. I wondered how badly you had to fuck up to get assigned a school counselor after hours, until I read her name-tag: Department of Community-Based Services, Division of Protection and Permanency, Child Caseworker (II).Oh. A foster kid.

      I like this part a lot because it gives the reader a new perspective of this character that is in the narrator's life. We also get to see how she comes to realize the deeper part of his life.

    1. Welcome back. In this lesson, I want to talk about CloudWatch, a core product inside AWS used for operational management and monitoring. CloudWatch performs three main jobs: it collects and manages operational data, monitors metrics, and performs actions based on these metrics.

      CloudWatch collects and manages operational data generated by an environment, including performance details, nominal operations, and logging data. It can be considered three products in one: CloudWatch, CloudWatch Logs, and CloudWatch Events.

      Firstly, CloudWatch allows the collection, monitoring, and actions based on metrics related to AWS products, applications, or on-premises systems. Metrics include data such as CPU utilization, disk space usage, or website traffic. CloudWatch can gather metrics from AWS, on-premises environments, or other cloud platforms using a public internet connection. Some metrics are gathered natively by CloudWatch, while others require the CloudWatch Agent, especially for monitoring non-AWS environments or specific processes on AWS instances.

      CloudWatch provides a user interface, command line interface, or API to access and manage this data. The second part of CloudWatch, CloudWatch Logs, handles the collection, monitoring, and actions based on logging data from various sources like Windows event logs, web server logs, and more. For custom logs or non-AWS systems, the CloudWatch Agent is also needed.

      The third part is CloudWatch Events, which functions as an event hub. It generates events based on AWS service actions (e.g., starting or stopping an EC2 instance) and can also create scheduled events for specific times or days.

      The core concepts of CloudWatch include namespaces, metrics, datapoints, and dimensions. A namespace is a container for monitoring data, helping to organize and separate different areas of data. AWS uses a reserved namespace format (e.g., AWS/EC2 for EC2 metrics), while you can create custom namespaces for your data. Metrics are collections of related data points in a time-ordered structure, such as CPU utilization. Each datapoint includes a timestamp and value. Dimensions, which are name-value pairs, help separate and identify data within a metric, like distinguishing datapoints from different EC2 instances.

      CloudWatch also uses alarms to take actions based on metrics. Alarms can be in an OK state (indicating no issues), an ALARM state (indicating a problem), or an INSUFFICIENT_DATA state (indicating not enough data to assess). Actions could include notifications or more complex responses. You’ve already seen an example of this with the billing alarm created at the start of the course.

      In the next demo lesson, we’ll provision an EC2 instance, let it run, and then create an alarm to monitor CPU usage, providing practical exposure to how CloudWatch works.

      Thanks for watching. Complete this video and join me in the demo when you’re ready.

    1. Reviewer #1 (Public Review):

      Summary:

      This study addresses the question of how task-relevant sensory information affects activity in the motor cortex. The authors use various approaches to address this question, looking at single units and population activity. They find that there are three subtypes of modulation by sensory information at the single unit level. Population analyses reveal that sensory information affects the neural activity orthogonally to motor output. The authors then compare both single unit and population activity to computational models to investigate how encoding of sensory information at the single unit level is coordinated in a network. They find that an RNN that displays similar orbital dynamics and sensory modulation to the motor cortex also contains nodes that are modulated similarly to the three subtypes identified by the single unit analysis.

      Strengths:

      The strengths of this study lie in the population analyses and the approach of comparing single-unit encoding to population dynamics. In particular, the analysis in Figure 3 is very elegant and informative about the effect of sensory information on motor cortical activity. The task is also well designed to suit the questions being asked and well controlled.

      It is commendable that the authors compare single units to population modulation. The addition of the RNN model and perturbations strengthen the conclusion that the subtypes of individual units all contribute to the population dynamics. However, the subtypes (PD shift, gain, and addition) are not sufficiently justified. The authors also do not address that single units exhibit mixed modulation, but RNN units are not treated as such.

      Weaknesses:

      The main weaknesses of the study lie in the categorization of the single units into PD shift, gain, and addition types. The single units exhibit clear mixed selectivity, as the authors highlight. Therefore, the subsequent analyses looking only at the individual classes in the RNN are a little limited. Another weakness of the paper is that the choice of windows for analyses is not properly justified and the dependence of the results on the time windows chosen for single-unit analyses is not assessed. This is particularly pertinent because tuning curves are known to rotate during movements (Sergio et al. 2005 Journal of Neurophysiology).

      This paper shows sensory information can affect motor cortical activity whilst not affecting motor output. However, it is not the first to do so and fails to cite other papers that have investigated sensory modulation of the motor cortex (Stavinksy et al. 2017 Neuron, Pruszynski et al. 2011 Nature, Omrani et al. 2016 eLife). These studies should be mentioned in the Introduction to capture better the context around the present study. It would also be beneficial to add a discussion of how the results compare to the findings from these other works.

      This study also uses insights from single-unit analysis to inform mechanistic models of these population dynamics, which is a powerful approach, but is dependent on the validity of the single-cell analysis, which I have expanded on below.

      I have clarified some of the areas that would benefit from further analysis below:

      (1) Task:<br /> The task is well designed, although it would have benefited from perhaps one more target speed (for each direction). One monkey appears to have experienced one more target speed than the others (seen in Figure 3C). It would have been nice to have this data for all monkeys.

      (2) Single unit analyses:<br /> In some analyses, the effects of target speed look more driven by target movement direction (e.g. Figures 1D and E). To confirm target speed is the main modulator, it would be good to compare how much more variance is explained by models including speed rather than just direction. More target speeds may have been helpful here too.

      The choice of the three categories (PD shift, gain addition) is not completely justified in a satisfactory way. It would be nice to see whether these three main categories are confirmed by unsupervised methods.

      The decoder analyses in Figure 2 provide evidence that target speed modulation may change over the trial. Therefore, it is important to see how the window considered for the firing rate in Figure 1 (currently 100ms pre - 100ms post movement onset) affects the results.

      (3) Decoder:<br /> One feature of the task is that the reach endpoints tile the entire perimeter of the target circle (Figure 1B). However, this feature is not exploited for much of the single-unit analyses. This is most notable in Figure 2, where the use of a SVM limits the decoding to discrete values (the endpoints are divided into 8 categories). Using continuous decoding of hand kinematics would be more appropriate for this task.

      (4) RNN:<br /> Mixed selectivity is not analysed in the RNN, which would help to compare the model to the real data where mixed selectivity is common. Furthermore, it would be informative to compare the neural data to the RNN activity using canonical correlation or Procrustes analyses. These would help validate the claim of similarity between RNN and neural dynamics, rather than allowing comparisons to be dominated by geometric similarities that may be features of the task. There is also an absence of alternate models to compare the perturbation model results to.

    2. Author response:

      eLife assessment

      This useful study examines the neural activity in the motor cortex as a monkey reaches to intercept moving targets, focusing on how tuned single neurons contribute to an interesting overall population geometry. The presented results and analyses are solid, though the investigation of this novel task could be strengthened by clarifying the assumptions behind the single neuron analyses, and further analyses of the neural population activity and its relation to different features of behaviour.

      Thanks for recognizing the content of our research, and please stay tuned for our follow-up studies on neural dynamics during interception.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      This study addresses the question of how task-relevant sensory information affects activity in the motor cortex. The authors use various approaches to address this question, looking at single units and population activity. They find that there are three subtypes of modulation by sensory information at the single unit level. Population analyses reveal that sensory information affects the neural activity orthogonally to motor output. The authors then compare both single unit and population activity to computational models to investigate how encoding of sensory information at the single unit level is coordinated in a network. They find that an RNN that displays similar orbital dynamics and sensory modulation to the motor cortex also contains nodes that are modulated similarly to the three subtypes identified by the single unit analysis.

      Strengths:

      The strengths of this study lie in the population analyses and the approach of comparing single-unit encoding to population dynamics. In particular, the analysis in Figure 3 is very elegant and informative about the effect of sensory information on motor cortical activity. The task is also well designed to suit the questions being asked and well controlled.

      We appreciate these kind comments.

      It is commendable that the authors compare single units to population modulation. The addition of the RNN model and perturbations strengthen the conclusion that the subtypes of individual units all contribute to the population dynamics. However, the subtypes (PD shift, gain, and addition) are not sufficiently justified. The authors also do not address that single units exhibit mixed modulation, but RNN units are not treated as such.

      We’re sorry for not providing sufficient grounds to introduce the subtypes. We determined the PD shift, gain, and addition as pertinent subtypes based on classical cosine tuning model (Georgopoulos et al., 1982) and referred to some gain modulation studies (e.g. Pesaran et al. 2010, Bremner and Andersen, 2012). Here, we applied this subtype analysis as a criteria to identify the modulation in neuronal population rather than to sort neuron into distinct cell types. We will update Methods in the revised version of manuscript.

      Weaknesses:

      The main weaknesses of the study lie in the categorization of the single units into PD shift, gain, and addition types. The single units exhibit clear mixed selectivity, as the authors highlight. Therefore, the subsequent analyses looking only at the individual classes in the RNN are a little limited. Another weakness of the paper is that the choice of windows for analyses is not properly justified and the dependence of the results on the time windows chosen for single-unit analyses is not assessed. This is particularly pertinent because tuning curves are known to rotate during movements (Sergio et al. 2005 Journal of Neurophysiology).

      The mixed selectivity or precisely the mixed modulation is indeed a significant feature of neuronal population in the present study. The purpose of the subtype analysis was to serve as a criterion for the potential modulation mechanisms. However, the results appear to be a spectrum than clusters. It still through some insights to understand the modulation distribution and we will refine the description in the next version. In the current version, we observed single-unit tuning and population neural state with sliding windows, focusing on the period around movement onset (MO) due to the emergence of a ring-like structure. We will clarify the choice of windows and the dependence assessment in the next version. It’s a great suggestion to consider the role of rotating tuning curves in neural dynamics during interception.

      This paper shows sensory information can affect motor cortical activity whilst not affecting motor output. However, it is not the first to do so and fails to cite other papers that have investigated sensory modulation of the motor cortex (Stavinksy et al. 2017 Neuron, Pruszynski et al. 2011 Nature, Omrani et al. 2016 eLife). These studies should be mentioned in the Introduction to capture better the context around the present study. It would also be beneficial to add a discussion of how the results compare to the findings from these other works.

      Thanks for the reminder. We will introduce the relevant research in the next version of manuscript.

      This study also uses insights from single-unit analysis to inform mechanistic models of these population dynamics, which is a powerful approach, but is dependent on the validity of the single-cell analysis, which I have expanded on below.

      I have clarified some of the areas that would benefit from further analysis below:

      (1) Task:

      The task is well designed, although it would have benefited from perhaps one more target speed (for each direction). One monkey appears to have experienced one more target speed than the others (seen in Figure 3C). It would have been nice to have this data for all monkeys.

      Great suggestion! However, it’s hard to implement as the implanted arrays have been removed.

      (2) Single unit analyses:

      In some analyses, the effects of target speed look more driven by target movement direction (e.g. Figures 1D and E). To confirm target speed is the main modulator, it would be good to compare how much more variance is explained by models including speed rather than just direction. More target speeds may have been helpful here too.

      Nice suggestion! The fitting goodness of the simple model (just motor direction) is much less than the complex model (including target speed). We will update the results in the next version.

      The choice of the three categories (PD shift, gain addition) is not completely justified in a satisfactory way. It would be nice to see whether these three main categories are confirmed by unsupervised methods.

      A good point. We will have a try with unsupervised methods. 

      The decoder analyses in Figure 2 provide evidence that target speed modulation may change over the trial. Therefore, it is important to see how the window considered for the firing rate in Figure 1 (currently 100ms pre - 100ms post movement onset) affects the results.

      Thanks for the suggestion and close reading. We will test the decoder in other epochs.

      (3) Decoder:

      One feature of the task is that the reach endpoints tile the entire perimeter of the target circle (Figure 1B). However, this feature is not exploited for much of the single-unit analyses. This is most notable in Figure 2, where the use of a SVM limits the decoding to discrete values (the endpoints are divided into 8 categories). Using continuous decoding of hand kinematics would be more appropriate for this task.

      This is a very reasonable suggestion. In this study, we discrete the reach-direction as the previous studies (Li et al., 2018&2022) and thought that the discrete decoding was already enough to show the interaction of sensory and motor variables. In future studies, we will try continuous decoding of hand kinematics.

      (4) RNN:

      Mixed selectivity is not analysed in the RNN, which would help to compare the model to the real data where mixed selectivity is common. Furthermore, it would be informative to compare the neural data to the RNN activity using canonical correlation or Procrustes analyses. These would help validate the claim of similarity between RNN and neural dynamics, rather than allowing comparisons to be dominated by geometric similarities that may be features of the task. There is also an absence of alternate models to compare the perturbation model results to.

      Thank you for these helpful suggestions. We will perform decoding analysis on RNN units to verify if there is interaction of sensory and motor variables as in real data, as well as the canonical correlation or Procrustes analysis.

      Reviewer #2 (Public Review):

      Summary:

      In this manuscript, Zhang et al. examine neural activity in the motor cortex as monkeys make reaches in a novel target interception task. Zhang et al. begin by examining the single neuron tuning properties across different moving target conditions, finding several classes of neurons: those that shift their preferred direction, those that change their modulation gain, and those that shift their baseline firing rates. The authors go on to find an interesting, tilted ring structure of the neural population activity, depending on the target speed, and find that (1) the reach direction has consistent positioning around the ring, and (2) the tilt of the ring is highly predictive of the target movement speed. The authors then model the neural activity with a single neuron representational model and a recurrent neural network model, concluding that this population structure requires a mixture of the three types of single neurons described at the beginning of the manuscript.

      Strengths:

      I find the task the authors present here to be novel and exciting. It slots nicely into an overall trend to break away from a simple reach-to-static-target task to better characterize the breadth of how the motor cortex generates movements. I also appreciate the movement from single neuron characterization to population activity exploration, which generally serves to anchor the results and make them concrete. Further, the orbital ring structure of population activity is fascinating, and the modeling work at the end serves as a useful baseline control to see how it might arise.

      Thank you for recognizing our work.

      Weaknesses:

      While I find the behavioral task presented here to be excitingly novel, I find the presented analyses and results to be far less interesting than they could be. Key to this, I think, is that the authors are examining this task and related neural activity primarily with a single-neuron representational lens. This would be fine as an initial analysis since the population activity is of course composed of individual neurons, but the field seems to have largely moved towards a more abstract "computation through dynamics" framework that has, in the last several years, provided much more understanding of motor control than the representational framework has. As the manuscript stands now, I'm not entirely sure what interpretation to take away from the representational conclusions the authors made (i.e. the fact that the orbital population geometry arises from a mixture of different tuning types). As such, by the end of the manuscript, I'm not sure I understand any better how the motor cortex or its neural geometry might be contributing to the execution of this novel task.

      The present study shows the sensory modulation on motor tuning in single units and neural state during motor execution period. It’s a pity that the findings were constrained in certain time windows. We are still working this topic, and hopefully will address related questions in our follow-up studies.

      Main Comments:

      My main suggestions to the authors revolve around bringing in the computation through a dynamics framework to strengthen their population results. The authors cite the Vyas et al. review paper on the subject, so I believe they are aware of this framework. I have three suggestions for improving or adding to the population results:

      (1) Examination of delay period activity: one of the most interesting aspects of the task was the fact that the monkey had a random-length delay period before he could move to intercept the target. Presumably, the monkey had to prepare to intercept at any time between 400 and 800 ms, which means that there may be some interesting preparatory activity dynamics during this period. For example, after 400ms, does the preparatory activity rotate with the target such that once the go cue happens, the correct interception can be executed? There is some analysis of the delay period population activity in the supplement, but it doesn't quite get at the question of how the interception movement is prepared. This is perhaps the most interesting question that can be asked with this experiment, and it's one that I think may be quite novel for the field--it is a shame that it isn't discussed.

      Great idea! We are on the way, and close to complete the puzzle.

      (2) Supervised examination of population structure via potent and null spaces: simply examining the first three principal components revealed an orbital structure, with a seemingly conserved motor output space and a dimension orthogonal to it that relates to the visual input. However, the authors don't push this insight any further. One way to do that would be to find the "potent space" of motor cortical activity by regression to the arm movement and examine how the tilted rings look in that space (this is actually fairly easy to see in the reach direction components of the dPCA plot in the supplement--the rings will be highly aligned in this space). Presumably, then, the null space should contain information about the target movement. dPCA shows that there's not a single dimension that clearly delineates target speed, but the ring tilt is likely evident if the authors look at the highest variance neural dimension orthogonal to the potent space (the "null space")--this is akin to PC3 in the current figures, but it would be nice to see what comes out when you look in the data for it.

      Nice suggestion. Target-speed modulation mainly influences PC3, which is consistent with ‘null space’ hypothesis. We will try other methods of dimensionality reduction (e.g. dPCA, Manopt) to determine the potent and null space.

      (3) RNN perturbations: as it's currently written, the RNN modeling has promise, but the perturbations performed don't provide me with much insight. I think this is because the authors are trying to use the RNN to interpret the single neuron tuning, but it's unclear to me what was learned from perturbing the connectivity between what seems to me almost arbitrary groups of neurons (especially considering that 43% of nodes were unclassifiable). It seems to me that a better perturbation might be to move the neural state before the movement onset to see how it changes the output. For example, the authors could move the neural state from one tilted ring to another to see if the virtual hand then reaches a completely different (yet predictable) target. Moreover, if the authors can more clearly characterize the preparatory movement, perhaps perturbations in the delay period would provide even more insight into how the interception might be prepared.

      We are sorry that we didn’t clarify the definition of “none” type, which can be misleading. The 43% unclassified nodes include those inactive ones, when only activate (task-related) nodes included, the ratio of unclassified nodes would be much lower. By perturbing the connectivity, we intended to explore the interaction between different modulations.

      Thank you for the great advice. We tried moving neural states from one ring to another without changing the directional cluster, but this perturbation didn’t have a significant influence on network performance as expected. We will check this result again and try perturbations in the delay period.

      Reviewer #3 (Public Review):

      Summary:

      This experimental study investigates the influence of sensory information on neural population activity in M1 during a delayed reaching task. In the experiment, monkeys are trained to perform a delayed interception reach task, in which the goal is to intercept a potentially moving target.

      This paradigm allows the authors to investigate how, given a fixed reach endpoint (which is assumed to correspond to a fixed motor output), the sensory information regarding the target motion is encoded in neural activity.

      At the level of single neurons, the authors found that target motion modulates the activity in three main ways: gain modulation (scaling of the neural activity depending on the target direction), shift (shift of the preferred direction of neurons tuned to reach direction), or addition (offset to the neural activity).

      At the level of the neural population, target motion information was largely encoded along the 3rd PC of the neural activity, leading to a tilt of the manifold along which reach direction was encoded that was proportional to the target speed. The tilt of the neural manifold was found to be largely driven by the variation of activity of the population of gain-modulated neurons.

      Finally, the authors studied the behaviour of an RNN trained to generate the correct hand velocity given the sensory input and reach direction. The RNN units were found to similarly exhibit mixed selectivity to the sensory information, and the geometry of the « neural population » resembled that observed in the monkeys.

      Strengths:

      - The experiment is well set up to address the question of how sensory information that is directly relevant to the behaviour but does not lead to a direct change in behavioural output modulates motor cortical activity.

      - The finding that sensory information modulates the neural activity in M1 during motor preparation and execution is non trivial, given that this modulation of the activity must occur in the nullspace of the movement.

      - The paper gives a complete picture of the effect of the target motion on neural activity, by including analyses at the single neuron level as well as at the population level. Additionally, the authors link those two levels of representation by highlighting how gain modulation contributes to shaping the population representation.

      Thanks for your recognition.

      Weaknesses:

      - One of the main premises of the paper is the fact that the motor output for a given reach point is preserved across different target motions. However, as the authors briefly mention in the conclusion, they did not record muscle activity during the task, but only hand velocity, making it impossible to directly verify how preserved muscle patterns were across movements. While the authors highlight that they did not see any difference in their results when resampling the data to control for similar hand velocities across conditions, this seems like an important potential caveat of the paper whose implications should be discussed further or highlighted earlier in the paper.

      Thanks for the suggestion. We will highlight the resampling results as important control in the next version of manuscript.

      - The main takeaway of the RNN analysis is not fully clear. The authors find that an RNN trained given a sensory input representing a moving target displays modulation to target motion that resembles what is seen in real data. This is interesting, but the authors do not dissect why this representation arises, and how robust it is to various task design choices. For instance, it appears that the network should be able to solve the task using only the motion intention input, which contains the reach endpoint information. If the target motion input is not used for the task, it is not obvious why the RNN units would be modulated by this input (especially as this modulation must lie in the nullspace of the movement hand velocity if the velocity depends only on the reach endpoint). It would thus be important to see alternative models compared to true neural activity, in addition to the model currently included in the paper. Besides, for the model in the paper, it would therefore be interesting to study further how the details of the network setup (eg initial spectral radius of the connectivity, weight regularization, or using only the target position input) affect the modulation by the motion input, as well as the trained population geometry and the relative ratios of modulated cells after training.

      Great suggestions. It’s a considerable pity that we didn’t dissect the formation reason and influence factor of the representation in the current version. We’ve tried several combinations of inputs before: in the network which received only motor intention and GO inputs, there were rings but not tilting related to target-speed; in the network which received only target location and GO inputs, there were ring-like structures but not clear directional clusters. We will check these results and try alternative models in the next version. In future studies, we will examine the influence of network setup details.

      - Additionally, it is unclear what insights are gained from the perturbations to the network connectivity the authors perform, as it is generally expected that modulating the connectivity will degrade task performance and the geometry of the responses. If the authors wish the make claims about the role of the subpopulations, it could be interesting to test whether similar connectivity patterns develop in networks that are not initialized with an all-to-all random connectivity or to use ablation experiments to investigate whether the presence of multiple types of modulations confers any sort of robustness to the network.

      Thank you for the great suggestions. By perturbations, we intended to explore the contribution of interaction between certain subpopulations. We tried ablation experiments, but the result was not significant. Probably because the most units were of mixed selectivity, the units of only modulations were not enough for bootstrapping, or the random sampling from single subpopulation (bearing mixed selectivity) could be repeated. We will consider these suggestions carefully in the revised version.

      - The results suggest that the observed changes in motor cortical activity with target velocity result from M1 activity receiving an input that encodes the velocity information. This also appears to be the assumption in the RNN model. However, even though the input shown to the animal during preparation is indeed a continuously moving target, it appears that the only relevant quantity to the actual movement is the final endpoint of the reach. While this would have to be a function of the target velocity, one could imagine that the computation of where the monkeys should reach might be performed upstream of the motor cortex, in which case the actual target velocity would become irrelevant to the final motor output. This makes the results of the paper very interesting, but it would be nice if the authors could discuss further when one might expect to see modulation by sensory information that does not directly affect motor output in M1, and where those inputs may come from. It may also be interesting to discuss how the findings relate to previous work that has found behaviourally irrelevant information is being filtered out from M1 (for instance, Russo et al, Neuron 2020 found that in monkeys performing a cycling task, context can be decoded from SMA but not from M1, and Wang et al, Nature Communications 2019 found that perceptual information could not be decoded from PMd)?

      How and where sensory information modulates M1 are very interesting and open questions. We will discuss further about this topic in the next version.

    1. Reviewer #1 (Public Review):

      Summary:

      Previous research from the Margoliash laboratory has demonstrated that the intrinsic electrophysiological properties of one class of projection neurons in the song nucleus HVC, HVCX neurons, are similar within birds and differ between birds in a manner that relates to the bird's song. The current study builds on this research by addressing how intrinsic properties may relate to the temporal structure of the bird's song and by developing a computational model for how this can influence sequence propagation of activity within HVC during singing.

      First, the authors identify that the duration of the song motif is correlated with the duration of song syllables and particularly the length of harmonic stacks within the song. They next found positive correlations between some of the intrinsic properties, including firing frequency, sag ratio, and rebound excitation area with the duration of the birds' longest harmonic syllable and some other measure of motif duration. These results were extended by examining measures of firing frequency and sag ratio between two groups of birds that were experimentally raised to learn songs that only differed by the addition of a long terminal harmonic stack in one of the groups. Lastly, the authors present an HH-based model elucidating how the timing and magnitude of rebound excitation of HVCX neurons can function to support previously reported physiological network properties of these neurons during singing.

      Strengths:

      By trying to describe how intrinsic properties (IPs) may relate to the structure of learned behavior and providing a potentially plausible model (see below for more on this) for how differences in IPs can relate to sequence propagation in this neural network, this research is addressing an important and challenging issue. An understanding of how cell types develop IPs and how those IPs relate to the function and output of a network is a fundamental issue. Tackling this in the zebra finch HVC is an elegant approach because it provides a quantifiable and reliable behavior that is explicitly tied to the neurons that the authors are studying. Nonetheless, this is a difficult problem, and kudos to the authors for trying to unravel this.

      Correlations between harmonic stack durations and song durations are well-supported and interesting. This provides a new insight that can and will likely be used by other research groups in correlating neuronal activity patterns to song behavior and motif duration. Additionally, correlations between IPs associated with rebound excitation are also well supported in this study.

      The HH-model presented is important because it meaningfully relates how high or low rebound excitation can set the integration time window for HVCX neurons. Further, the synaptic connectivity of this model provides at least one plausible way in how this functions to permit the bursting activity of HVCX neurons during singing (and potentially during song playback experiments in sleeping birds). Thus, this model will be useful to the field for understanding how this network activity intersects with 'learned' IPs in an important class of neurons in this circuit.

      Weaknesses:

      The main weakness of the study is that there is somewhat of a disconnect between the physiological measurements described and the key components of the circuit model presented at the end of the paper. Thus, better support could be provided to link the magnitude of rebound excitation with song temporal structure. The rebound excitation area is shown to be positively correlated with the longest harmonic stack. Does this correlation hold when the four birds with unusually long stacks (>150ms) are excluded? Is rebound excitation area positively correlated with motif duration? Additionally, rebound excitation was not considered when examining experimentally tutored birds. Further analysis of these correlations can better link this research to the model presented.

      The HH model is of general interest, but I am concerned about the plausibility of some of this circuitry, particularly because synaptic connectivity underlying information flow is a central component of the model. At several steps in the model, excitatory drive onto HVCX neurons is coming from another HVCX neuron. Although disynaptic inhibition between HVCX neurons and between HVCRA and HVCX neurons is well established, I am not aware of any data indicating direct synaptic connections between HVCX neurons.

      Thus, how does the model change if all excitatory drive onto HVCX neurons are coming from HVCRA neurons? Currently, the model is realized through neurons being active at syllable or gesture transitions. What does the model predict about the distribution of HVCRA neurons activity across songs if they are the exclusive excitatory input to HVCX neurons? A better consideration of these issues can improve the suitability of the model in the context of known connectivity.

      If I understand the model and ideas correctly, birds with longer motifs should exhibit longer pauses in the activity of tonically active HVC interneurons during singing and they should exhibit longer post-rebound integration windows. Experimental evidence supporting either of these ideas is not provided and would strengthen this research.

    2. Reviewer #2 (Public Review):

      Intrinsic properties of a neuron refer to the ion channels that a neuron expresses. These ion channels determine how a neuron responds to its inputs. How intrinsic properties link to behavior remains poorly understood. Medina and Margoliash address this question using the zebra finch, a well-studied songbird. Previous studies from their lab and other labs have shown that the intrinsic properties of adult songbird basal-ganglia projecting premotor neurons, are more similar within a bird than across birds. Across birds, this similarity is related to the extent of similarity in the songs; the more similar the song between two birds, the more similar the intrinsic properties between the neurons of these two birds. Finally, the intrinsic properties of these neurons change over the course of development and are sensitive to intact auditory feedback. However, the song features that relate to these intrinsic properties and the function of the within-bird homogeneity of intrinsic properties are unclear.

      In this manuscript, the authors address these two questions by examining the intrinsic properties of basal-ganglia projecting premotor neurons in zebra finch brain slices. Specifically, they focus on the Ih current (as this is related to rhythmic activity in many pattern-generating circuits) and correlate the properties of the Ih current with song features. They find that the sag ratio (a measure of the driving force of the Ih current) and the rebound area (a measure of the post-inhibitory depolarisation) are both correlated with the temporal features of the song. First, they show the presence of correlations between the length of the song motif and the length of the longest syllable (most often a harmonic stack syllable). Based on this, they conclude that longer song motifs are composed of longer syllables. Second, they show that HVCX neurons within a bird have more similar sag ratios and rebound areas than across birds. Third, the mean sag ratio and mean rebound areas across birds were correlated with the duration of the longest harmonic stack within the song. These two results suggest that IPs are correlated with the temporal structure of the song. To further test this, the authors used natural and experimental tutoring procedures to have birds that learned two different types of songs that only differed in length; the longer song had an extra harmonic stack at the end. Using these two sets of birds, the authors find larger sag ratios and higher firing frequencies in birds with longer songs. Fifth, they show that the post-inhibitory rebound area allows neurons to respond to excitatory inputs and produce spikes. Neurons with a larger rebound area have a larger time window for responding to excitatory inputs. Based on this, they speculate that HVCX neurons with larger rebound areas integrate over larger time windows. Finally, they make a network model of HVC and show that one specific model could explain sequence-specific bursting of HVCX neurons.

      Strengths

      The question being addressed is an interesting question and the authors use appropriate techniques. The authors find a new temporal structure within the song, specifically, they find that longer songs typically have more syllables and longer syllables. As far as I know, this has not been shown earlier. The authors build on existing literature to suggest that IPs of HVCX neurons are correlated with the temporal structure of songs.

      Weaknesses

      I have a number of concerns with the statistics and interpretation of the results, insufficient controls for one experiment, and the specifics of the model that affect the implications of these results. These concerns are listed in the recommendations for the authors.

    1. Welcome back. In this lesson, I want to introduce another core AWS service, the simple storage service known as S3. If you use AWS in production, you need to understand S3. This lesson will give you the very basics because I'll be deep diving into a specific S3 section later in the course, and the product will feature constantly as we go. Pretty much every other AWS service has some kind of interaction with S3. So let's jump in and get started.

      S3 is a global storage platform. It's global because it runs from all of the AWS regions and can be accessed from anywhere with an internet connection. It's a public service. It's regional based because your data is stored in a specific AWS region at rest. So when it's not being used, it's stored in a specific region. And it never leaves that region unless you explicitly configure it to. S3 is regionally resilient, meaning the data is replicated across availability zones in that region. S3 can tolerate the failure of an AZ, and it also has some ability to replicate data between regions, but more on that in the S3 section of the course.

      Now S3 might initially appear confusing. If you utilize it from the UI, you appear not to have to select a region. Instead, you select the region when you create things inside S3, which I'll talk about soon. S3 is a public service, so it can be accessed from anywhere as long as you have an internet connection. The service itself runs from the AWS public zone. It can cope with unlimited data amounts and it's designed for multi-user usage of that data. So millions of users could be accessing cute cat pictures added by the Animals for Life Rescue Officers. S3 is perfect for hosting large amounts of data. So think movies or audio distribution, large scale photo storage like stock images, large textual data or big data sets. It could be just as easily used for millions or billions of IOT devices or to store images for a blog. It scales from nothing to near unlimited levels.

      Now S3 is economical, it's a great value service for storing and allowing access to data. And it can be accessed using a variety of methods. There's the GUI, you can use the command line, the AWS APIs or even standard methods such as HTTP. I want you to think of S3 as the default storage service in AWS. It should be your default starting point unless your requirement isn't delivered by S3. And I'll talk more about the limitations and use cases later in this lesson.

      S3 has two main things that it delivers: Objects and Buckets. Objects are the data the S3 stores, your cat picture, the latest episode of Game of Thrones, which you have stored legally, of course, or it could be large scale datasets showing the migration of the koala population in Australia after a major bushfire. Buckets are containers for objects. It's buckets and objects that I want to cover in this lesson as an introduction to the service.

      So first, let's talk about objects. An object in S3 is made up of two main components and some associated metadata. First, there is the object key. And for now you can think of the object key, similar to a file name. The key identifies the object in a bucket. So if you know the object key and the bucket, then you can uniquely access the object, assuming that you have permissions. Remember by default, even for public services, there is no access in AWS initially, except for the account root user.

      Now, the other main component of an object is its value. And the value is the data or the contents of the object. In this case, a sequence of binary data, which represents a koala in his house. In this course, I want to avoid suggesting that you remember pointless values. Sometimes though, there are things that you do need to commit to memory. And this is one of those times. The value of an object, in essence, how large the object is, can range from zero bytes up to five terabytes in size. So you can have an empty object or you can have one that is a huge five TB. This is one of the reasons why S3 is so scalable and so useful in a wide range of situations because of this range of sizes for objects.

      Now, the other components of an object, aren't that important to know at this stage, but just so you hear the terms that I'll use later, objects also have a version ID, metadata, some access control, as well as sub resources. Now don't worry about what they do for now, I'll be covering them all later. For this lesson, just try to commit to memory what an object is, what components it has and the size range for an object.

      So now that we've talked about objects, let's move on and look at buckets. Buckets are created in a specific AWS region. And let's use Sydney or ap-southeast-2 as an example. This has two main impacts. Firstly, your data that's inside a bucket has a primary home region. And it never leaves that region, unless you as an architect or one of your system admins configures that data to leave that region. That means that S3 has stable and controlled data sovereignty. By creating a bucket in a region, you can control what laws and rules apply to that data. What it also means is that the blast radius of a failure is that region.

      Now this might be a new term. What I mean by blast radius is that if a major failure occurs, say a natural disaster or a large scale data corruption, the effect of that will be contained within the region. Now a bucket is identified by its name, the bucket name in this case, koala data. A bucket name needs to be globally unique. So that's across all regions and all accounts of AWS. If I pick a bucket name, in this case, koala data, nobody else can use it in any AWS account. Now making a point of stressing this as it often comes up in the exam. Most AWS things are often unique in a region or unique in your account. For example, I can have an IAM user called Fred and you can also have an IAM user called Fred. Buckets though are different, with buckets, the name has to be totally unique, and that's across all regions and all AWS accounts. I've seen it come up in the exam a few times. So this is definitely a point to remember.

      Now buckets can hold an unlimited number of objects. And because objects can range from zero to five TB in size, that essentially means that a bucket can hold from zero to unlimited bytes of data. It's an infinitely scalable storage system. Now one of the most important things that I want to say in this lesson is that as an object storage system, an S3 bucket has no complex structure. It's flat, it has flat structure. All objects stored within the bucket at the same level. So this isn't like a file system where you can truly have files within folders, within folders. Everything is stored in the bucket at the root level.

      But, if you do a listing on an S3 bucket, you will see what you think are folders. Even the UI presents this as folders. But it is important for you to know at this stage that that's not how it actually is. Imagine a bucket where you see three image files, koala one, two and three dot JPEG. The first thing is that inside S3, there's no concept of file type based on the name. These are just three objects where the object key is koala1.JPEG, koala2.JPEG and koala3.JPEG. Now folders in S3 are represented when we have object names that are structured like these. So the objects have a key, a forward slash old forward slash koala one, two and three dot JPEG. When we create object names like this, then S3 presents them in the UI as a folder called old. So because we've got object names that begin with slash old, then S3 presents this as a folder called old. And then inside that folder, we've got koala one, two, and three dot JPEG.

      Now nine out of 10 times, this detail doesn't matter, but I want to make sure that you understand how it actually works. Folders are often referred to as prefixes in S3 because they're part of the object names. They prefix the object names. As you move through the various stages of your AWS learnings, this is gonna make more and more sense. And I'm gonna demonstrate this in the next lesson, which is a demo lesson.

      Now to summarize buckets are just containers, they're stored in a region, and for S3, they're generally where a lot of permissions and options are set. So remember that buckets are generally the default place where you should go to, to configure the way the S3 works.

      Now, I want to cover a few summary items and then step through some patterns and anti-patterns for S3, before we move to the demo. But first an exam powerup. These are things that you should try to remember and they will really help in the exam. First bucket names are globally unique. Remember that one because it will really help in the exam. I've seen a lot of times where AWS have included trick questions, which test your knowledge of this one. If you get any errors, you can't create a bucket a lot of the time it's because somebody else already has that bucket name.

      Now bucket names do have some restrictions. They need to be between 3 and 63 characters, all lower case and no underscores. They have to start with a lowercase letter or a number, and they can't be formatted like IP addresses. So you can't have 1.1.1.1 as your bucket name. Now there are some limits in terms of buckets. Now limits are often things that you don't need to remember for the exam, but this is one of the things that you do. There is a limit of a hundred buckets that you can have in an AWS account. So this is not per region, it's for the entire account. There's a soft limit of 100 and a hard limit so you can increase all the way up to this hard limit using support requests, and this hard limit is a thousand.

      Now this matters for architectural reasons. It's not just an arbitrary number. If you're designing a system which uses S3 and users of that system store data inside S3, you can implement a solution that has one bucket per user if you have anywhere near this number of users. So if you have anywhere from a hundred to a thousand users or more of a system, then you're not gonna be able to have one bucket per user because you'll hit this hard limit. You tend to find this in the exam quite often, it'll ask you how to structure a system, which has potentially thousands of users. What you can do is take a single bucket and divide it up using prefixes, so those folders that aren't really folders, and then in that way, you can have multiple users using one bucket. Remember the 100/1000, it's a 100 soft limit and a 1000 hard limit.

      You aren't limited in terms of objects in a bucket, you can have zero to an infinite number of objects in a bucket. And each object can range in size from zero bytes to 5 TB in size. And then finally, in terms of the object structure, an object consists of a key, which is its name and then the value, which is the data. And there are other elements to an object which I'll discuss later in the course, but for now, just remember the two main components, the key and the value. Now, all of these points are worth noting down, maybe make them into a set of flashcards and you can use them later on during your studies.

      S3 is pretty straightforward and that there tend to be a number of things that it's really good at and a fairly small set of things that it's not suitable for. So let's take a look. S3 is an object storage system. It's not a file storage system, and it's not a block storage system, which are the other main types. What this means is that if you have a requirement where you're accessing the whole of these entities, so the whole of an object, so an image, an audio file, and you're doing all of that at once, then it's a candidate for object storage. If you have a Window server which needs to access a network file system, then it's not S3 that needs to be file-based storage. S3 has no file system, it's flat. So you can't browse to an S3 bucket like you would a file share in Windows. Likewise, it's not block storage, which means you can't mount it as a mount point or a volume on the Linux or Windows. When you're dealing with virtual machines or instances, you mount block storage to them. Block storage is basically virtual hard disks. In EC2, you have EBS, which is block storage. Block storage is generally limited to one thing accessing it at a time, one instance in the case of EBS. S3 doesn't have that single user limitation and it's not block storage, but that means you can't mount it as a drive.

      S3 is great for large scale data storage or distribution. Many examples I'll show you throughout the course will fit into that category. And it's also good for offloading things. If you have a blog with lots of posts and lots of images or audio or movies, instead of storing that data on an expensive compute instance, you can move it to an S3 bucket and configure your blog software to point your users at S3 directly. You can often shrink your instance by offloading data onto S3. And don't worry, I'll be demoing this later in the course. Finally, S3 should be your default thought for any input to AWS services or output from AWS services. Most services which consume data and or output data can have S3 as an option to take data from or put data to when it's finished. So if you're faced with any exam questions and there's a number of options on where to store data, S3 should be your default. There are plenty of AWS services which can output large quantities of data or ingest large quantities of data. And most of the time, it's S3, which is an ideal storage platform for that service.

      Okay time for a quick demo. And in this demo, we're just gonna run through the process of creating a simple S3 bucket, uploading some objects to that bucket, and demonstrating exactly how the folder functionality works inside S3. And I'm also gonna demonstrate a number of elements of how access and permissions work with S3. So go ahead and complete this video, and when you're ready join me in the next, which is gonna be a demo of S3.

    1. Welcome back. In this lesson, we will introduce an essential AWS service, the Elastic Compute Cloud (EC2). EC2 is a critical service you need to understand for the AWS exam and for designing real-world AWS solutions, as it is the default compute service within AWS. EC2 provides access to virtual machines known as instances. When deploying compute resources that require an operating system, runtime environment, database dependencies, applications, and application interfaces, EC2 is the service to use. It should be the starting point for any compute requirement in AWS.

      Let's review some important features of EC2 before discussing architectural points. EC2 is an Infrastructure as a Service (IaaS), providing access to virtual machines called EC2 instances. As an IaaS, the unit of consumption is the instance, which is an operating system configured with allocated resources. EC2 is a private AWS service, meaning it runs in the private AWS zone by default. An EC2 instance is configured to launch into a single VPC subnet, set during instance launch, and requires configuration for public access if needed.

      With the default VPC, public access configuration is handled automatically. For custom VPCs, you need to manage this configuration. An EC2 instance's availability zone (AZ) resilience means that if the AZ fails, the instance will likely fail too. Understanding this is crucial for the course and the exam. When launching an instance, you can choose from various sizes and capabilities, influencing the resources and additional features like GPUs, advanced storage, or networking. Some configurations can be changed post-launch.

      As a consumer of EC2, you manage the operating system and upward in the infrastructure stack, while AWS handles virtualization, physical hardware, networking, storage, and facilities. EC2 offers on-demand billing, either by the second or hour, depending on the instance's software, charging only for consumed resources. Instance charges include CPU and memory consumption, storage usage, and extras for any commercial software.

      EC2 instances can use different storage types, including local host storage and Elastic Block Store (EBS), a network storage service. An EC2 instance's state, such as running, stopped, or terminated, indicates its condition. Running instances consume resources and generate charges, while stopped instances do not incur CPU, memory, or networking costs but still generate storage charges. Terminated instances stop all resource usage and delete allocated storage, making this action irreversible.

      An Amazon Machine Image (AMI) is an image of an EC2 instance used to create or derived from an instance. AMIs contain attached permissions, determining which accounts can use them. They can be public, owner-accessible, or explicitly shared with specific AWS accounts. AMIs include the boot volume and a block device mapping, linking volumes to the device IDs expected by the operating system.

      In the next lesson, we will connect to EC2 instances. Windows instances use the Remote Desktop Protocol (RDP) on port 3389, while Linux instances use the SSH protocol on port 22, authenticated with an SSH key pair. When creating an EC2 instance, you select a key pair, download the private key, and AWS retains the public key. For Windows instances, the private key retrieves the local administrator password for RDP access. For Linux instances, the private and public key match enables SSH connection.

      Next, we will demonstrate creating an EC2 instance in the default VPC, assigning a public IPv4 address, and connecting using the discussed methods. Finish this video and join me in the demo lesson when ready.

    1. Author response:

      “Overall, the paper has several strengths, including leveraging large-scale, multi-modal datasets, using computational reasonable tools, and having an in-depth discussion of the significant results.”

      We thank the reviewer for the very supportive comments.

      Based on the comments and questions, we have grouped the concerns and corresponding responses into three categories.

      (1) The scope and data selection

      “The results are somewhat inconclusive or not validated.

      The overall results are carefully designed, but most of the results are descriptive. While the authors are able to find additional evidence either from the literature or explain the results with their existing knowledge, none of the results have been biologically validated. Especially, the last three result sections (signaling pathways, eQTLs, and TF binding) further extended their findings, but the authors did not put the major results into any of the figures in the main text.”

      The goal of this manuscript is to provide a list of putative childhood obesity target genes to yield new insights and help drive further experimentation. Moreover, the outputs from signaling pathways, eQTLs, and TF binding, although noteworthy and supportive of our method, were not particularly novel. In our manuscript we placed our focus on the novel findings from the analyses. We did, however, report the part of the eQTLs analysis concerning ADCY3, which brought new insight to the pathology of obesity, in Figure 4C.

      “The manuscript would benefit from an explanation regarding the rationale behind the selection of the 57 human cell types analyzed. it is essential to clarify whether these cell types have unique functions or relevance to childhood development and obesity.”

      We elected to comprehensively investigate the GWAS-informed cellular underpinnings of childhood development and obesity. By including a diverse range of cell types from different tissues and organs, we sought to capture the multifaceted nature of cellular contributions to obesity-related mechanisms, and open new avenues for targeted therapeutic interventions.

      There are clearly cell types that are already established as being key to the pathogenesis of obesity when dysregulated: adipocytes for energy storage, immune cell types regulating inflammation and metabolic homeostasis, hepatocytes regulating lipid metabolism, pancreatic cell types intricately involved in glucose and lipid metabolism, skeletal muscle for glucose uptake and metabolism, and brain cell types in the regulation of appetite, energy expenditure, and metabolic homeostasis.

      While it is practical to focus on cell types already proven to be associated with or relevant to obesity, this approach has its limitations. It confines our understanding to established knowledge and rules out the potential for discovering novel insights from new cellular mechanisms or pathways that could play significant roles in the pathogenesis if obesity. Therefore, it is was essential to reflect known biology against the unexplored cell types to expand our overall understanding and potentially identify innovative targets for treatment or prevention.

      “I wonder whether the used epigenome datasets are all from children. Although the authors use literature to support that body weight and obesity remain stable from infancy to adulthood, it remains uncertain whether epigenomic data from other life stages might overlook significant genetic variants that uniquely contribute to childhood obesity.”

      The datasets utilized in our study were derived from a combination of sources, both pediatric and adult. We recognize that epigenetic profiles can vary across different life stages but our principal effort was to characterize susceptibility BEFORE disease onset.

      “Given that the GTEx tissue samples are derived from adult donors, there appears to be a mismatch with the study's focus on childhood obesity. If possible, identifying alternative validation strategies or datasets more closely related to the pediatric population could strengthen the study's findings.” 

      We thank the reviewer for raising this important point. We acknowledge that the GTEx tissue samples are derived from adult donors, which might not perfectly align with the study's focus on childhood obesity. The ideal strategy would be a longitudinal design that follows individuals from childhood into adulthood to bridge the gap between pediatric and adult data, offering systematic insights into how early-life epigenetic markers influencing obesity later in life. In future work, we aim to carry out such efforts, which will represent substantial time and financial commitment.

      Along the same lines, the Developmental Genotype-Tissue Expression (dGTEx) Project is a new effort to study development-specific genetic effects on gene expression at 4 developmental windows spanning from infant to post-puberty (0-18 years). Donor recruitment began in August 2023 and remains ongoing. Tissue characterization and data production are underway. We hope that with the establishment of this resource, our future research in the field of pediatric health will be further enhanced.

      “Figure 1B: in subplots c and d, the results are either from Hi-C or capture-C. Although the authors use different colors to denote them, I cannot help wondering how much difference between Hi-C and capture-C brings in. Did the authors explore the difference between the Hi-C and capture-C?”.

      Thank you for your comment. It is not within the scope of our paper to explore the differences between the Hi-C and Capture-C methods. In the context of our study, both methods serve the same purpose of detecting chromatin loops that bring putative enhancers to sometimes genomically distant gene promoters. Consequently, our focus was on utilizing these methods to identify relevant chromatin interactions rather than comparing their technical differences.

      (2) Details on defining different categories of the regions of interest

      “Some technical details are missing.

      While the authors described all of their analysis steps, a lot of the time, they did not mention the motivation. Sometimes, the details were also omitted.”

      We will add a section to the revision to address the rationale behind different OCRs categories.

      “Line 129: should "-1,500/+500bp" be "-500/+500bp"? 

      A gene promoter was defined as a region 1,500 bases upstream to 500 bases downstream of the TSS. Most transcription factor binding sites are distributes upstream (5’) from TSS, and the assembly of transcription machinery occurs up to 1000 bases 5’ from TSS. Given our interest in SNPs that can potentially disrupt transcription factor binding, this defined promoter length allowed us to capture such SNPs in our analyses.

      “How did the authors define a contact region?”

      Chromatin contact regions identified by Hi-C or Capture-C assays are always reported as pairs of chromatin regions. The Supplementary eMethods provide details on the method of processing and interaction calling from the Hi-C and Capture-C data.

      “The manuscript would benefit from a detailed explanation of the methods used to define cREs, particularly the process of intersecting OCRs with chromatin conformation data. The current description does not fully clarify how the cREs are defined.”

      “In the result section titled "Consistency and diversity of childhood obesity proxy variants mapped to cREs", the authors introduced the different types of cREs in the context of open chromatin regions and chromatin contact regions, and TSS. Figure 2A is helpful in some way, but more explanation is definitely needed. For example, it seems that the authors introduced three chromatin contacts on purpose, but I did not quite get the overall motivation.”

      We apologize for the confusion. Our definition of cREs is consistent throughout the study. Figure 2A will be the first Figure 1A in the revision in order to aid the reader.

      The 3 representative chromatin loops illustrate different ways the chromatin contact regions (pairs of blue regions under blue arcs) can overlap with OCRs (yellow regions under yellow triangles – ATAC peaks) and gene promoters.

      [1] The first chromatin loop has one contact region that overlaps with OCRs at one end and with the gene promoter at the other. This satisfies the formation of cREs; thus, the area under the yellow ATAC-peak triangle is green.

      [2] The second loop only overlapped with OCR at one end, and there was no gene promoter nearby, so it is unqualified as cREs formation.

      [3] The third chromatin loop has OCR and promoter overlapping at one end. We defined this as a special cRE formation; thus, the area under the yellow ATAC-peak triangle is green.

      To avoid further confusion for the reader, we will eliminate this variation in the new illustration for the revised manuscript.

      “Figure 2A: The authors used triangles filled differently to denote different types of cREs but I wonder what the height of the triangles implies. Please specify.”

      The triangles are illustrations for ATAC-seq peaks, and the yellow chromatin regions under them are OCRs. The different heights of ATAC-seq peaks are usually quantified as intensity values for OCRs. However, in our study, when an ATAC-seq peak passed the significance threshold from the data pipeline, we only considered their locations, regardless of their intensities. To avoid further confusion for the reader, we will eliminate this variation in the new illustration for the revised manuscript.

      “Figure 1B-c. the title should be "OCRs at putative cREs". Similarly in Figure 1B-d.”

      cREs are a subset of OCRs.

      - In the section "Cell type specific partitioned heritability", the authors used "4 defined sets of input genomic regions". Are you corresponding to the four types of regions in Figure 2A? 

      Figure 2A will be the first Figure 1A in the revision and will be modified to showcase how we define OCRs and cREs.

      “It seems that the authors described the 771 proxies in "Genetic loci included in variant-to-genes mapping" (ln 154), and then somehow narrowed down from 771 to 94 (according to ln 199) because they are cREs. It would be great if the authors could describe the selection procedure together, rather than isolated, which made it quite difficult to understand.”

      In the Methods section entitled “Genetic loci included in variant-to-genes mapping," we described the process of LD expansion to include 771 proxies from 19 sentinel obesity-significantly associated signals. Not all of these proxies are located within our defined cREs. Figure 2B, now Figure 2A in the revision, illustrates different proportions of these proxies located within different types of regions, reducing the proxy list to 94 located within our defined cREs.

      “Figure 2. What's the difference between the 771 and 758 proxies? “

      13 out of 771 proxies did not fall within any defined regions. The remaining 758 were located within contact regions of at least one cell type regardless of chromatin state.

      (3) Typos

      “In the paragraph "Childhood obesity GWAS summary statistics", the authors may want to describe the case/control numbers in two stages differently. "in stage 1" and "921 cases" together made me think "1,921" is one number.”

      This will be amended in the revision.

      “Hi-C technology should be spelled as Hi-C. There are many places, it is miss-spelled as "hi-C". In Figure 1, the author used "hiC" in the legend. Similarly, Capture-C sometime was spelled as "capture-C" in the manuscript.”

      “At the end of the fifth row in the second paragraph of the Introduction section: "exisit" should be "exist".

      “In Figure 2A: "Within open chromatin contract region" should be "Within open chromatin contact region". 

      These typos and terminology inconsistencies will be amended in the revision.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      The development of effective computational methods for protein-ligand binding remains an outstanding challenge to the field of drug design. This impressive computational study combines a variety of structure prediction (AlphaFold2) and sampling (RAVE) tools to generate holo-like protein structures of three kinases (DDR1, Abl1, and Src kinases) for binding to type I and type II inhibitors. Of central importance to the work is the conformational state of the Asp-Phy-Gly "DFG motif" where the Asp points inward (DFG-in) in the active state and outward (DFG-out) in the inactive state. The kinases bind to type I or type II inhibitors when in the DFG-in or DFG-out states, respectively.

      It is noted that while AlphaFold2 can be effective in generating ligand-free apo protein structures, it is ineffective at generating holo-structures appropriate for ligand binding. Starting from the native apo structure, structural fluctuations are necessary to access holo-like structures appropriate for ligand binding. A variety of methods, including reduced multiple sequence alignment (rMSA), AF2-cluster, and AlphaFlow may be used to create decoy structures. However, those methods can be limited in the diversity of structures generated and lack a physics-based analysis of Boltzmann weight critical to their relative evaluation.

      To address this need, the authors combine AlphaFold2 with the Reweighted Autoencoded Variational Bayes for Enhanced Sampling (RAVE) method, to explore metastable states and create a Boltzmann ranking. With that variety of structures in hand, grid-based docking methods Glide and Induced-Fit Docking (IFD) were used to generate protein-ligand (kinase-inhibitor) complexes.

      The authors demonstrate that using AlphaFold2 alone, there is a failure to generate DFG-out structures needed for binding to type II inhibitors. By applying the AlphaFold2 with rMSA followed by RAVE (using short MD trajectories, SPIB-based collective variable analysis, and enhanced sampling using umbrella sampling), metastable DFG-out structures with Boltzmann weighting are generated enabling protein-ligand binding. Moreover, the authors found that the successful sampling of DFG-out states for one kinase (DDR1) could be used to model similar states for other proteins (Abl1 and Src kinase). The AF2RAVE approach is shown to result in a set of holo-like protein structures with a 50% rate of docking type II inhibitors.

      Overall, this is excellent work and a valuable contribution to the field that demonstrates the strengths and weaknesses of state-of-the-art computational methods for protein-ligand binding. The authors also suggest promising directions for future study, noting that potential enhancements in the workflow may result from the use of binding site prediction models and free energy perturbation calculations.

      Reviewer #2 (Public Review):

      Summary:

      This manuscript explores the utility of AlphaFold2 (AF2) and the author's own AF2-RAVE method for drug discovery. As has been observed elsewhere, the predictive power of docking against AF2 structures is quite limited, particularly for proteins like kinases that have non-trivial conformational dynamics. However, using enhanced sampling methods like RAVE to explore beyond AF2 starting structures leads to a significant improvement.

      Strengths:

      This is a nice demonstration of the utility of the authors' previously published RAVE method.

      Weaknesses:

      My only concern is the authors' discussion of induced fit. I'm quite confident the structures discussed are present in the absence of ligand binding, consistent with conformational selection. It seems the author's own data also argues for an important role in conformational selection. It would be nice to acknowledge this instead of going along with the common practice in drug discovery of attributing any conformational changes to induced fit without thoughtful consideration of conformational selection.

      The reviewer is correct. We aim to highlight the significant role of conformational selection. To clarify this, we have expanded the discussion on conformational selection in the introduction.

      Reviewer #3 (Public Review):

      In this manuscript, the authors aim to enhance AlphaFold2 for protein conformation-selective drug discovery through the integration of AlphaFold2 and physics-based methods, focusing on improving the accuracy of predicting protein structures ensemble and small molecule binding of metastable protein conformations to facilitate targeted drug design.

      The major strength of the paper lies in the methodology, which includes the innovative integration of AlphaFold2 with all-atom enhanced sampling molecular dynamics and induced fit docking to produce protein ensembles with structural diversity. Moreover, the generated structures can be used as reliable crystal-like decoys to enrich metastable conformations of holo-like structures. The authors demonstrate the effectiveness of the proposed approach in producing metastable structures of three different protein kinases and perform docking with their type I and II inhibitors. The paper provides strong evidence supporting the potential impact of this technology in drug discovery. However, limitations may exist in the generalizability of the approach across other structures, especially complex structures such as protein-protein or DNA-protein complexes.

      Proteins undergo thermodynamic fluctuations and can occasionally reach metastable configurations. It can be assumed that other biomolecules, such as proteins and DNA, stabilize these metastable states when forming protein-protein or protein-DNA complexes. Since our method has the potential to identify these metastable states, it shows promise for designing drugs targeting proteins in allosteric configurations induced by other biomolecules.

      The authors largely achieved their aims by demonstrating that the AF2RAVE-Glide workflow can generate holo-like structure candidates with a 50% successful docking rate for known type II inhibitors. This work is likely to have a significant impact on the field by offering a more precise and efficient method for predicting protein structure ensemble, which is essential for designing targeted drugs. The utility of the integrated AF2RAVE-Glide approach may streamline the drug discovery process, potentially leading to the development of more effective and specific medications for various diseases.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      Suggestions

      (1) The computational protocol is found to be insufficient to generate precise values of the relative free energies between structures generated. The authors note in the Conclusion that an enhancement in the workflow might result from the addition of free energy calculations. Can the authors comment on the prospects for generating more accurate estimates of the free energy that might be used to qualitatively evaluate poses and the free energy landscape surrounding putative metastable states? What are the principal challenges and what might help overcome them? What would the most effective computational protocol be?

      More accurate estimates of the free energy can theoretically be achieved by increasing the number of umbrella sampling windows and extending the simulation length until the PMF converges. However, there is always a trade-off between PMF accuracy and computational costs, so we have chosen to stick with the current setup. Metadynamics is another method to obtain a more accurate free energy profile, which we have used in previous versions of AlphaFold2-RAVE, but for the specific systems we investigated, it had issues in achieving back and forth movement given the high entropic nature of the activation loop. Research in enhanced sampling methods and dimensionality reduction techniques for reaction coordinates is continually evolving and will play a critical role in alleviating this problem.

      (2) I was surprised that there was not more correlation of a funnel-like shape in Figures S16 and S18, showing a stronger correlation between low RMSD and better docking score. This is true for both the ponatinib and imatinib applications in DDR1 and Abl1. That also seems true for the trimmed results for Src kinase in Figure S19. I was also surprised that there are structures with very large RMSD but docking scores comparable to the best structures of the lowest RMSD. Might something be done to make the docking score a more effective discriminator?

      The docking algorithm and docking score are used to filter out highly improbable docking poses. False positives in predicted docking poses are a common issue across all docking methods as described for instance in:

      Fan, Jiyu, Ailing Fu, and Le Zhang. "Progress in molecular docking." Quantitative Biology 7 (2019): 83-89.

      Ferreira, R.S., Simeonov, A., Jadhav, A., Eidam, O., Mott, B.T., Keiser, M.J., McKerrow, J.H., Maloney, D.J., Irwin, J.J. and Shoichet, B.K., 2010. "Complementarity between a docking and a high-throughput screen in discovering new cruzain inhibitors." Journal of medicinal chemistry, 53(13), pp.4891-4905.

      Moreover, there is always a trade-off between docking accuracy and computational cost. While employing more accurate docking methods may decrease false positives, it can also be resource-intensive. In such scenarios, our approach to enriching holo-structures can be impactful by reducing the number of pocket structures in the input ensembles and significantly enhancing docking efficiency.

      (3) I think that it is fine to identify one structure as "IFD winner" but also feel that its significance is overstressed, especially given that it can be identified only in a retrospective analysis rather than through de novo prediction.

      We agree with the reviewer. We did not intend to emphasize the specific structure "IFD winner". Rather, we aimed to demonstrate that our method can enrich promising candidates for holo-structures. We verified this by showing that our holo-structure candidates performed well in retrospective docking using IFD, which we previously referred to as "IFD winner". We have now revised this term to "holo-model".

      Minor Points

      p. 3 "DymanicBind" should be "DynamicBind"

      p. 3 Change "We chosen" to "We have chosen" or "we chose."

      p. 3 In identifying the Schrödinger software Glide and IFD, I recommend removing the subjective modifier "industry-leading."

      Modifications done.

      Reviewer #2 (Recommendations For The Authors):

      In the view of this reviewer, the writing is 'choppy'.

      We have tried to improve the writing.

      Reviewer #3 (Recommendations For The Authors):

      (1) In Figure 1, the workflow labels (i) to (iv) are not shown on the figures, making it difficult for readers to follow. Consider adding these labels to the figures.

      Modifications done.

      (2) Explain how Boltzmann ranks were calculated based on unbiased MD simulations to guide the enrichment of holo-like structures in metastable states.

      The Methods section is now updated for clarification.

      (3) The authors could clarify how the classical DFG-out decoys in the DDR1 rMSA AF2 ensemble are transferred to Abl1 kinase in the Methods section.

      The Methods section is now updated for clarification.

      (4) The authors can clarify the methodology section by providing more detailed explanations about how the unbiased MD simulations are performed, including which MD simulation software was used and whether energy minimization and equilibrium steps were needed as in conventional MD simulations, and other setup details.

      The Methods section is now updated for clarification.

      (5) The validation of the proposed approach in this work used three kinase proteins. The authors can enhance the discussion section by addressing other types of protein structure prediction that can use the proposed approach in drug discovery, beyond the three kinase proteins tested.

      The proposed approach is theoretically applicable to other types of proteins, such as GPCRs, where both conformational selection and the induced-fit effect are crucial. We have expanded the discussion on the generalization of our protocol in the Conclusion section.

      (6) The authors should add appropriate citations for the software and tools used in the manuscript. For example, a reference should be added for the Glide XP docking experiments that utilized the Maestro software. Double-check all related software citations.

      We have now updated the citations for docking experiments based on the instruction of the Maestro Glide User manual and IFD User manual.

      (7) The authors should consider offering a comprehensive list of software tools and databases utilized in the study to assist in replicating the experiments and further validating the results.

      We have now added a summary of tools used in the Methods section.

  5. Jul 2024
    1. several everyday programs such as SAPGUI. You can do so by opening the Software Center or Software Corner from the Windows start menu.

      It might be helpful to have a list of the everyday programs the user will need to install, along with links to their corresponding pages in the help portal.

    1. El entorno de investigación reproducible para esta tesis

      El entorno de investigación reproducible para esta tesis articuló un conjunto de herramientas que permitiesen la instación de software, el control de versiones históricas del texto de la tesis y otros productos de investigación asociados, la colaboración con el tutor y la publicación de productos intermedios, así como la exploración de los datos extraídos y su visualización.

      Este conjunto de herramientas se eligió sobre otras que podrían funcionar para el mismo propósito, debido a su simpleza y flexibilidad. Por ejemplo, Fossil es más sencillo de usar que Git y tiene más funcionalidades integradas (ver "[Fossil versus Git)[https://www.fossil-scm.org/home/doc/trunk/www/fossil-v-git.wiki]" y Lepiter+Grafoscopio/MiniDocs provee mayor flexibilidad de personalización sobre otras herramientas más populares como los Jupyter Notebooks (Luna 2019, cap tal).

      La selección de herramientas se hizo de manera tal que la escritura de la tesis, incluyendo el análisis de datos se pudiera hacer en distintos computadores con diferentes capacidades de procesamiento, incluyendo computadores familiares (Mac y Windows), así como computadores del LabCI. Esto acarrea cierta redudancia de herramientas como Zettlr y LiteXL, ambos usados para escritra en Markdown, pero el segundo más ligero y que funciona en máquinas más modestas.

    1. Reviewer #2 (Public Review):

      Summary:

      Spargo and colleagues present an analysis of the shared genetic architectures of Schizoprehnia and several late-onset neurological disorders. In contrast to many polygenic traits for which global genetic correlation estimates are substantial, global genetic correlation estimates for neurological conditions are relatively small, likely for several reasons. One is that assortative mating, which will spuriously inflate genetic correlation estimates, is likely to be less salient for late-onset conditions. Another, which the authors explore in the current manuscript, is that some loci affecting two or more conditions (i.e., pleiotropic loci) may have effects in opposite directions, or shared loci are sparse, such that the global genetic correlation signal washes out.

      The authors apply a local genetic correlation approach that assesses the presence and direction of pleiotropy in much smaller spatial windows across the genome. Then, within regions evidencing local genetic correlations for a given trait pair, they apply fine-mapping and colocalization methods to attempt to differentiate between two scenarios: that the two traits share the same causal variant in the region or that distinct loci within the region influence the traits. Interestingly, the authors only discover one instance of the former: an SNP in the HLA region appearing to confer risk for both AD and ALS. This is in contrast to six regions with distinct causal loci, and twenty regions with no clear shared loci.

      Finally, the authors have published their analysis pipeline such that other researchers might easily apply the same techniques to other collections of traits.

      Strengths:<br /> - All such analysis pipelines involve many decision points where there is often no clear correct option. Nonetheless, the authors clearly present their reasoning behind each such decision.<br /> - The authors have published their analytic pipeline such that future researchers might easily replicate and extend their findings.

      Weaknesses:<br /> - The majority of regions display no clear candidate causal variants for the traits, whether shared or distinct. Further, despite the potential of local genetic correlation analysis to identify regions with effects in opposing directions, all of the regions for causal variants were identified for both traits evidenced positive correlations. The reasons for this aren't clear and the authors would do well to explore this in greater detail.<br /> - The authors very briefly discuss how their findings differ from previous analyses because of their strict inclusion for "high-quality" variants. This might be the case, but the authors do not attempt to demonstrate this via simulation or otherwise, making it difficult to evaluate their explanation.

      These concerns were addressed in the revised version of this manuscript.

    2. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews: 

      Reviewer #1 (Public Review): 

      The authors investigate pleiotropy in the genetic loci previously associated to a range of neuropsychiatric disorders: Alzheimer's disease, amyotrophic lateral sclerosis (ALS), frontotemporal dementia, Parkinson's disease, and schizophrenia. The local statistical fine-mapping and variant colocalisation approaches they use have the potential to uncover not only shared loci but also shared causal variants between these disorders. There is existing literature describing the pleiotropy between ALS and these other disorders but here the authors apply state of the art, local genetic correlation approaches to further refine any relationships. 

      Complex disease and GWAS is not my area of expertise but the authors managed to present their methods and results in a clear, easy to follow manner. Their results statistically support several correlations between the disorders and, for ALS and AD, a shared variant in the vicinity of the lead SNP from the original ALS GWAS. Such findings could have important implications for our understanding of the mechanisms of such disorders and eventually the possibility of managing and treating them. 

      The authors have built a useful pipeline that plugs together all the gold-standard, existing software to perform this analysis and made it openly available which is commendable. However, there is little discussion of what software is available to perform global and local correlation analysis and, if there are multiple tools available, why they consider the ones they selected to be the gold-standard. 

      There is some mention of previous findings of genetic pleiotropy between ALS and these other disorders in the introduction, and discussion of their improved ALS-AD evidence relative to previous work. However, detailed comparisons of their other correlations to what was described before for the same pairs of disorders (if any) is missing. Adding this would strengthen the impact of this paper. 

      Finally, being new to this approach I found the abstract a little confusing. Initially, the shared causal variant between ALS and AD is mentioned but immediately in the following sentence they describe how their study "suggested that disease- implicated variants in these loci often differ between traits". After reading the whole paper I understood that the ALS-AD shared variant was the exception but it may be best to restructure this part of the abstract. Additionally, in the abstract the authors state that different variants "suggests the role of distinct mechanisms across diseases despite shared loci". Is it not possible that different variants in the same regulatory region or protein-coding parts of a gene could be having the same effect and mechanism? Or does the methodology to establish that different variants are involved automatically mean that the variants are too distant for this to be possible? 

      We thank reviewer one for their considered review of this manuscript and for highlighting points that would benefit from further exploration. Itemised responses are provided below.

      (1) The reviewer noted that we did not adequately explain our choice of software for global and local genetic correlation analysis, and why we consider the techniques chosen as gold standard. We agree that the paper would benefit from clarification around this aspect of the study.

      Briefly, we firstly selected LAVA for the local genetic correlation analysis because it offers several advantages above competing software and was developed by a reputable team previously known for developing MAGMA, which is well-established in the statistical genetics field. In the manuscript (page 8), we added the following clarification: “LAVA was the most appropriate local genetic correlation approach for this study for several reasons. First, unlike SUPERGNOVA and rho-HESS, LAVA makes specific accommodations for analysis of binary traits. Second, other tools focus on bivariate correlation between traits whilst LAVA offers this alongside multivariate tests such as multiple regression and partial correlation, enabling rigorous testing of pleiotropic effects. Lastly, LAVA is shown to provide results which are less biased than those from other tools.”

      LDSC was selected for the global genetic correlation analysis because the software is well-established and likely the most widely adopted global genetic correlation tool. Reflecting its prevalence, the software is also compatible with LAVA, which adjusts for sample overlap based on the bivariate intercept estimate returned by LDSC. Since global genetic correlations were not the primary focus of this study, having been tested across several previous investigations (see response 2), we did not prioritise comparison of correlation estimates from LDSC against other available software. In the manuscript (pages 7-8) we now include the following statement: “[LDSC] was also applied to derive ‘global’ (i.e., genome-wide) genetic correlation estimates between trait pairs and estimate sample overlap from the bivariate intercept. The latter of these outputs was taken forward as an input for the local genetic correlation analysis using LAVA (see 2.2.2.2). Since global genetic correlation analysis across the traits studied here is not novel and associations reported in past studies are congruent across different tools, the compatibility between LDSC and LAVA motivated our use of LDSC for this analysis”.

      (2) The second comment was that the paper would be strengthened by contextualising our study with detail around what is previously known about associations between the studied traits. Accordingly, we have added clarifying text at the end of the introduction, stating: “although previous studies have performed global genetic correlation analyses between various combinations of these traits {references}, this is the first to compare them at a genome-wide scale using a local genetic correlation approach“. In the discussion, we link back to these studies, stating that “Through genetic correlation analysis, we replicated genome-wide correlations previously described between the studied traits {references}”.

      (3) The reviewer highlighted that the abstract as originally written may mislead or confuse the reader and we agree that clarity could be improved with some restructuring. This has now been revised and should read more logically.

      (4) They also enquired about our reasons for suggesting that the implication of distinct variants for each trait from a colocalisation analysis suggests a distinct causal mechanism. We thank them for this question as it encouraged us to reconsider how best to present the results of this analysis. To answer their question:

      It is certainly true that nearby but distinct variants can confer the same effect. In a scenario where multiple distinct variants result in the same effect and thus increase susceptibility towards two or more related phenotypes, you would expect to find evidence of association to each relevant variant in GWAS across these related traits (even if the magnitude of the associations differ). Where biological mechanisms are shared, post-GWAS finemapping analysis would be expected to yield credible sets overlapping across the traits, and likewise, colocalisation analysis should converge on a set of credible SNPs that are candidates for the shared effect. Where multiple distinct variants confer the same effect, you would expect to see separate fine-mapping credible sets for these distinct variants that colocalise pairwise between the jointly-affected traits. Generally, therefore, evidence supporting the two distinct variants hypothesis would suggest the role of two distinct mechanisms except when certain credible sets identified through fine-mapping converge on a colocalised effect.

      There is a further caveat which we also explored in response to Reviewer two: if a region includes long-spanning LD (and hence a larger number of variants are considered in the analysis), then the colocalisation analysis is more likely to favour the two distinct variants hypothesis since the probability of the variants implicated in both traits being shared decreases. It is likely that support for the two independent variants hypothesis is correct in most of the comparisons from this study that favour this conclusion. This is because, generally, the fine-mapping credible sets do not overlap across trait pairs (Figure S4) and consequently the colocalisation analysis does not find any support for the shared variant hypothesis. An exception is the analysis of PD and schizophrenia at the MAPT locus on chromosome 17. We have accordingly added the following clarification to the (page 18): “However, the colocalisation analysis will increasingly favour the two independent variants hypothesis as the number of analysed variants increases. Hence, the wide-spanning LD of this region may have obstructed identification of variants and mechanisms shared between the traits.”

      Reviewer #2 (Public Review): 

      Summary: 

      Spargo and colleagues present an analysis of the shared genetic architectures of Schizoprehnia and several late-onset neurological disorders. In contrast to many polygenic traits for which global genetic correlation estimates are substantial, global genetic correlation estimates for neurological conditions are relatively small, likely for several reasons. One is that assortative mating, which will spuriously inflate genetic correlation estimates, is likely to be less salient for late-onset conditions. Another, which the authors explore in the current manuscript, is that some loci affecting two or more conditions (i.e., pleiotropic loci) may have effects in opposite directions, or shared loci are sparse, such that the global genetic correlation signal washes out. 

      The authors apply a local genetic correlation approach that assesses the presence and direction of pleiotropy in much smaller spatial windows across the genome. Then, within regions evidencing local genetic correlations for a given trait pair, they apply fine-mapping and colocalization methods to attempt to differentiate between two scenarios: that the two traits share the same causal variant in the region or that distinct loci within the region influence the traits. Interestingly, the authors only discover one instance of the former: an SNP in the HLA region appearing to confer risk for both AD and ALS. This is in contrast to six regions with distinct causal loci, and twenty regions with no clear shared loci. 

      Finally, the authors have published their analysis pipeline such that other researchers might easily apply the same techniques to other collections of traits. 

      Strengths: 

      - All such analysis pipelines involve many decision points where there is often no clear correct option. Nonetheless, the authors clearly present their reasoning behind each such decision. <br /> - The authors have published their analytic pipeline such that future researchers might easily replicate and extend their findings. 

      Weaknesses:

      - The majority of regions display no clear candidate causal variants for the traits, whether shared or distinct. Further, despite the potential of local genetic correlation analysis to identify regions with effects in opposing directions, all of the regions for causal variants were identified for both traits evidenced positive correlations. The reasons for this aren't clear and the authors would do well to explore this in greater detail. 

      - The authors very briefly discuss how their findings differ from previous analyses because of their strict inclusion for "high-quality" variants. This might be the case, but the authors do not attempt to demonstrate this via simulation or otherwise, making it difficult to evaluate their explanation. 

      We thank Reviewer two for their appraisal of this manuscript and kind comments regarding its strengths. We will now aim to address the identified weaknesses.

      (1) The reviewer comments that we did not adequately investigate why loci with causal variants identified in both traits all had positive local genetic correlations. We agree that it would be helpful to better understand the underlying reasons. To address this issue, we have added a new supplementary figure to compare the positive and negative local genetic correlation results (see Figure S2). In the main-text we add the following clarification. ”Although both positive and negative local genetic correlations passed the FDR-adjusted significance threshold, we observed only positive local genetic correlations in loci where fine-mapping credible sets were identified for both traits in the pair. This reflects that the correlation coefficients and variant associations from the analysed GWAS studies were generally stronger in the positively correlated loci (see Figure S2).”

      (2) The reviewer rightly suggests that the manuscript would benefit from an improved explanation of the somewhat inconsistent results for the colocalisation analysis of ALS and AD at the locus around the rs9275477 SNP from this work and a previous study.  We have now further investigated this and believe that the discrepancy results partly from an inherent empirical characteristic of the colocalisation analysis. We have explained this in the manuscript (page 22) as follows: “The previous study analysed a 200Kb window of over 2,000 SNPs around the lead genome-wide significant SNP from the ALS GWAS, rs9275477, and found ~0.50 posterior probability for each of the shared and two independent variant(s) hypotheses. The current analysis used 475 SNPs occurring within a semi-independent LD block of ~50kb in this locus. Since the posterior probability of the two independent variants hypothesis (H3) increases exponentially with the number of variants in the region whilst the shared variant hypothesis (H4) scales linearly, it is expected that our analysis would give stronger support for the latter. Given that the previous study defined regions for analysis based on an arbitrary window of ±100kb around each lead genome-wide significant SNP from the ALS GWAS and we defined each analysis region based on patterns of LD in European ancestry populations, it is reasonable to favour the current finding.”

    1. Welcome back, this is part two of this lesson where we continue immediately from the end of part one. In my case, it took about three minutes for the status check to change to "two out of two checks passed," indicating we're good to continue. AWS runs several checks on an EC2 instance as the final stage of the provisioning process, so ensure an instance has passed these checks before proceeding. You can see the instance has been provisioned into US-East-1C, with its own public IP version 4 address and public IP version 4 DNS name for internet access. Additional details such as the security group name, key name, and launch time are also available. Clicking on the instance reveals more detailed information in the bottom pane across various tabs like Details, Security, Networking, Storage, Status Checks, and Monitoring. To connect to the terminal of this EC2 instance, right-click the instance, select "connect," and choose one of the connection methods. EC2 Instance Connect allows access via a web browser. For this lesson, we focus on connecting using a local SSH client. Ensure you have the local SSH client and the A4L.PEM key file downloaded earlier. In your terminal, navigate to the folder containing the key file and adjust the file's permissions using the chmod 400 command. If on Windows, follow the detailed steps linked in the lesson to correct permissions. After setting permissions, use the provided SSH connection command to connect to the EC2 instance. This command uses the key pair to authenticate your connection. AWS stores the public part of the key pair on the EC2 instance, while you have the private part, allowing secure access. Throughout the course, we will use various methods to connect to EC2 instances, including local SSH clients, Instance Connect, and Session Manager. For Windows EC2 instances, we'll use Remote Desktop, which is beyond the scope of this lesson. To finish up, terminate the instance by selecting it, right-clicking, and choosing "terminate instance." Additionally, delete the security group created for this instance by navigating to the security groups, selecting it, and deleting it once the instance is terminated. Complete these steps and then proceed to the next lesson.

    1. Welcome to this demo lesson where you'll create your first EC2 instance as part of this course. This lesson falls under the AWS Free Tier, so assuming you're using a new account created for this course, there won't be any charges. Ensure you're logged into the general AWS account and have the Northern Virginia region (US-East-1) selected. Once there, navigate to the EC2 console by typing "EC2" in the search box, finding the EC2 service, and clicking to access the console. At the dashboard, the first step is to create an SSH key pair, essential for accessing the EC2 instance. If you don’t already have one, scroll down the left menu, click on Keypairs under Network and Security, and create a key pair. Remember, you'll only have one opportunity to download your part of the key pair, so keep it safe. Name it "A4L" (Animals for Life), choose the key pair type and private key file format based on your operating system (PEM for Mac OS/Linux or modern Windows, PPK for older Windows/using PuTTY), and save it to your default download location.

      Next, return to the dashboard to begin creating your EC2 instance. Click on "Instances" to see all launched instances (you shouldn't have any yet), then click "Launch Instances" to go to the launch instance console. Name the instance "my first EC2 instance" and select the Amazon Machine Image (AMI) for the operating system—use Amazon Linux 2023 for this demo, as it’s free tier eligible. Leave all defaults and choose the instance type, ensuring it’s a free tier eligible type like t2.micro. Select the SSH key pair "A4L" you created earlier.

      Configure network settings by selecting the default VPC and ensure the option to auto-assign a public IP is enabled for internet access. Create a new Security Group named "my_first_instance_sg" to allow SSH from anywhere, understanding you'll need the key pair for access. Configure storage, leaving the default 8GIB GP3 root volume, and expand advanced details if necessary for later use. Launch the EC2 instance and wait for the creation process, which may take a few moments. Refresh to see the instance status change from pending to running, and wait for the status check to complete. This concludes part one of the lesson. Take a break and join part two to continue from here.

    1. With Blue - uncertain - stumbling Buzz - Between the light - and me - And then the Windows failed - and then I could not see to see -

      These lines describe the loss of sight. First, something gets between the narrator and the light. Then their eyes stop working. In the end, they can't see at all. The writer uses "Windows" to describe the eyes. The repetition of the word "see" emphasizes the importance of sight. The dashes make the writing feel jerky. This matches how suddenly losing sight might feel-- abrupt and incohesive.

    1. nisation des outils de configuration du CAP4000 Transition vers une nouvelle application web Auteur Joshua JOURDAM Date de publication 17 juillet 2024 Résumé au dos ou au début du mémoire synthèse carte de visite anglais / français ~15 lignes 3 à 5 mot clés Le renouvellement constant des technologies dans l’industrie est un phénomène de plus en plus marquant de nos jours. Les entreprises qui souhaitent rester compétitives sur le marché doivent s’adapter à ces changements et innover en permanence. Dans ce contexte, le présent projet a pour objectif de développer une nouvelle solution pour répondre aux besoins actuels des industriels. Le projet s’inscrit dans la conception et la réalisation d’une plateforme web permettant de faciliter la gestion des opérations de maintenance et de suivi des équipements SDEL Contrôle Commande. Mots clés Calculateur, Développement WEB, API REST, Authentification Table des matières 1 Environnement et Contexte 1.1 Entreprise, service et position 1.1.1 Vinci 1.1.2 Vinci Energies 1.1.3 Hiérarchie et Fonctionnement de l’Entreprise 1.1.4 SDEL Contrôle Commande 1.1.5 Le service Recherche et Développement 1.2 Contexte du projet 2 Problématique 3 Buts et Objectifs 3.1 Objectifs Stratégiques et Opérationnels 3.1.1 Sélectionner un Cadre de Développement Optimal 3.1.2 Préparer la Transition Technologique 3.1.3 Développer la Nouvelle Application Web 3.1.4 Objectif 4 : Développer les Compétences en Gestion de Projet et Techniques 4 Démarche 4.1 Démarche générale 4.2 Méthodologie, techniques et technologies 4.3 Analyse des risques 4.4 Acteurs 4.5 Lotissements 4.6 Planning prévisionnel 4.7 Planning effectif 4.7.1 Jalons 4.7.2 Livrables 4.8 Budget 5 Résultats 5.1 Evaluation des technologies du marché 5.2 Transition 5.2.1 OpenAPI 5.2.2 Authentification 5.2.3 POC supervision stage 5.3 Développement 5.3.1 Spécification de l’application 5.3.2 Fonctionnement de la solution 5.3.3 Architecture/Workspace (expliquer comment fonctionne un monorepo) 5.3.4 Bibliothèques (zod, react-hook-form, react-query, react-router, orval, react-testing-library, vitest) 5.3.5 Gestion des erreurs et des chargements 5.3.6 Authentification et autorisation 5.3.7 Application 5.3.8 Modules 5.3.9 Tests 5.3.10 Performances 5.3.11 Application 6 Conclusion 6.1 Bilan 7 Annexes 7.1 Références 7.2 Grille des compétences Acronymes HSP PRP REST API SDELCC JWT HTTP HTTPS HMAC WBS CSR SPA Remerciements Ce projet s’intègre dans le cadre de mon apprentissage au sein de l’entreprise SDEL Contrôle Commande. SDEL Contrôle Commande appartient à la filiale Energies du groupe VINCI. Grâce à son expertise technique, SDEL Contrôle commande propose son accompagnement auprès des gestionnaires de réseaux de transport et de distribution d’énergie. J’ai intégré l’entreprise en tant qu’apprenti ingénieur en informatique. J’ai été affecté au service Recherche et Développement, sous la responsabilité de Monsieur Sébastien BARRE, responsable du développement logiciel. Le service Recherche et Développement est en charge de la conception et du développement de calculateurs utilisés principalement dans le domaine de l’énergie notamment dans les postes de transformation du réseau électrique français. Il est également en charge de la maintenance des produits existants. Cette maintenance peut s’effectuer sur de longues périodes, de l’ordre de 20 ans. Le projet s’inscrit dans une dynamique globale qui vise à moderniser les logiciels et les outils que nos équipements embarquent. Cette modernisation à pour objectif de proposer des outils plus ergonomiques et de répondre à des contraintes de cybersécurité plus strictes. Le projet à pour objectif de développer une nouvelle application d’administration et de configuration des automates SDEL. IL se déroule sur la dernière année de ma formation d’ingénieur en apprentissage sur la période de Juin 2023 à août 2024. Ce rapport est découpé en 3 parties : Environnement et contexte : Dans un premier temps, je présenterai le contexte de mon entreprise, puis mon poste et mes rôles au sein de celle-ci. Je présenterai également le contexte, la problématique spécifique ainsi que les buts et objectif du projet. Mise en œuvre et analyse des résultats : Dans cette partie, j’aborderai la méthodologie de travail choisie et les outils utilisés pour assurer son bon déroulement. Je présenterai ensuite les résultats obtenus en évoquant les écarts avec les objectif établis. Bilan et perspectives : Enfin, je ferai un bilan et j’évoquerai les perspectives d’évolution du projet. Je présenterai également les compétences acquises et les apports de ce projet dans mon parcours de formation. Pour finir, je présenterai mes perspectives futures pour ma carrière professionnelle. 1 Environnement et Contexte développer contexte entreprise, Vinci et hiérarchie (energies, omexom), nos marchés (où) ce qu’il faut retenir de chaque chapitre (bullet list) 1.1 Entreprise, service et position 1.1.1 Vinci image/svg+xml Figure 1: Logo VINCI Vinci est une entreprise multinationale française spécialisée dans la construction et les concessions. Fondée en 1899 sous le nom de Société Générale d’Entreprises (SGE), elle est devenue Vinci en 2000. Vinci est l’une des plus grandes entreprises de construction et de concessions dans le monde, avec des activités diversifiées dans le domaine de la construction, des infrastructures, des services énergétiques et de la gestion des infrastructures. Secteurs d’activité : Construction : Vinci Construction est spécialisée dans le bâtiment, les travaux publics, le génie civil, les fondations et les infrastructures de transport. Concessions : Vinci Autoroutes et Vinci Airports gèrent des réseaux autoroutiers et des aéroports dans plusieurs pays. Énergie : Vinci Energies intervient dans les domaines des infrastructures d’énergie, des industries, des technologies de l’information, et de la transition énergétique. Immobilier : Vinci Immobilier est active dans le développement immobilier résidentiel, tertiaire et commercial. #mermaid-1{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#mermaid-1 .error-icon{fill:#a44141;}#mermaid-1 .error-text{fill:#ddd;stroke:#ddd;}#mermaid-1 .edge-thickness-normal{stroke-width:2px;}#mermaid-1 .edge-thickness-thick{stroke-width:3.5px;}#mermaid-1 .edge-pattern-solid{stroke-dasharray:0;}#mermaid-1 .edge-pattern-dashed{stroke-dasharray:3;}#mermaid-1 .edge-pattern-dotted{stroke-dasharray:2;}#mermaid-1 .marker{fill:lightgrey;stroke:lightgrey;}#mermaid-1 .marker.cross{stroke:lightgrey;}#mermaid-1 svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#mermaid-1 .edge{stroke-width:3;}#mermaid-1 .section--1 rect,#mermaid-1 .section--1 path,#mermaid-1 .section--1 circle,#mermaid-1 .section--1 polygon,#mermaid-1 .section--1 path{fill:#1f2020;}#mermaid-1 .section--1 text{fill:lightgrey;}#mermaid-1 .node-icon--1{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge--1{stroke:#1f2020;}#mermaid-1 .edge-depth--1{stroke-width:17;}#mermaid-1 .section--1 line{stroke:#e0dfdf;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-0 rect,#mermaid-1 .section-0 path,#mermaid-1 .section-0 circle,#mermaid-1 .section-0 polygon,#mermaid-1 .section-0 path{fill:#0b0000;}#mermaid-1 .section-0 text{fill:lightgrey;}#mermaid-1 .node-icon-0{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-0{stroke:#0b0000;}#mermaid-1 .edge-depth-0{stroke-width:14;}#mermaid-1 .section-0 line{stroke:#f4ffff;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-1 rect,#mermaid-1 .section-1 path,#mermaid-1 .section-1 circle,#mermaid-1 .section-1 polygon,#mermaid-1 .section-1 path{fill:#4d1037;}#mermaid-1 .section-1 text{fill:lightgrey;}#mermaid-1 .node-icon-1{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-1{stroke:#4d1037;}#mermaid-1 .edge-depth-1{stroke-width:11;}#mermaid-1 .section-1 line{stroke:#b2efc8;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-2 rect,#mermaid-1 .section-2 path,#mermaid-1 .section-2 circle,#mermaid-1 .section-2 polygon,#mermaid-1 .section-2 path{fill:#3f5258;}#mermaid-1 .section-2 text{fill:lightgrey;}#mermaid-1 .node-icon-2{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-2{stroke:#3f5258;}#mermaid-1 .edge-depth-2{stroke-width:8;}#mermaid-1 .section-2 line{stroke:#c0ada7;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-3 rect,#mermaid-1 .section-3 path,#mermaid-1 .section-3 circle,#mermaid-1 .section-3 polygon,#mermaid-1 .section-3 path{fill:#4f2f1b;}#mermaid-1 .section-3 text{fill:lightgrey;}#mermaid-1 .node-icon-3{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-3{stroke:#4f2f1b;}#mermaid-1 .edge-depth-3{stroke-width:5;}#mermaid-1 .section-3 line{stroke:#b0d0e4;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-4 rect,#mermaid-1 .section-4 path,#mermaid-1 .section-4 circle,#mermaid-1 .section-4 polygon,#mermaid-1 .section-4 path{fill:#6e0a0a;}#mermaid-1 .section-4 text{fill:lightgrey;}#mermaid-1 .node-icon-4{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-4{stroke:#6e0a0a;}#mermaid-1 .edge-depth-4{stroke-width:2;}#mermaid-1 .section-4 line{stroke:#91f5f5;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-5 rect,#mermaid-1 .section-5 path,#mermaid-1 .section-5 circle,#mermaid-1 .section-5 polygon,#mermaid-1 .section-5 path{fill:#3b0048;}#mermaid-1 .section-5 text{fill:lightgrey;}#mermaid-1 .node-icon-5{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-5{stroke:#3b0048;}#mermaid-1 .edge-depth-5{stroke-width:-1;}#mermaid-1 .section-5 line{stroke:#c4ffb7;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-6 rect,#mermaid-1 .section-6 path,#mermaid-1 .section-6 circle,#mermaid-1 .section-6 polygon,#mermaid-1 .section-6 path{fill:#995a01;}#mermaid-1 .section-6 text{fill:lightgrey;}#mermaid-1 .node-icon-6{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-6{stroke:#995a01;}#mermaid-1 .edge-depth-6{stroke-width:-4;}#mermaid-1 .section-6 line{stroke:#66a5fe;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-7 rect,#mermaid-1 .section-7 path,#mermaid-1 .section-7 circle,#mermaid-1 .section-7 polygon,#mermaid-1 .section-7 path{fill:#154706;}#mermaid-1 .section-7 text{fill:lightgrey;}#mermaid-1 .node-icon-7{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-7{stroke:#154706;}#mermaid-1 .edge-depth-7{stroke-width:-7;}#mermaid-1 .section-7 line{stroke:#eab8f9;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-8 rect,#mermaid-1 .section-8 path,#mermaid-1 .section-8 circle,#mermaid-1 .section-8 polygon,#mermaid-1 .section-8 path{fill:#161722;}#mermaid-1 .section-8 text{fill:lightgrey;}#mermaid-1 .node-icon-8{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-8{stroke:#161722;}#mermaid-1 .edge-depth-8{stroke-width:-10;}#mermaid-1 .section-8 line{stroke:#e9e8dd;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-9 rect,#mermaid-1 .section-9 path,#mermaid-1 .section-9 circle,#mermaid-1 .section-9 polygon,#mermaid-1 .section-9 path{fill:#00296f;}#mermaid-1 .section-9 text{fill:lightgrey;}#mermaid-1 .node-icon-9{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-9{stroke:#00296f;}#mermaid-1 .edge-depth-9{stroke-width:-13;}#mermaid-1 .section-9 line{stroke:#ffd690;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-10 rect,#mermaid-1 .section-10 path,#mermaid-1 .section-10 circle,#mermaid-1 .section-10 polygon,#mermaid-1 .section-10 path{fill:#01629c;}#mermaid-1 .section-10 text{fill:lightgrey;}#mermaid-1 .node-icon-10{font-size:40px;color:lightgrey;}#mermaid-1 .section-edge-10{stroke:#01629c;}#mermaid-1 .edge-depth-10{stroke-width:-16;}#mermaid-1 .section-10 line{stroke:#fe9d63;stroke-width:3;}#mermaid-1 .disabled,#mermaid-1 .disabled circle,#mermaid-1 .disabled text{fill:lightgray;}#mermaid-1 .disabled text{fill:#efefef;}#mermaid-1 .section-root rect,#mermaid-1 .section-root path,#mermaid-1 .section-root circle,#mermaid-1 .section-root polygon{fill:hsl(180, 1.5873015873%, 48.3529411765%);}#mermaid-1 .section-root text{fill:#2c2c2c;}#mermaid-1 .icon-container{height:100%;display:flex;justify-content:center;align-items:center;}#mermaid-1 .edge{fill:none;}#mermaid-1 .mindmap-node-label{dy:1em;alignment-baseline:middle;text-anchor:middle;dominant-baseline:middle;text-align:center;}#mermaid-1 :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;}VINCIVinci EnergiesVinci ConstructionVinci AutoroutesVinci AirportsVinci Immobilier 1.1.2 Vinci Energies Vinci Energies, une filiale de Vinci, est spécialisée dans les services énergétiques et les technologies de l’information. Elle propose des solutions dans les domaines de l’énergie, des technologies de l’information, et des télécommunications. Ses principales activités incluent : Infrastructure Énergie : Gestion des réseaux électriques et des infrastructures de distribution d’énergie. Industrie : Optimisation des processus industriels et amélioration de l’efficacité énergétique. TIC (Technologies de l’Information et de la Communication) : Solutions pour les systèmes d’information et les télécommunications. Transition Énergétique et Environnementale : Développement de solutions pour une énergie plus durable et respectueuse de l’environnement. Vinci Energies, une division du groupe VINCI, regroupe plusieurs marques spécialisées dans divers domaines des services énergétiques et des technologies de l’information. Les cinq principales marques de Vinci Energies sont : Actemium : Spécialisée dans les solutions et les services pour les processus industriels, couvrant l’ensemble du cycle de vie des installations industrielles. Axians : Focalisée sur les technologies de l’information et de la communication (TIC), offrant des solutions pour les infrastructures IT, la cybersécurité, le cloud, les réseaux et la collaboration. Cegelec : Fournit des services et des solutions en ingénierie électrique et maintenance pour les infrastructures et les bâtiments. Omexom : Se concentre sur les infrastructures énergétiques, notamment les réseaux de transport et de distribution d’électricité, les énergies renouvelables et les systèmes de stockage d’énergie. VINCI Facilities : Offre des services de gestion et de maintenance des bâtiments, incluant des solutions de facility management intégrées pour optimiser les performances des installations. 1.1.3 Hiérarchie et Fonctionnement de l’Entreprise 1.1.3.1 Organisation Vinci est organisée en plusieurs divisions opérationnelles, chacune ayant une structure hiérarchique propre. La hiérarchie de Vinci et de ses filiales, comme Vinci Energies, est généralement structurée de la manière suivante : Conseil d’Administration : Organe suprême de la société, responsable de la stratégie globale et de la surveillance de la direction exécutive. Direction Générale : Composée du Président-Directeur Général (PDG) et d’autres membres de la direction exécutive, responsables de la mise en œuvre de la stratégie et de la gestion quotidienne de l’entreprise. Directeurs de Divisions/Branches : Chacun responsable d’une branche spécifique (par exemple, Vinci Construction, Vinci Energies). Directeurs de Filières/Entités : Supervisent des sous-divisions ou entités spécifiques au sein de chaque branche, telles que des régions géographiques ou des domaines spécialisés. Chefs de Projet et Managers Opérationnels : Responsables de la gestion quotidienne des projets et des équipes sur le terrain. Équipes Opérationnelles : Constituent les employés qui travaillent directement sur les projets, incluant ingénieurs, techniciens, ouvriers, et autres professionnels spécialisés. 1.1.3.2 Fonctionnement Décentralisation : Vinci privilégie une approche décentralisée, permettant à ses différentes divisions et filiales de bénéficier d’une grande autonomie. Cela favorise la réactivité et l’adaptabilité aux marchés locaux. Innovation : L’entreprise met l’accent sur l’innovation technologique et l’efficacité énergétique, soutenant les projets de recherche et développement pour anticiper les besoins futurs. Responsabilité Sociétale et Environnementale : Vinci est engagée dans une démarche de développement durable, visant à réduire son empreinte écologique et à améliorer ses performances environnementales. En résumé, Vinci est une entreprise diversifiée avec une structure hiérarchique bien définie, soutenant une approche décentralisée et innovante pour répondre aux besoins de ses marchés dans le domaine de la construction, des concessions, et des services énergétiques. 1.1.4 SDEL Contrôle Commande Figure 2: Entrée principale de SDEL Contrôle Commande SDEL Contrôle Commande appartient à la filiale Energies du groupe VINCI. Elle est composée de plusieurs marques, comme Omexom, Actemium et Axians. Chaque marque est spécialisée dans un domaine d’activité précis. Ainsi VINCI Energies offre une gamme complète de services et de solutions dans le domaine de l’énergie. SDEL Contrôle Commande est une entreprise mono-site basée à Saint-Aignan-Grandlieu. Au sein du réseau Omexom, SDEL Contrôle Commande intervient dans la gestion de projets clé en main, la conception, l’ingénierie, l’intégration, l’installation, la configuration, les essais, la mise en service et la maintenance de systèmes de contrôle commande de postes électriques et d’automatismes. Figure 3: Tranches SDEL Composée de plus de 300 salariés, elle bénéficie de plus de 50 années d’expérience dans le contrôle commande. Ainsi, grâce à son expertise technique, SDEL Contrôle commande propose son accompagnement auprès des gestionnaires de réseaux de transport et de distribution d’énergie. Nos clients majeurs sont RTE et Enedis. Nous proposons une offre sur mesure de produits à destination des postes du réseau électrique français. Nous fournissons également des systèmes de contrôle commande à Thales pour le pilotage de batteries marines et à la RATP pour la supervision du système de ventilation du métro parisien. RTE Enedis Thales RATP EDF Nos principaux concurrents sont Actia Telecom, SCLE, Eiffage Energie. Ils proposent une gamme de produits similaire sur le marché de la distribution et du transport de l’énergie. Quelques entreprises du groupe VINCI sont également en compétition directe avec SDEL Contrôle Commande sur certains appels d’offre. #fig-market-mermaid{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#fig-market-mermaid .error-icon{fill:#a44141;}#fig-market-mermaid .error-text{fill:#ddd;stroke:#ddd;}#fig-market-mermaid .edge-thickness-normal{stroke-width:2px;}#fig-market-mermaid .edge-thickness-thick{stroke-width:3.5px;}#fig-market-mermaid .edge-pattern-solid{stroke-dasharray:0;}#fig-market-mermaid .edge-pattern-dashed{stroke-dasharray:3;}#fig-market-mermaid .edge-pattern-dotted{stroke-dasharray:2;}#fig-market-mermaid .marker{fill:lightgrey;stroke:lightgrey;}#fig-market-mermaid .marker.cross{stroke:lightgrey;}#fig-market-mermaid svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#fig-market-mermaid .pieCircle{stroke:black;stroke-width:2px;opacity:0.7;}#fig-market-mermaid .pieOuterCircle{stroke:black;stroke-width:2px;fill:none;}#fig-market-mermaid .pieTitleText{text-anchor:middle;font-size:25px;fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);font-family:"trebuchet ms",verdana,arial,sans-serif;}#fig-market-mermaid .slice{font-family:"trebuchet ms",verdana,arial,sans-serif;fill:#ccc;font-size:17px;}#fig-market-mermaid .legend text{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:17px;}#fig-market-mermaid :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;}85%10%4%1%EnergieDéfenseTransportAutres industries Figure 4: Répartition des marchés de SDEL Contrôle Commande #fig-organisation-mermaid{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#fig-organisation-mermaid .error-icon{fill:#a44141;}#fig-organisation-mermaid .error-text{fill:#ddd;stroke:#ddd;}#fig-organisation-mermaid .edge-thickness-normal{stroke-width:2px;}#fig-organisation-mermaid .edge-thickness-thick{stroke-width:3.5px;}#fig-organisation-mermaid .edge-pattern-solid{stroke-dasharray:0;}#fig-organisation-mermaid .edge-pattern-dashed{stroke-dasharray:3;}#fig-organisation-mermaid .edge-pattern-dotted{stroke-dasharray:2;}#fig-organisation-mermaid .marker{fill:lightgrey;stroke:lightgrey;}#fig-organisation-mermaid .marker.cross{stroke:lightgrey;}#fig-organisation-mermaid svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#fig-organisation-mermaid .edge{stroke-width:3;}#fig-organisation-mermaid .section--1 rect,#fig-organisation-mermaid .section--1 path,#fig-organisation-mermaid .section--1 circle,#fig-organisation-mermaid .section--1 polygon,#fig-organisation-mermaid .section--1 path{fill:#1f2020;}#fig-organisation-mermaid .section--1 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon--1{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge--1{stroke:#1f2020;}#fig-organisation-mermaid .edge-depth--1{stroke-width:17;}#fig-organisation-mermaid .section--1 line{stroke:#e0dfdf;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-0 rect,#fig-organisation-mermaid .section-0 path,#fig-organisation-mermaid .section-0 circle,#fig-organisation-mermaid .section-0 polygon,#fig-organisation-mermaid .section-0 path{fill:#0b0000;}#fig-organisation-mermaid .section-0 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-0{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-0{stroke:#0b0000;}#fig-organisation-mermaid .edge-depth-0{stroke-width:14;}#fig-organisation-mermaid .section-0 line{stroke:#f4ffff;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-1 rect,#fig-organisation-mermaid .section-1 path,#fig-organisation-mermaid .section-1 circle,#fig-organisation-mermaid .section-1 polygon,#fig-organisation-mermaid .section-1 path{fill:#4d1037;}#fig-organisation-mermaid .section-1 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-1{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-1{stroke:#4d1037;}#fig-organisation-mermaid .edge-depth-1{stroke-width:11;}#fig-organisation-mermaid .section-1 line{stroke:#b2efc8;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-2 rect,#fig-organisation-mermaid .section-2 path,#fig-organisation-mermaid .section-2 circle,#fig-organisation-mermaid .section-2 polygon,#fig-organisation-mermaid .section-2 path{fill:#3f5258;}#fig-organisation-mermaid .section-2 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-2{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-2{stroke:#3f5258;}#fig-organisation-mermaid .edge-depth-2{stroke-width:8;}#fig-organisation-mermaid .section-2 line{stroke:#c0ada7;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-3 rect,#fig-organisation-mermaid .section-3 path,#fig-organisation-mermaid .section-3 circle,#fig-organisation-mermaid .section-3 polygon,#fig-organisation-mermaid .section-3 path{fill:#4f2f1b;}#fig-organisation-mermaid .section-3 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-3{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-3{stroke:#4f2f1b;}#fig-organisation-mermaid .edge-depth-3{stroke-width:5;}#fig-organisation-mermaid .section-3 line{stroke:#b0d0e4;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-4 rect,#fig-organisation-mermaid .section-4 path,#fig-organisation-mermaid .section-4 circle,#fig-organisation-mermaid .section-4 polygon,#fig-organisation-mermaid .section-4 path{fill:#6e0a0a;}#fig-organisation-mermaid .section-4 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-4{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-4{stroke:#6e0a0a;}#fig-organisation-mermaid .edge-depth-4{stroke-width:2;}#fig-organisation-mermaid .section-4 line{stroke:#91f5f5;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-5 rect,#fig-organisation-mermaid .section-5 path,#fig-organisation-mermaid .section-5 circle,#fig-organisation-mermaid .section-5 polygon,#fig-organisation-mermaid .section-5 path{fill:#3b0048;}#fig-organisation-mermaid .section-5 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-5{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-5{stroke:#3b0048;}#fig-organisation-mermaid .edge-depth-5{stroke-width:-1;}#fig-organisation-mermaid .section-5 line{stroke:#c4ffb7;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-6 rect,#fig-organisation-mermaid .section-6 path,#fig-organisation-mermaid .section-6 circle,#fig-organisation-mermaid .section-6 polygon,#fig-organisation-mermaid .section-6 path{fill:#995a01;}#fig-organisation-mermaid .section-6 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-6{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-6{stroke:#995a01;}#fig-organisation-mermaid .edge-depth-6{stroke-width:-4;}#fig-organisation-mermaid .section-6 line{stroke:#66a5fe;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-7 rect,#fig-organisation-mermaid .section-7 path,#fig-organisation-mermaid .section-7 circle,#fig-organisation-mermaid .section-7 polygon,#fig-organisation-mermaid .section-7 path{fill:#154706;}#fig-organisation-mermaid .section-7 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-7{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-7{stroke:#154706;}#fig-organisation-mermaid .edge-depth-7{stroke-width:-7;}#fig-organisation-mermaid .section-7 line{stroke:#eab8f9;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-8 rect,#fig-organisation-mermaid .section-8 path,#fig-organisation-mermaid .section-8 circle,#fig-organisation-mermaid .section-8 polygon,#fig-organisation-mermaid .section-8 path{fill:#161722;}#fig-organisation-mermaid .section-8 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-8{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-8{stroke:#161722;}#fig-organisation-mermaid .edge-depth-8{stroke-width:-10;}#fig-organisation-mermaid .section-8 line{stroke:#e9e8dd;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-9 rect,#fig-organisation-mermaid .section-9 path,#fig-organisation-mermaid .section-9 circle,#fig-organisation-mermaid .section-9 polygon,#fig-organisation-mermaid .section-9 path{fill:#00296f;}#fig-organisation-mermaid .section-9 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-9{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-9{stroke:#00296f;}#fig-organisation-mermaid .edge-depth-9{stroke-width:-13;}#fig-organisation-mermaid .section-9 line{stroke:#ffd690;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-10 rect,#fig-organisation-mermaid .section-10 path,#fig-organisation-mermaid .section-10 circle,#fig-organisation-mermaid .section-10 polygon,#fig-organisation-mermaid .section-10 path{fill:#01629c;}#fig-organisation-mermaid .section-10 text{fill:lightgrey;}#fig-organisation-mermaid .node-icon-10{font-size:40px;color:lightgrey;}#fig-organisation-mermaid .section-edge-10{stroke:#01629c;}#fig-organisation-mermaid .edge-depth-10{stroke-width:-16;}#fig-organisation-mermaid .section-10 line{stroke:#fe9d63;stroke-width:3;}#fig-organisation-mermaid .disabled,#fig-organisation-mermaid .disabled circle,#fig-organisation-mermaid .disabled text{fill:lightgray;}#fig-organisation-mermaid .disabled text{fill:#efefef;}#fig-organisation-mermaid .section-root rect,#fig-organisation-mermaid .section-root path,#fig-organisation-mermaid .section-root circle,#fig-organisation-mermaid .section-root polygon{fill:hsl(180, 1.5873015873%, 48.3529411765%);}#fig-organisation-mermaid .section-root text{fill:#2c2c2c;}#fig-organisation-mermaid .icon-container{height:100%;display:flex;justify-content:center;align-items:center;}#fig-organisation-mermaid .edge{fill:none;}#fig-organisation-mermaid .mindmap-node-label{dy:1em;alignment-baseline:middle;text-anchor:middle;dominant-baseline:middle;text-align:center;}#fig-organisation-mermaid :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;}SDEL Contrôle CommandeRecherche etDéveloppementIngénierieInnovationGestion affairesStratégie IndustrielleBureau d 'étudesIntégrationLogistiqueChaîne achatsApprovisionnementEssaisMise en serviceTravauxInterventions Figure 5: Les services de SDEL Contrôle Commande 1.1.5 Le service Recherche et Développement SDEL Contrôle Commande propose des solutions techniques adaptées aux besoins de ses clients grâce à ses services Recherche et Développement, Ingénierie et Innovation. En plus de notre connaissance approfondie des différents constructeurs, notre expertise dans le domaine d’application nous permet de répondre aux besoins en ingénierie, dimensionnement, configuration et déploiement d’équipements de protection et de contrôle pour les réseaux électriques de transport et de distribution d’énergie. Le service Recherche et Développement est composé de 16 personnes. Il est spécialisé dans la conception de calculateurs et d’équipements utilisés dans les réseaux de distribution électriques. #fig-software-team-mermaid{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#fig-software-team-mermaid .error-icon{fill:#a44141;}#fig-software-team-mermaid .error-text{fill:#ddd;stroke:#ddd;}#fig-software-team-mermaid .edge-thickness-normal{stroke-width:2px;}#fig-software-team-mermaid .edge-thickness-thick{stroke-width:3.5px;}#fig-software-team-mermaid .edge-pattern-solid{stroke-dasharray:0;}#fig-software-team-mermaid .edge-pattern-dashed{stroke-dasharray:3;}#fig-software-team-mermaid .edge-pattern-dotted{stroke-dasharray:2;}#fig-software-team-mermaid .marker{fill:lightgrey;stroke:lightgrey;}#fig-software-team-mermaid .marker.cross{stroke:lightgrey;}#fig-software-team-mermaid svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#fig-software-team-mermaid .edge{stroke-width:3;}#fig-software-team-mermaid .section--1 rect,#fig-software-team-mermaid .section--1 path,#fig-software-team-mermaid .section--1 circle,#fig-software-team-mermaid .section--1 polygon,#fig-software-team-mermaid .section--1 path{fill:#1f2020;}#fig-software-team-mermaid .section--1 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon--1{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge--1{stroke:#1f2020;}#fig-software-team-mermaid .edge-depth--1{stroke-width:17;}#fig-software-team-mermaid .section--1 line{stroke:#e0dfdf;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-0 rect,#fig-software-team-mermaid .section-0 path,#fig-software-team-mermaid .section-0 circle,#fig-software-team-mermaid .section-0 polygon,#fig-software-team-mermaid .section-0 path{fill:#0b0000;}#fig-software-team-mermaid .section-0 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-0{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-0{stroke:#0b0000;}#fig-software-team-mermaid .edge-depth-0{stroke-width:14;}#fig-software-team-mermaid .section-0 line{stroke:#f4ffff;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-1 rect,#fig-software-team-mermaid .section-1 path,#fig-software-team-mermaid .section-1 circle,#fig-software-team-mermaid .section-1 polygon,#fig-software-team-mermaid .section-1 path{fill:#4d1037;}#fig-software-team-mermaid .section-1 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-1{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-1{stroke:#4d1037;}#fig-software-team-mermaid .edge-depth-1{stroke-width:11;}#fig-software-team-mermaid .section-1 line{stroke:#b2efc8;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-2 rect,#fig-software-team-mermaid .section-2 path,#fig-software-team-mermaid .section-2 circle,#fig-software-team-mermaid .section-2 polygon,#fig-software-team-mermaid .section-2 path{fill:#3f5258;}#fig-software-team-mermaid .section-2 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-2{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-2{stroke:#3f5258;}#fig-software-team-mermaid .edge-depth-2{stroke-width:8;}#fig-software-team-mermaid .section-2 line{stroke:#c0ada7;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-3 rect,#fig-software-team-mermaid .section-3 path,#fig-software-team-mermaid .section-3 circle,#fig-software-team-mermaid .section-3 polygon,#fig-software-team-mermaid .section-3 path{fill:#4f2f1b;}#fig-software-team-mermaid .section-3 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-3{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-3{stroke:#4f2f1b;}#fig-software-team-mermaid .edge-depth-3{stroke-width:5;}#fig-software-team-mermaid .section-3 line{stroke:#b0d0e4;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-4 rect,#fig-software-team-mermaid .section-4 path,#fig-software-team-mermaid .section-4 circle,#fig-software-team-mermaid .section-4 polygon,#fig-software-team-mermaid .section-4 path{fill:#6e0a0a;}#fig-software-team-mermaid .section-4 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-4{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-4{stroke:#6e0a0a;}#fig-software-team-mermaid .edge-depth-4{stroke-width:2;}#fig-software-team-mermaid .section-4 line{stroke:#91f5f5;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-5 rect,#fig-software-team-mermaid .section-5 path,#fig-software-team-mermaid .section-5 circle,#fig-software-team-mermaid .section-5 polygon,#fig-software-team-mermaid .section-5 path{fill:#3b0048;}#fig-software-team-mermaid .section-5 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-5{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-5{stroke:#3b0048;}#fig-software-team-mermaid .edge-depth-5{stroke-width:-1;}#fig-software-team-mermaid .section-5 line{stroke:#c4ffb7;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-6 rect,#fig-software-team-mermaid .section-6 path,#fig-software-team-mermaid .section-6 circle,#fig-software-team-mermaid .section-6 polygon,#fig-software-team-mermaid .section-6 path{fill:#995a01;}#fig-software-team-mermaid .section-6 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-6{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-6{stroke:#995a01;}#fig-software-team-mermaid .edge-depth-6{stroke-width:-4;}#fig-software-team-mermaid .section-6 line{stroke:#66a5fe;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-7 rect,#fig-software-team-mermaid .section-7 path,#fig-software-team-mermaid .section-7 circle,#fig-software-team-mermaid .section-7 polygon,#fig-software-team-mermaid .section-7 path{fill:#154706;}#fig-software-team-mermaid .section-7 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-7{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-7{stroke:#154706;}#fig-software-team-mermaid .edge-depth-7{stroke-width:-7;}#fig-software-team-mermaid .section-7 line{stroke:#eab8f9;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-8 rect,#fig-software-team-mermaid .section-8 path,#fig-software-team-mermaid .section-8 circle,#fig-software-team-mermaid .section-8 polygon,#fig-software-team-mermaid .section-8 path{fill:#161722;}#fig-software-team-mermaid .section-8 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-8{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-8{stroke:#161722;}#fig-software-team-mermaid .edge-depth-8{stroke-width:-10;}#fig-software-team-mermaid .section-8 line{stroke:#e9e8dd;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-9 rect,#fig-software-team-mermaid .section-9 path,#fig-software-team-mermaid .section-9 circle,#fig-software-team-mermaid .section-9 polygon,#fig-software-team-mermaid .section-9 path{fill:#00296f;}#fig-software-team-mermaid .section-9 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-9{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-9{stroke:#00296f;}#fig-software-team-mermaid .edge-depth-9{stroke-width:-13;}#fig-software-team-mermaid .section-9 line{stroke:#ffd690;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-10 rect,#fig-software-team-mermaid .section-10 path,#fig-software-team-mermaid .section-10 circle,#fig-software-team-mermaid .section-10 polygon,#fig-software-team-mermaid .section-10 path{fill:#01629c;}#fig-software-team-mermaid .section-10 text{fill:lightgrey;}#fig-software-team-mermaid .node-icon-10{font-size:40px;color:lightgrey;}#fig-software-team-mermaid .section-edge-10{stroke:#01629c;}#fig-software-team-mermaid .edge-depth-10{stroke-width:-16;}#fig-software-team-mermaid .section-10 line{stroke:#fe9d63;stroke-width:3;}#fig-software-team-mermaid .disabled,#fig-software-team-mermaid .disabled circle,#fig-software-team-mermaid .disabled text{fill:lightgray;}#fig-software-team-mermaid .disabled text{fill:#efefef;}#fig-software-team-mermaid .section-root rect,#fig-software-team-mermaid .section-root path,#fig-software-team-mermaid .section-root circle,#fig-software-team-mermaid .section-root polygon{fill:hsl(180, 1.5873015873%, 48.3529411765%);}#fig-software-team-mermaid .section-root text{fill:#2c2c2c;}#fig-software-team-mermaid .icon-container{height:100%;display:flex;justify-content:center;align-items:center;}#fig-software-team-mermaid .edge{fill:none;}#fig-software-team-mermaid .mindmap-node-label{dy:1em;alignment-baseline:middle;text-anchor:middle;dominant-baseline:middle;text-align:center;}#fig-software-team-mermaid :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;}BARRE SebastienChef de groupeIngénieurBAUDOUIN Jean LouisIHMTechnicienJANNIERE SylvainCAP4000 /IHMIngénieurCECILLON LucasCAP4000 /CybersécuritéApprentiDUPONT DavidIHMIngénieurHELARD FlorentCAP4000IngénieurMINIER BertrandYocto /CAP4000IngénieurJOURDAM JoshuaYocto /IHMApprenti Figure 6: Organigramme équipe logiciel Monsieur Sébastien BARRE, est responsable du développement logiciel dans le service Recherche et Développement. L’équipe logiciel travaille les applicatifs et le système d’exploitation des équipements que nous développons. Ces matériels sont destinés à être commercialisés via l’intégration dans les produits et services fournis par SDEL contrôle commande. Par exemple lors de la vente d’armoires de contrôle commande ou lors de l’installation d’un poste électrique. Figure 7: DigiBOX, calculateur SDELCC dédié aux applications des postes électriques Applications : Système de supervision de poste Passerelle de téléconduite Synoptique de poste local Consignateur d’état Concrètement -> embarqué, linux, yocto En tant qu’apprenti j’ai pour rôle de contribuer au développement de nos logiciels. Je suis aussi mobilisé pour tester et étudier l’intégration de nouvelles technologies sur les produits. Mon travail comprends l’étude, l’analyse de données, la création de prototypes, la mise en place de tests et l’assistance aux techniciens et aux ingénieurs dans leur travail. Je suis également amené à rédiger de la documentation et la communiquer les résultats de mes travaux. J’ai l’occasion de travailler sur des projets concrets et de développer mes compétences dans le domaine de l’informatique embarquée principalement. Avec ma montée en compétences au fur et à mesure de ma formation, je suis maintenant amené à travailler sur des projets plus complexes et à diriger le présent projet. 1.2 Contexte du projet Le CAP4000 est une base logicielle modulaire qui est utilisée par tous les équipements développés par le service Recherche et Développement. C’est une application qui fonctionne en permanence. Elle gère les principales fonctionnalités de nos produits en s’interfaçant avec différents composants tels que : Système d’exploitation (réseau, alimentation, processus …) Périphériques (affichage, cartes d’extensions, clé usb, …) Moteur d’automatisme Panneau frontal (leds, boutons, …) Nos équipements sont hautement configurables. Pour simplifier l’utilisation par nos client, l’administration des équipements doit pouvoir être effectuée à distance. Ces tâches sont réalisées grâce à aux outils de configuration historiques qui sont inclus dans une application de bureau windows. Figure 8: Utilitaire de configuration Depuis 2020 l’équipe logiciel travaille sur le remplacement de cette application. L’objectif est de moderniser nos outils pour simplifier et intégrer des contraintes de cybersécurité plus strictes dans les processus de gestion et de maintenance des équipements. Pour moderniser les principes d’accès aux ressources d’un calculateur, une API de type REST à été implémentée dans le logiciel CAP4000. L’objectif est de mettre à disposition un système de dialogue fiable et sécurisé avec un équipement dans le but de créer des outils basés sur les technologies Web pour les produits SDELCC. API (Interface de Programmation d’Application) Les API pour Application Programming Interface permettent à 2 ordinateur de communiquer entre eux. Imaginez cela comme l’utilisation d’un site web, mais au lieu de cliquer sur des boutons, vous écrivez du code pour demander explicitement des données à un serveur. Une API dite RESTful, suit un ensemble de règles et contraintes imposées par l’architecture REST (REpresentational State Transfer). Une API REST fonctionne sur le protocole HTTP (Hypertext Transfer Protocol). Elle mets à disposition des ressources (données) accessible par des URL uniques. Une requête permet d’accéder aux ressources. Chaque requête (GET, POST, PATCH, DELETE) respecte le principe CRUD (Create, Read, Update, Delete) et suit un format spécifique : méthode, URL, en-têtes (métadonnées) et corps (données). Un client effectue une requête sur le serveur via une URL, le serveur exécute du code (généralement accède à une base de données), formate les données dans une réponse avec un code d’état (indiquant le succès, une erreur client ou serveur). Les API REST sont stateless, ce qui signifie que chaque interaction est indépendante des précédentes, rendant les applications prévisibles et fiables. Les API de type REST sont devenues le standard de facto pour le développement d’API web depuis le début des années 2000. #fig-rest-api-flow-mermaid{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#fig-rest-api-flow-mermaid .error-icon{fill:#a44141;}#fig-rest-api-flow-mermaid .error-text{fill:#ddd;stroke:#ddd;}#fig-rest-api-flow-mermaid .edge-thickness-normal{stroke-width:2px;}#fig-rest-api-flow-mermaid .edge-thickness-thick{stroke-width:3.5px;}#fig-rest-api-flow-mermaid .edge-pattern-solid{stroke-dasharray:0;}#fig-rest-api-flow-mermaid .edge-pattern-dashed{stroke-dasharray:3;}#fig-rest-api-flow-mermaid .edge-pattern-dotted{stroke-dasharray:2;}#fig-rest-api-flow-mermaid .marker{fill:lightgrey;stroke:lightgrey;}#fig-rest-api-flow-mermaid .marker.cross{stroke:lightgrey;}#fig-rest-api-flow-mermaid svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#fig-rest-api-flow-mermaid .label{font-family:"trebuchet ms",verdana,arial,sans-serif;color:#ccc;}#fig-rest-api-flow-mermaid .cluster-label text{fill:#F9FFFE;}#fig-rest-api-flow-mermaid .cluster-label span,#fig-rest-api-flow-mermaid p{color:#F9FFFE;}#fig-rest-api-flow-mermaid .label text,#fig-rest-api-flow-mermaid span,#fig-rest-api-flow-mermaid p{fill:#ccc;color:#ccc;}#fig-rest-api-flow-mermaid .node rect,#fig-rest-api-flow-mermaid .node circle,#fig-rest-api-flow-mermaid .node ellipse,#fig-rest-api-flow-mermaid .node polygon,#fig-rest-api-flow-mermaid .node path{fill:#1f2020;stroke:#81B1DB;stroke-width:1px;}#fig-rest-api-flow-mermaid .flowchart-label text{text-anchor:middle;}#fig-rest-api-flow-mermaid .node .label{text-align:center;}#fig-rest-api-flow-mermaid .node.clickable{cursor:pointer;}#fig-rest-api-flow-mermaid .arrowheadPath{fill:lightgrey;}#fig-rest-api-flow-mermaid .edgePath .path{stroke:lightgrey;stroke-width:2.0px;}#fig-rest-api-flow-mermaid .flowchart-link{stroke:lightgrey;fill:none;}#fig-rest-api-flow-mermaid .edgeLabel{background-color:hsl(0, 0%, 34.4117647059%);text-align:center;}#fig-rest-api-flow-mermaid .edgeLabel rect{opacity:0.5;background-color:hsl(0, 0%, 34.4117647059%);fill:hsl(0, 0%, 34.4117647059%);}#fig-rest-api-flow-mermaid .cluster rect{fill:hsl(180, 1.5873015873%, 28.3529411765%);stroke:rgba(255, 255, 255, 0.25);stroke-width:1px;}#fig-rest-api-flow-mermaid .cluster text{fill:#F9FFFE;}#fig-rest-api-flow-mermaid .cluster span,#fig-rest-api-flow-mermaid p{color:#F9FFFE;}#fig-rest-api-flow-mermaid div.mermaidTooltip{position:absolute;text-align:center;max-width:200px;padding:2px;font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:12px;background:hsl(20, 1.5873015873%, 12.3529411765%);border:1px solid rgba(255, 255, 255, 0.25);border-radius:2px;pointer-events:none;z-index:100;}#fig-rest-api-flow-mermaid .flowchartTitleText{text-anchor:middle;font-size:18px;fill:#ccc;}#fig-rest-api-flow-mermaid :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;}OdrinateurÉquipementNavigateurCAP4000VisiteRequêteGET/POST/PUT/DELETERéponseJSONLit et écritExecuteApplication webAPIrestModule XBase de donnéesProcéduresUtilisateur Figure 9: Principe d’utilisation de l’API REST du CAP4000 Cette API à pour principal objectif de permettre la configuration et la gestion des équipements à distance. Elle permet de récupérer des informations sur l’état de l’équipement et de ses composants. Elle permet également la modification la configuration de l’équipement et le déclenchement de procédures comme le redémarrage. Par exemple, il est possible de récupérer la version du logiciel installé sur l’équipement en exécutant la requête HTTP GET suivante : http://<ip-equipement>/api/v1/identification. identification { "CAP4000": { "IndiceValidation": 0, "Nom": "DigiBOX", "VersionCorrectif": 1, "VersionMajeure": 6, "VersionMineure": 0, "VersionProtocole": 2 }, "OS": { "Nom": "digibox-debug-os", "VersionCorrectif": 5, "VersionMajeure": 0, "VersionMineure": 4 }, "Produit": { "Nom": "digibox-debug", "VersionCorrectif": 1, "VersionMajeure": 2, "VersionMineure": 0 } } Une expérimentation d’application à été développée. Cette application est disponible pour deux plateformes : Windows et Web. Elle dialogue avec l’API REST du CAP4000. L’application a été développée avec le framework Qt. C’est un framework que nous utilisons déjà sur d’autres projets. Il a donc été choisi afin de capitaliser sur l’expertise du service. L’application est découpée en modules qui représentent une fonctionnalité dans l’interface (debug, configuration réseau, import de configuration, …). Qt (prononcé “cute”) Qt est un framework de développement d’applications multi-plateformes basé sur le language C++. Il permet de créer des logiciels avec une interface utilisateur graphique. Les applications peuvent être exécutées sur différents systèmes d’exploitation tels que Windows, macOS, Linux, etc., mais aussi dans un navigateur sans nécessiter de modifications majeures du code source. Pour fonctionner dans un navigateur, une application Qt doit être compilée au format WebAssembly. WebAssembly est un format binaire qui permet d’exécuter du code bas niveau de manière portable et sécurisée dans les navigateurs web modernes. Il est conçu pour compléter les langages de programmation traditionnels utilisés sur le web, tels que JavaScript, en offrant des performances plus élevées pour les applications web. Cette application expérimentale est destinée à être utilisée en interne par les collaborateurs de SDEL Contrôle Commande ainsi que certains clients qui souhaitent l’évaluer. Figure 10: Interface web du produit DigiBox (menu principal) On retrouve des fonctionnalités comme : Configuration des interfaces réseau Gestion des certificats État des cartes entrées/sorties Affichage des informations de débogage Gestion de l’alimentation Import de configuration … L’application gère les utilisateurs du système d’exploitation et les rôles du CAP4000. Ainsi, il est possible de restreindre l’accès à certaines fonctionnalités. Par exemple, un utilisateur avec le rôle “observateur” ne pourra pas accéder à la configuration réseau de l’équipement. L’architecture du cette solution peut être représentée par la pile technologique suivante : Table 1: Pile technologique Frontend API Backend Qt API REST CAP4000 Pile technologique D’abord c’est quoi ? Une pile technologique peut être découpée en trois catégories : Frontend La couche frontend inclus les outils requis pour construire une interface homme machine pour les utilisateurs finaux. Elle peut être développée pour fonctionner dans un navigateur internet ou sur un système d’exploitation (windows, linux, macos, android, ios, …). Backend La couche backend inclus un runtime (environnement d’exécution) serveur généralement accompagné d’une base de données. Elle permet de gérer les données et la logique métier de l’application (gestion des utilisateurs, des droits, fonctionnalités …). API La couche APIs permet de connecter le frontend et le backend (REST, GraphQL). Elle gère aussi l’intégration avec des services tiers (paiement, gestion des identités, messagerie, …) L’application web est accessible localement en accédant à un équipement via son adresse ip depuis un navigateur internet. Elle est compatible avec les navigateurs modernes (Chrome, Firefox, Edge, Safari, …). L’application est donc distribuée directement dans le système d’exploitation d’un équipement et ne nécessite aucune installation sur les postes de travail de nos clients. Chaque type d’équipement dispose de sa propre version de l’application intégrant plus ou moins de fonctionnalités. Dans la suite du rapport, nous allons présenter les différentes étapes de la réalisation de ce projet. Nous commencerons par présenter la problématique et les objectifs du projet. Nous détaillerons ensuite la méthodologie utilisée pour réaliser le projet. Enfin, nous présenterons les résultats obtenus et les perspectives pour la suite du projet. Parler des licenses libres / open sources https://www.diatem.net/les-licences-open-source/ 2 Problématique Aujourd’hui, nous souhaitons intégrer l’application web sur l’ensemble des produits que nous commercialisons. En informatique, les logiciels sont distribués sous plusieurs types de licenses. Les licenses open-sources permettent aux développeurs de les utiliser, de les modifier et de les redistribuer en suivant certaines règles. Il existe beaucoup de licences open-sources et chacune dispose de ses propres règles d’utilisation. Il est important de prendre en compte ces conditions avant de commercialiser un produit qui intègre un composant open-source. Le framework Qt est la propriété de The Qt Company. Il est distribué sous plusieurs licenses. Jusqu’ici nous avons utilisé la version open-source de ce logiciel pour commercialiser nos produits. Cependant cette version ne permet pas de distribuer une application au format WebAssembly sans également distribuer le code source de celle-ci. La direction de VINCI ne souhaite pas publier les sources de ses applications. Nous avons également écarté la possibilité de payer des licenses commerciales. La commercialisation des produits intégrant de l’application web sous sa forme actuelle n’est donc pas envisageable. Le présent projet vise à développer une nouvelle application en se basant sur une autre technologie. Nous souhaitons une interface graphique web reposant sur le dialogue avec un calculateur via l’API REST. Nous souhaitons utiliser un framework open source et entièrement gratuit, adapté aux besoins de l’embarqué. Ce projet permettra également d’explorer et monter en compétences sur les technologies web ainsi que consolider notre solution en utilisant des technologies pérennes qui autorisent la diffusion de nos applications sous licence propriétaire. Problématique Comment choisir judicieusement un framework open source pour le développement d’une nouvelle application web intégrée à l’ensemble de la gamme de produits, tout en respectant les contraintes de confidentialité imposées par la direction de l’entreprise et en garantissant la possibilité de diffuser les applications sous licence propriétaire ? Ce projet aura un impact sur les utilisateurs de l’application Qt. Comme cette application était encore en phase expérimentale, seuls les développeurs du service recherche et développement et les intervenants internes divers (test, bureau d’études, etc …) seront impactés. Nos client qui profitent déjà de l’application seront limités à l’utilisation sur la plateforme Windows uniquement. Lorsque la prochaine application sera développée, ils pourront bénéficier d’une mise à niveau vers la nouvelle version système de leur produit. L’application Qt continuera à être développée par Jean-Louis Baudoin. Un autre groupe de travail a réussi à reproduire le fonctionnement de l’application web en explorant des solutions alternatives au WebAssembly. Cela est rendu possible en installant l’application de bureau directement sur les calculateurs et partageant l’environnement de bureau grace aux technologies de type VNC. Cette solution, rapide à mettre en place, restera cependant temporaire en raison de ses faible performances et sera à terme définitivement remplacée définitivement lorsque la future application sera publiée. 3 Buts et Objectifs Choisir un framework Établir une liste de minimum 10 critères permettant la comparaison des frameworks Évaluer et tester au moins trois frameworks alternatifs en fonction la liste de critères Réaliser un prototype (POC) fonctionnel comprenant un menu ainsi que 3 modules Préparer la transition Mettre à niveau la spécification de l’API pour répondre au standard OpenAPI 3 Modifier la méthode d’authentification des requêtes pour simplifier l’implémentation de l’API dans la nouvelle application Changer le système d’authentification afin de garantir la sécurité de l’API Effectuer une semaine d’auto-formation sur le framework choisi Établir une liste ordonnée de modules à porter, à créer ou à supprimer comprenant au minimum les modules existants Développer la nouvelle application web Développer au moins 20 modules Atteindre des performances au moins équivalentes (dégradation maximale de 5%) à l’application actuelle Acquérir les compétences d’un ingénieur débutant Mettre en place une démarche de gestion de projet permettant d’atteindre les buts et objectifs fixés Fournir l’ensemble des livrables demandés par l’ESEO Avoir une note supérieure à 14 aux 3 évaluations PING Bien sûr, voici une proposition de structure alternative pour les objectifs, avec des formulations plus étoffées et professionnelles : 3.1 Objectifs Stratégiques et Opérationnels 3.1.1 Sélectionner un Cadre de Développement Optimal Identifier et choisir un framework adapté aux besoins spécifiques du projet. Élaborer une Liste de Critères de Sélection : Objectif : Développer une liste détaillée d’au moins 10 critères pertinents pour la comparaison des frameworks. Délai : 2 semaines. Description : Inclure des critères tels que la performance, la scalabilité, la facilité d’intégration, le support communautaire, la documentation, et la sécurité. Évaluer et Tester des Frameworks Alternatifs : Objectif : Évaluer au moins trois frameworks alternatifs en fonction de la liste de critères établie. Délai : 4 semaines. Description : Effectuer des tests pratiques pour chaque framework afin de vérifier leur conformité aux critères et documenter les résultats. Développer un Prototype Fonctionnel : Objectif : Réaliser un Proof of Concept (POC) comprenant un menu principal et trois modules fonctionnels. Délai : 6 semaines. Description : Utiliser le framework sélectionné pour développer un prototype démontrant les capacités du framework à répondre aux exigences du projet. 3.1.2 Préparer la Transition Technologique Assurer une transition fluide et sécurisée vers le nouveau framework. Mettre à Niveau la Spécification de l’API : Objectif : Conformer la spécification de l’API au standard OpenAPI 3. Délai : 2 semaines. Description : Revoir et modifier la spécification actuelle pour garantir la compatibilité et les bonnes pratiques. Modifier le Système d’Authentification : Objectif : Mettre en place un système d’authentification robuste pour garantir la sécurité de l’API. Délai : 3 semaines. Description : Intégrer des méthodes modernes d’authentification (par exemple, OAuth2) pour renforcer la sécurité et simplifier l’implémentation. Former l’Équipe sur le Nouveau Framework : Objectif : Effectuer une semaine d’auto-formation pour l’ensemble de l’équipe sur le framework sélectionné. Délai : 1 semaine. Description : Utiliser des ressources en ligne et des formations internes pour acquérir les compétences nécessaires. Établir une Liste de Modules à Migrer : Objectif : Créer une liste ordonnée des modules existants à porter, créer ou supprimer. Délai : 2 semaines. Description : Prioriser les modules en fonction de leur importance et de leur complexité, en incluant une évaluation des efforts requis pour chaque module. Minimum viable product : Liste modules 3.1.3 Développer la Nouvelle Application Web Concevoir et déployer une nouvelle application web performante et fonctionnelle. Développer les Modules Nécessaires : Objectif : Concevoir et développer au moins 20 modules fonctionnels pour la nouvelle application. Délai : 3 mois. Description : Chaque module doit être testé et validé selon les critères de qualité et de performance. Optimiser les Performances de l’Application : Objectif : Assurer que les performances de la nouvelle application ne se dégradent pas de plus de 5% par rapport à l’application actuelle. Délai : 1 mois. Description : Effectuer des tests de performance réguliers et optimiser le code et l’infrastructure en conséquence. 3.1.4 Objectif 4 : Développer les Compétences en Gestion de Projet et Techniques But : Acquérir et démontrer les compétences nécessaires pour réussir dans le cadre du projet. Mettre en Place une Démarche de Gestion de Projet : Objectif : Développer une approche structurée de gestion de projet pour atteindre les objectifs fixés. Délai : 1 semaine. Description : Utiliser des outils de gestion de projet (comme Jira ou Trello) pour suivre les tâches, les progrès et les délais. Fournir les Livrables Requis : Objectif : Produire et livrer tous les livrables exigés par l’ESEO. Délai : Selon les échéances établies. Description : Assurer la qualité et la complétude de tous les documents et livrables. Obtenir des Notes Élevées aux Évaluations PING : Objectif : Obtenir une note supérieure à 14/20 aux trois évaluations PING. Délai : Prochaine session d’évaluations. Description : Préparer et réviser les matières évaluées pour garantir une performance optimale. 4 Démarche Comment fonctionne le service R&D de base outils de gestion de projet Redmine Fiche de développement Notes Outils - Gestion de projet - Gestion de version - Environnement de développement - Produits de développement - Langages de programmation 4.1 Démarche générale La démarché générale du projet est définie par le présent document. Ce document défini les buts et objectif attendus. Il établis également un plan d’action pour les atteindre. Il mets en place un processus de communication pour garantir la circulation de l’information. Enfin, il identifie les risques et problèmes potentiels qui pourraient survenir et établi des solutions pour y faire face. 4.2 Méthodologie, techniques et technologies Le développement logiciel suit plusieurs étapes essentielles pour assurer la création et le bon fonctionnement d’un logiciel. Analyse des besoins : Identifier et définir les besoins des utilisateurs et les objectifs du logiciel. Les exigences fonctionnelles et non fonctionnelles sont spécifiées, ainsi que les contraintes du projet. Conception : Créer une architecture logicielle en se basant sur l’analyse des besoins. Cela implique la création de diagrammes, de schémas et de modèles qui servent de guide pour la réalisation du logiciel. Tests : Vérifier la qualité et la conformité du logiciel. Des tests unitaires, d’intégration et de validation sont réalisés pour détecter et corriger les éventuelles erreurs et bugs. Intégration : Assembler les différentes parties du logiciel en un ensemble fonctionnel et cohérent. Cela inclut l’ajout de fonctionnalités supplémentaires et l’optimisation des performances. Déploiement : Rendre le logiciel disponible aux utilisateurs finaux. Cela peut impliquer l’installation sur des serveurs, la distribution de fichiers d’installation ou la mise à disposition sur des plateformes en ligne. Maintenance : Maintenir le logiciel en état de fonctionnement. Les erreurs sont corrigées, des mises à jour sont effectuées, des nouvelles fonctionnalités peuvent être ajoutées et des améliorations sont apportées pour répondre aux besoins changeants des utilisateurs. Dans le cadre de ce projet, une approche de gestion de projet agile sera utilisée. Cette approche permettra une meilleure gestion des changements et des imprévus, tout en limitant l’effet tunnel. Méthodologie agile La méthodologie agile est une approche de gestion de projet qui se caractérise par sa flexibilité, sa collaboration continue avec les parties prenantes et sa capacité à s’adapter aux changements tout au long du cycle de développement. Les méthodologies agiles mettent l’accent sur la livraison itérative et incrémentale du produit, favorisant des cycles de développement courts et des retours fréquents des utilisateurs. L’agilité est souvent utilisée dans le développement logiciel, mais elle peut également être appliquée à d’autres domaines. Pour plus d’informations sur les méthodologies agiles, voir https://www.atlassian.com/fr/agile. La phase de développement sera découpée en sprints. Chaque période entreprise (2 à 4 semaines) intégrera un unique sprint. Cette segmentation permettra d’effectuer des contrôles réguliers de l’avancement et favorisera un dialogue continu. Sprint Un sprint en développement agile est un cycle de travail itératif et incrémental qui permet à une équipe de développer et de livrer des fonctionnalités de manière régulière, tout en restant flexible et adaptative aux changements et aux retours client. Un sprint intègre généralement les étapes suivantes : Planification des tâches en début du sprint Développement et test continu Revues et rétrospectives en fin de sprint Livraison logicielle potentielle Le projet utilisera un backlog pour recueillir, organiser et hiérarchiser l’ensemble des travaux à réaliser dans le cadre du projet. Chaque fonctionnalité à développer sera spécifiée en une user story et ajoutée comme ticket dans le backlog. Une user story est une technique de description des besoins du client du point de vue de ce dernier. L’objectif de chaque user story est de décrire une fonctionnalité du logiciel de manière simple et compréhensible, tout en se concentrant sur la valeur apportée à l’utilisateur final. Chaque user story pourra dépendre de plusieurs tâches de développements, qui seront rattachés à ce ticket. Au début de chaque sprint, le backlog devra être complété en ajoutant les tâches à traiter pour les trois semaines à venir. Si nécessaire, une modélisation de la fonctionnalité à implémenter pourra être réalisée pour répondre aux spécifications établies précédemment. Chaque sprint devra intégrer la spécification, le développement, les tests et la documentation d’une ou plusieurs user story. Un réunion sera effectué à la fin de chaque sprint afin de suivre l’état d’avancement et d’exercer un esprit critique. Le but de ces réunions sera de replacer les tickets qui n’ont pas pu être traités et d’effectuer un bilan de ce qui à marché et ce qui n’a pas fonctionné pour mieux organiser les sprints futurs. Ces réunions seront basés sur un unique document de suivi, qui sera mis à jour durant chaque sprint. L’utilisation d’UML sera privilégiée pour la modélisation des différents aspects du projet, tels que les cas d’utilisation, les diagrammes de classes, les diagrammes de séquence, etc. Cela permettra de décrire plus précisément les besoins du projet et de mieux visualiser les différentes parties impliquées. 4.3 Analyse des risques Table 2: SWOT (Strengths Weaknesses Opportunities Threats) Atouts Handicapes Interne Forces Maintenir la confidentialité du code source Economies de coûts à long terme Étendre et renforcer les compétences techniques de l’équipe de développement Faiblesses Coûteux en termes de temps et de ressources Peut ne pas répondre aux besoins Difficultés à apprendre un nouveau framework Externe Opportunités Nouvelles opportunités de développement pour l’entreprise Élargir son champ d’expertise Bénéficier de l’expertise de la communauté open source Menaces Coûts de licence prohibitifs ou des restrictions d’utilisation Difficulté à trouver une alternative qui soit compatible avec la cible embarquée Retards dans le développement en raison de l’apprentissage d’un nouveau framework Manque de compétences sur le nouveau framework et les technologies associées Établir un calendrier pour l’apprentissage et la mise en œuvre du nouveau framework Offrir une formation à l’équipe de développement pour aider à l’apprentissage du nouveau framework Mobiliser des ressources supplémentaires pour maintenir les délais Communauté du framework limitée Effectuer une recherche sur les communautés de développeurs pour identifier les frameworks les plus populaires Évaluer la taille de la communauté et la fréquence des mises à jour Évaluer la qualité de la documentation et des exemples Manque de ressources Augmenter l’équipe de développement pour combler les manques Étendre les délais du projet Réduire les fonctionnalités à développer Incompatibilité avec la cible embarquée Examiner les exigences de la cible embarquée et les spécifications de l’alternative Effectuer des tests de compatibilité pour vérifier si l’alternative répond aux exigences de la cible embarquée Résistance au changement Communiquer sur les avantages du nouveau framework choisi Mettre en place une formation pour les développeurs qui ne connaissent pas le nouveau framework Difficulté à trouver une alternative viable Effectuer une analyse des besoins pour identifier les caractéristiques essentielles que l’alternative doit posséder Effectuer une recherche sur les coûts des licences pour les différentes alternatives Quadrant chart risk analysis 4.4 Acteurs Table 3: Acteurs Acteur Rôle Description Joshua Jourdam Chef de projet, Développeur Fourni le livrable du projet. Conduit et pilote le projet. Anime les différentes réunions. Principal développeur. Sébastien Barré Sponsor du projet, Client Responsable global du projet et est un soutien pour le chef de projet. Prend les décisions importantes et arbitre entre deux choix, notamment en situation de crise. Détermine les attendus du projet et les exigences liées. Validera la conformité du livrable à ses attentes et exigences. Opérateurs, testeurs, développeurs Utilisateurs finaux Personnes qui vont utiliser le produit ou le service au quotidien. Exprime leur soutien ou leur mécontentement vis à vis du projet. Jean-Louis Baudoin Utilisateur clé Référent métier, développeur de l’application Qt. Sera le principal interlocuteur pour collecter des informations sur l’application actuelle. David Dupond Utilisateur clé Référent licenses, pourra valider la license d’utilisation du nouvel outil. Lucas Cecillon Utilisateur clé Référent cybersécurité, pourra être consulté sur les problématiques de sécurité de l’application et de l’API. Stagiaires Support Peuvent intervenir en renfort sur le projet, notamment lors des périodes estivales. 4.5 Lotissements %%| label: fig-work-breakdown-structure %%| fig-cap: WBS (Work Breakdown Structure) %%| file: ./graphics/wbs.mmd 4.6 Planning prévisionnel #fig-forecast-schedule-mermaid{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#fig-forecast-schedule-mermaid .error-icon{fill:#a44141;}#fig-forecast-schedule-mermaid .error-text{fill:#ddd;stroke:#ddd;}#fig-forecast-schedule-mermaid .edge-thickness-normal{stroke-width:2px;}#fig-forecast-schedule-mermaid .edge-thickness-thick{stroke-width:3.5px;}#fig-forecast-schedule-mermaid .edge-pattern-solid{stroke-dasharray:0;}#fig-forecast-schedule-mermaid .edge-pattern-dashed{stroke-dasharray:3;}#fig-forecast-schedule-mermaid .edge-pattern-dotted{stroke-dasharray:2;}#fig-forecast-schedule-mermaid .marker{fill:lightgrey;stroke:lightgrey;}#fig-forecast-schedule-mermaid .marker.cross{stroke:lightgrey;}#fig-forecast-schedule-mermaid svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#fig-forecast-schedule-mermaid .mermaid-main-font{font-family:"trebuchet ms",verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-forecast-schedule-mermaid .exclude-range{fill:hsl(52.9411764706, 28.813559322%, 48.431372549%);}#fig-forecast-schedule-mermaid .section{stroke:none;opacity:0.2;}#fig-forecast-schedule-mermaid .section0{fill:hsl(52.9411764706, 28.813559322%, 58.431372549%);}#fig-forecast-schedule-mermaid .section2{fill:#EAE8D9;}#fig-forecast-schedule-mermaid .section1,#fig-forecast-schedule-mermaid .section3{fill:#333;opacity:0.2;}#fig-forecast-schedule-mermaid .sectionTitle0{fill:#F9FFFE;}#fig-forecast-schedule-mermaid .sectionTitle1{fill:#F9FFFE;}#fig-forecast-schedule-mermaid .sectionTitle2{fill:#F9FFFE;}#fig-forecast-schedule-mermaid .sectionTitle3{fill:#F9FFFE;}#fig-forecast-schedule-mermaid .sectionTitle{text-anchor:start;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-forecast-schedule-mermaid .grid .tick{stroke:lightgrey;opacity:0.8;shape-rendering:crispEdges;}#fig-forecast-schedule-mermaid .grid .tick text{font-family:"trebuchet ms",verdana,arial,sans-serif;fill:#ccc;}#fig-forecast-schedule-mermaid .grid path{stroke-width:0;}#fig-forecast-schedule-mermaid .today{fill:none;stroke:#DB5757;stroke-width:2px;}#fig-forecast-schedule-mermaid .task{stroke-width:2;}#fig-forecast-schedule-mermaid .taskText{text-anchor:middle;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-forecast-schedule-mermaid .taskTextOutsideRight{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);text-anchor:start;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-forecast-schedule-mermaid .taskTextOutsideLeft{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);text-anchor:end;}#fig-forecast-schedule-mermaid .task.clickable{cursor:pointer;}#fig-forecast-schedule-mermaid .taskText.clickable{cursor:pointer;fill:#003163!important;font-weight:bold;}#fig-forecast-schedule-mermaid .taskTextOutsideLeft.clickable{cursor:pointer;fill:#003163!important;font-weight:bold;}#fig-forecast-schedule-mermaid .taskTextOutsideRight.clickable{cursor:pointer;fill:#003163!important;font-weight:bold;}#fig-forecast-schedule-mermaid .taskText0,#fig-forecast-schedule-mermaid .taskText1,#fig-forecast-schedule-mermaid .taskText2,#fig-forecast-schedule-mermaid .taskText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);}#fig-forecast-schedule-mermaid .task0,#fig-forecast-schedule-mermaid .task1,#fig-forecast-schedule-mermaid .task2,#fig-forecast-schedule-mermaid .task3{fill:hsl(180, 1.5873015873%, 35.3529411765%);stroke:#ffffff;}#fig-forecast-schedule-mermaid .taskTextOutside0,#fig-forecast-schedule-mermaid .taskTextOutside2{fill:lightgrey;}#fig-forecast-schedule-mermaid .taskTextOutside1,#fig-forecast-schedule-mermaid .taskTextOutside3{fill:lightgrey;}#fig-forecast-schedule-mermaid .active0,#fig-forecast-schedule-mermaid .active1,#fig-forecast-schedule-mermaid .active2,#fig-forecast-schedule-mermaid .active3{fill:#81B1DB;stroke:#ffffff;}#fig-forecast-schedule-mermaid .activeText0,#fig-forecast-schedule-mermaid .activeText1,#fig-forecast-schedule-mermaid .activeText2,#fig-forecast-schedule-mermaid .activeText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-forecast-schedule-mermaid .done0,#fig-forecast-schedule-mermaid .done1,#fig-forecast-schedule-mermaid .done2,#fig-forecast-schedule-mermaid .done3{stroke:grey;fill:lightgrey;stroke-width:2;}#fig-forecast-schedule-mermaid .doneText0,#fig-forecast-schedule-mermaid .doneText1,#fig-forecast-schedule-mermaid .doneText2,#fig-forecast-schedule-mermaid .doneText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-forecast-schedule-mermaid .crit0,#fig-forecast-schedule-mermaid .crit1,#fig-forecast-schedule-mermaid .crit2,#fig-forecast-schedule-mermaid .crit3{stroke:#E83737;fill:#E83737;stroke-width:2;}#fig-forecast-schedule-mermaid .activeCrit0,#fig-forecast-schedule-mermaid .activeCrit1,#fig-forecast-schedule-mermaid .activeCrit2,#fig-forecast-schedule-mermaid .activeCrit3{stroke:#E83737;fill:#81B1DB;stroke-width:2;}#fig-forecast-schedule-mermaid .doneCrit0,#fig-forecast-schedule-mermaid .doneCrit1,#fig-forecast-schedule-mermaid .doneCrit2,#fig-forecast-schedule-mermaid .doneCrit3{stroke:#E83737;fill:lightgrey;stroke-width:2;cursor:pointer;shape-rendering:crispEdges;}#fig-forecast-schedule-mermaid .milestone{transform:rotate(45deg) scale(0.8,0.8);}#fig-forecast-schedule-mermaid .milestoneText{font-style:italic;}#fig-forecast-schedule-mermaid .doneCritText0,#fig-forecast-schedule-mermaid .doneCritText1,#fig-forecast-schedule-mermaid .doneCritText2,#fig-forecast-schedule-mermaid .doneCritText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-forecast-schedule-mermaid .activeCritText0,#fig-forecast-schedule-mermaid .activeCritText1,#fig-forecast-schedule-mermaid .activeCritText2,#fig-forecast-schedule-mermaid .activeCritText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-forecast-schedule-mermaid .titleText{text-anchor:middle;font-size:18px;fill:#ccc;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-forecast-schedule-mermaid :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;} 01/06 01/07 01/08 01/09 01/10 01/11 01/12 01/01 01/02 01/03 01/04 01/05 01/06 01/07 01/08 01/09Réunion de lancement Collecte informations Rapport avant projet Réunion premier choix OpenAPI Soutenance avant projet Authentification FormationPOC Réunion de validation Feuille de route Sprints Rapport intermédiaire Soutenance intermédiaire Réunion de clôture Rapport final Soutenance final EvaluationPréparationDéveloppementEseo Figure 11: Diagramme de GANTT prévisionnel 4.7 Planning effectif #fig-actual-schedule-mermaid{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#fig-actual-schedule-mermaid .error-icon{fill:#a44141;}#fig-actual-schedule-mermaid .error-text{fill:#ddd;stroke:#ddd;}#fig-actual-schedule-mermaid .edge-thickness-normal{stroke-width:2px;}#fig-actual-schedule-mermaid .edge-thickness-thick{stroke-width:3.5px;}#fig-actual-schedule-mermaid .edge-pattern-solid{stroke-dasharray:0;}#fig-actual-schedule-mermaid .edge-pattern-dashed{stroke-dasharray:3;}#fig-actual-schedule-mermaid .edge-pattern-dotted{stroke-dasharray:2;}#fig-actual-schedule-mermaid .marker{fill:lightgrey;stroke:lightgrey;}#fig-actual-schedule-mermaid .marker.cross{stroke:lightgrey;}#fig-actual-schedule-mermaid svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#fig-actual-schedule-mermaid .mermaid-main-font{font-family:"trebuchet ms",verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-actual-schedule-mermaid .exclude-range{fill:hsl(52.9411764706, 28.813559322%, 48.431372549%);}#fig-actual-schedule-mermaid .section{stroke:none;opacity:0.2;}#fig-actual-schedule-mermaid .section0{fill:hsl(52.9411764706, 28.813559322%, 58.431372549%);}#fig-actual-schedule-mermaid .section2{fill:#EAE8D9;}#fig-actual-schedule-mermaid .section1,#fig-actual-schedule-mermaid .section3{fill:#333;opacity:0.2;}#fig-actual-schedule-mermaid .sectionTitle0{fill:#F9FFFE;}#fig-actual-schedule-mermaid .sectionTitle1{fill:#F9FFFE;}#fig-actual-schedule-mermaid .sectionTitle2{fill:#F9FFFE;}#fig-actual-schedule-mermaid .sectionTitle3{fill:#F9FFFE;}#fig-actual-schedule-mermaid .sectionTitle{text-anchor:start;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-actual-schedule-mermaid .grid .tick{stroke:lightgrey;opacity:0.8;shape-rendering:crispEdges;}#fig-actual-schedule-mermaid .grid .tick text{font-family:"trebuchet ms",verdana,arial,sans-serif;fill:#ccc;}#fig-actual-schedule-mermaid .grid path{stroke-width:0;}#fig-actual-schedule-mermaid .today{fill:none;stroke:#DB5757;stroke-width:2px;}#fig-actual-schedule-mermaid .task{stroke-width:2;}#fig-actual-schedule-mermaid .taskText{text-anchor:middle;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-actual-schedule-mermaid .taskTextOutsideRight{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);text-anchor:start;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-actual-schedule-mermaid .taskTextOutsideLeft{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);text-anchor:end;}#fig-actual-schedule-mermaid .task.clickable{cursor:pointer;}#fig-actual-schedule-mermaid .taskText.clickable{cursor:pointer;fill:#003163!important;font-weight:bold;}#fig-actual-schedule-mermaid .taskTextOutsideLeft.clickable{cursor:pointer;fill:#003163!important;font-weight:bold;}#fig-actual-schedule-mermaid .taskTextOutsideRight.clickable{cursor:pointer;fill:#003163!important;font-weight:bold;}#fig-actual-schedule-mermaid .taskText0,#fig-actual-schedule-mermaid .taskText1,#fig-actual-schedule-mermaid .taskText2,#fig-actual-schedule-mermaid .taskText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%);}#fig-actual-schedule-mermaid .task0,#fig-actual-schedule-mermaid .task1,#fig-actual-schedule-mermaid .task2,#fig-actual-schedule-mermaid .task3{fill:hsl(180, 1.5873015873%, 35.3529411765%);stroke:#ffffff;}#fig-actual-schedule-mermaid .taskTextOutside0,#fig-actual-schedule-mermaid .taskTextOutside2{fill:lightgrey;}#fig-actual-schedule-mermaid .taskTextOutside1,#fig-actual-schedule-mermaid .taskTextOutside3{fill:lightgrey;}#fig-actual-schedule-mermaid .active0,#fig-actual-schedule-mermaid .active1,#fig-actual-schedule-mermaid .active2,#fig-actual-schedule-mermaid .active3{fill:#81B1DB;stroke:#ffffff;}#fig-actual-schedule-mermaid .activeText0,#fig-actual-schedule-mermaid .activeText1,#fig-actual-schedule-mermaid .activeText2,#fig-actual-schedule-mermaid .activeText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-actual-schedule-mermaid .done0,#fig-actual-schedule-mermaid .done1,#fig-actual-schedule-mermaid .done2,#fig-actual-schedule-mermaid .done3{stroke:grey;fill:lightgrey;stroke-width:2;}#fig-actual-schedule-mermaid .doneText0,#fig-actual-schedule-mermaid .doneText1,#fig-actual-schedule-mermaid .doneText2,#fig-actual-schedule-mermaid .doneText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-actual-schedule-mermaid .crit0,#fig-actual-schedule-mermaid .crit1,#fig-actual-schedule-mermaid .crit2,#fig-actual-schedule-mermaid .crit3{stroke:#E83737;fill:#E83737;stroke-width:2;}#fig-actual-schedule-mermaid .activeCrit0,#fig-actual-schedule-mermaid .activeCrit1,#fig-actual-schedule-mermaid .activeCrit2,#fig-actual-schedule-mermaid .activeCrit3{stroke:#E83737;fill:#81B1DB;stroke-width:2;}#fig-actual-schedule-mermaid .doneCrit0,#fig-actual-schedule-mermaid .doneCrit1,#fig-actual-schedule-mermaid .doneCrit2,#fig-actual-schedule-mermaid .doneCrit3{stroke:#E83737;fill:lightgrey;stroke-width:2;cursor:pointer;shape-rendering:crispEdges;}#fig-actual-schedule-mermaid .milestone{transform:rotate(45deg) scale(0.8,0.8);}#fig-actual-schedule-mermaid .milestoneText{font-style:italic;}#fig-actual-schedule-mermaid .doneCritText0,#fig-actual-schedule-mermaid .doneCritText1,#fig-actual-schedule-mermaid .doneCritText2,#fig-actual-schedule-mermaid .doneCritText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-actual-schedule-mermaid .activeCritText0,#fig-actual-schedule-mermaid .activeCritText1,#fig-actual-schedule-mermaid .activeCritText2,#fig-actual-schedule-mermaid .activeCritText3{fill:hsl(28.5714285714, 17.3553719008%, 86.2745098039%)!important;}#fig-actual-schedule-mermaid .titleText{text-anchor:middle;font-size:18px;fill:#ccc;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#fig-actual-schedule-mermaid :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;} 01/06 01/07 01/08 01/09 01/10 01/11 01/12 01/01 01/02 01/03 01/04 01/05 01/06 01/07 01/08 01/09Réunion de lancement Recherche technologies Rapport avant projet Soutenance avant projet POC Spécification API Spécification fonctionelle Présentation des frameworks Présentation application React / Choix définitif Création du projet Sprint 1 Point authentification Authentification Point stratégie de tests Validation specification API Fiche de développement V1 Test déploiement sur cible Sprint 2 Sprint 3 Soutenance intermédiaire Sprint 4 Formation React Introduction plateforme de développementSprint 5 Rapport final Soutenance finale EvaluationPréparationDéveloppementEseo Figure 12: Diagramme de GANTT effectif 4.7.1 Jalons Réunion de lancement : Cadre le projet, enjeux, objectifs, finalité, jalons projet, risques identifiés, calendrier projet, … Réunion premier choix : Choix du framework pour commencer le POC Réunion de validation : Validation définitive du choix du framework, planification des étapes de développement Réunion en début de sprint : Tri, organisation, revue des tâches à réaliser Réunion de fin de sprint: Bilan, présentation des tâches réalisés, revue des spécification, replanification des tâches non réalisées, … Réunion de clôture : Fin du projet, assure une transition pour la maintenance et l’exploitation du projet 4.7.2 Livrables Document de comparaison des frameworks avec une évaluation quantitative de chaque critère, basée sur une analyse détaillée. Document de spécification, détaillant les fonctionnalités, les exigences système, les ressources requises et les performances attendues. Prototype (POC) fonctionnel de l’interface web, respectant toutes les spécifications documentées. Feuille de route pour la mise en place du framework, détaillant les étapes à suivre pour la pour le développement des fonctionnalités, les ressources requises et les délais. Documentation développeur qui comprends un guide de démarrage et les bonnes pratiques de développement. Documentation utilisateur qui comprends un guide d’utilisation et les bonnes pratiques d’utilisation. Rapports et soutenances de projet. 4.8 Budget La durée du projet PING est estimée à 500 heures minimum. Ce projet devrait nécessiter plus de ressources. Le temps total nécessaire à la réalisation du projet est difficilement quantifiable pour le moment. Étant donné que le projet est cadré pour une seule personne, le coût peut être calculé en utilisant le tarif horaire du salaire brut annuel moyen d’un ingénieur débutant. Cela représente environ 10 000 €. Des coûts d’intervention de développement externes supplémentaires ponctuels peuvent s’ajouter, mais ils sont plus difficiles à chiffrer. La mobilisation de certains membres de l’équipe logiciel lors des réunions peut également être prise en compte. 5 Résultats 5.1 Evaluation des technologies du marché La première phase du projet visait à trouver une technologie alternative au framework Qt pour développer des application pour la plateforme web. Pour cela, j’ai établi une liste de critères permettant d’évaluer chaque solution. Besoins développement Besoins fonctionnels Langage de programmation Communauté et support Proximité avec l’existant Date de sortie Documentation Évolutivité Licence Outils Patron de conception Plateformes Popularité Possibles freins J’ai identifié l’ensemble des technologies qui permettent de développer des application pour la plateforme Web. Ces technologies peuvent être regroupées en 2 catégories, natif et multi-plateforme. La première catégorie comprends l’ensemble des frameworks basés sur le langage JavaScript. La seconde réuni quand à elle des solutions plus diverses qui ce basent sur différent langages comme le C#, le Dart ou le C++. Ces technologies peuvent permettre de cibler plusieurs plateformes (web, bureau, mobile) avec une base de code unique. Ce premier découpage m’a amener à tester 5 solutions. Natif : React Vue Angular Multi-plateforme : Flutter Uno platform J’ai établi une procédure de test unique. L’objectif était de comprendre le fonctionnement basique et l’architecture des frameworks ainsi que les langages sur lesquels ils sont basés. Le test consistait à réaliser une page affichant des données reçues par l’API REST et un formulaire permettant de modifier le nom d’hôte de l’équipement. A partir de cette petite étude, j’ai conclu que les frameworks multi-plateformes sont des solution moins adaptés à notre besoin. Elle sont plus complexes et n’ont pas une intégration complète avec le fonctionnalités d’un navigateur web. Les solutions natives semblaient plus simples à utiliser. La prochaine étape consistait à réaliser un POC plus complet qui intègre les fonctionnalités de certains modules de l’application Qt. J’ai opté pour l’utilisation de React, le framework les plus largement utilisé. J’ai supervisé activement la période de stage de Romain LeDivenah, qui a pris en charge la réalisation de ce POC. Malgré mon statut de débutant, j’ai pu guider Romain dans la prise en main de React. Au cours du mois de supervision, l’objectif était de développer un prototype fonctionnel, répondant aux critères définis, comprenant un menu ainsi que trois modules distincts. 5.2 Transition 5.2.1 OpenAPI La spécification de l’API est définie dans un document word qui évolue au fur et à mesure de l’ajout de fonctionnalités. Cette documentation peut être améliorée. Il existe des outils spécialisés respectant les standards de l’industrie. Ils permettent de générer la documentation, la modélisation et du code a partir d’un fichier de description. J’ai souhaité intégrer la migration vers une spécification de notre API REST basée sur la norme OpenAPI 3 pour avoir a disposition une spécification qui suis des règles appliquées au marché actuel. Il existe de nombreux outil qui utilisent cette norme que je voulais également intégrer. Le stage de Romain LeDivenah m’a permis de travailler sur d’autres aspects en parallèle comme la spécification de l’api, et l’intégration d’outils. J’ai rédigé la spécification de l’API REST en suivant la norme OpenAPI 3.0.0. J’ai également mis en place l’outil Swagger UI pour mettre à disposition une documentation interactive de l’API à partir de cette spécification. J’ai aussi utilisé OpenAPI Generator pour générer un SDK client en JavaScript. Cet étape permettra de simplifier l’utilisation de l’API REST dans l’application à développer. Enfin le dernier outil que j’ai déployé est Spotlight Prism. C’est un serveur simulé HTTP open source qui permet d’émuler le comportement de notre API à partir d’un jeu d’exemple qui peut être défini dans la spécification. Cet outil permet de tester l’API sans avoir d’équipement à disposition ou de développer la partie graphique avant ou en parallèle de l’intégration d’une nouvelle fonctionnalité dans le CAP4000. Exemple de documentation générée avec Swagger UI 5.2.2 Authentification L’authentification de l’API est basée sur un secret partagé. Cette solution pouvait être utilisé avec l’application Qt car celle-ci était compilée en un format binaire. Cela présente cependant des problématiques majeures, particulièrement dans le contexte du langage JavaScript. Étant un langage interprété, le code source JavaScript est généralement envoyé au client, exposant ainsi le secret partagé au sein du code. Cette exposition représente une menace sérieuse pour la sécurité de l’API, car un utilisateur malveillant pourrait potentiellement accéder au code source, récupérer le secret partagé et compromettre l’authentification. L’utilisation du protocole HTTP (non sécurisé) pour communiquer avec l’API représente un risque supplémentaire. En effet, les données envoyées sur le réseau ne sont pas chiffrées, ce qui permet à un attaquant d’intercepter les requêtes et d’obtenir le secret partagé. La vulnérabilité de cette approche souligne la nécessité de stratégies plus sécurisées, telles que l’utilisation de méthodes d’authentification basées sur des jetons et l’utilisation du protocole HTTPS pour communiquer avec l’API. Lucas CECILLON, référent cybersécurité, a pu travailler sur cette partie en spécifiant et en implémentant un nouveau mécanisme d’authentification basé sur l’utilisation d’un jeton JWT et l’utilisation du protocole HTTPS. Ce mécanisme à également introduit d’autres problématiques comme la gestion des certificats utilisés pour chiffrer la connexion HTTPS sur un équipement. Le protocole HTTP est utilisé pour communiquer avec l’API REST. L’authentification est basée sur un secret partagé. Ce secret est stocké dans le code source de l’application Qt. L’authentification est effectuée avec plusieurs sécurités : Signature des requêtes avec un HMAC (basé sur le secret partagé) Timestamp pour éviter les attaques de rejeu Contrairement à l’IHM Web en QT utilisant WebAssembly (projet compilé puis exécuté directement par le navigateur), React envoie les sources au navigateur (JS, HTML, CSS) pour y être exécutées. Le secret est donc facilement interceptable et visible par n’importe qui. Par conséquent, il est essentiel de mettre en place une gestion plus sécurisée de la session utilisateur. Plusieurs solutions ont été envisagées pour sécuriser l’authentification : - JWT - Bearer - Session Cookie - Basic La solution mise en place s’axe autour de trois points : Utilisation du protocole HTTPS pour sécuriser les échanges Utilisation de JWT (JSON Web Token) pour l’authentification Gestion des certificats pour chiffrer les communications Suivant les recommandations de l’anssi 5.2.2.1 Mesures de protection SSL CSRF CSP CORS Protection des cookies HSTS XSS Rafraîchissement du token Cryptographie Rejeux 5.2.2.2 HTTPS Selon l’étude [DR01] réalisée, il est essentiel d’établir une communication entre le client et l’API en utilisant le protocole HTTPS (HTTP sécurisé avec TLS). Cela signifie que chaque appel à l’API doit obligatoirement être effectué via le port HTTPS. Par exemple : https://192.168.0.1:3001/api/v3/test. De la même manière, il est impératif que l’utilisateur ne puisse accéder à l’application Web qu’en utilisant le protocole HTTPS. Aucun accès en HTTP ne pourra être fait. Tout cela nécessite une gestion appropriée des certificats sur le calculateur afin de prévenir d’éventuels problèmes d’accès en cas de certificats corrompus ou expirés. 5.2.2.3 JWT Dorénavant, l’authentification des requêtes sera basée sur l’utilisation de jetons d’accès. Ces jetons seront générés par l’API et renvoyés une fois l’authentification effectuée. Par la suite, pour chaque requête jusqu’à la déconnexion du client, le jeton devra être inclus dans la demande envoyée à l’API. Avant de traiter une demande, l’API vérifiera systématiquement le contenu et l’intégrité du jeton de la manière suivante : • Est-il valide ? ◦ SHA256(header + « . » + payload) == signature  • Est-il expiré ? ◦ Valeur « exp » • L’utilisateur a-t-il le droit d’effectuer cette requête ? ◦ Valeur « role » Selon l’étude [DR01] réalisée, l’API doit renvoyer le token dans un cookie sécurisé. Voici les paramètres à appliquer au cookie : • name : name (nom du cookie) • value : value (valeur du token) • domain : domain (nom de domaine du site, ou ip) • path : « / » (chemin du cookie) • sameSite : strict (politique de cloisonnement) • httpOnly : true (inaccessible côté client) • session : true (cookie de session) • secure : true (portée limitée aux canaux sécurisés) L’API doit spécifier le domaine du cookie en fonction de l’adresse IP à partir de laquelle elle a été sollicitée. 5.2.2.4 Certificats La mise en place d’une nouvelle version de l’IHM et d’API qui communiquent exclusivement via HTTPS soulève des questions concernant le gestion des certificats sur le calculateur. En effet, si l’utilisateur de l’IHM parvient à déposer sur le calculateur un couple de clé-privée/certificat corrompus, alors le serveur web ne sera pas joignable en HTTPS. Cela empechera toute configuration par le client, ce qui n’est pas souhaitable. Pour éviter cette situation, il est essentiel de mettre en œuvre une solution de gestion de la configuration TLS robuste, reposant sur trois piliers principaux : la génération automatique de certificats auto-signés de secours, la sécurisation de l’import des certificats et de leur utilisation. Flowchart fallback certificats 5.2.3 POC supervision stage 5.3 Développement 5.3.1 Spécification de l’application Figure 13: Redmine IHM Web CAP4000 Figure 14: User story : configuration des interfaces réseaux L’application Qt n’a pas de spécification propre. J’ai utilisé le logiciel Redmine pour créer un backlog de l’ensemble des fonctionnalités à développer. J’ai également utilisé ce logiciel pour créer des user stories qui décrivent les fonctionnalités à développer. Ces user stories sont utilisées pour définir les critères d’acceptation des fonctionnalités. Elles sont également utilisées pour suivre l’avancement du développement. Evolution spécification Avant / Après Comparaison Utilité Changements 5.3.2 Fonctionnement de la solution On garde le meme principe voir schema contexte Fonctionnement de React TODO 5.3.3 Architecture/Workspace (expliquer comment fonctionne un monorepo) CI/CD (commitlint, semantic versioning, changelog) Développement (branches, pull request) Comparaison SVN Git -> gitlab J’ai créé l’environnement du projet dans un “monorepo”. C’est un référentiel unique qui permet de contenir plusieurs projet. J’ai choisi cette approche car nous souhaitons développer une application par produit. Comme l’application Qt, le code sera découpé en modules qui pourront être intégrés ou non dans l’application d’un produit. J’ai également mis en place une démarche qualité du code et des méthodologies de développement. L’application adopte la norme semantic versioning pour gérer les versions des projets. J’ai préparé un changelog pour répertorier toutes le évolution de l’application. J’ai finalement intégré l’outil commitlint pour vérifier que les messages de commit suivent une convention de rédaction. Le projet est versionné avec l’outil de gestion de configuration Subversion. Le projet suivra une méthodologie de développement en branches. Chaque fonctionnalité sera développée dans une branche dédiée. Une fois la fonctionnalité terminée, la branche sera fusionnée dans la branche principale. Cette approche permet de travailler sur plusieurs fonctionnalités en parallèle et de garder une branche principale stable. #mermaid-8{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;fill:#ccc;}#mermaid-8 .error-icon{fill:#a44141;}#mermaid-8 .error-text{fill:#ddd;stroke:#ddd;}#mermaid-8 .edge-thickness-normal{stroke-width:2px;}#mermaid-8 .edge-thickness-thick{stroke-width:3.5px;}#mermaid-8 .edge-pattern-solid{stroke-dasharray:0;}#mermaid-8 .edge-pattern-dashed{stroke-dasharray:3;}#mermaid-8 .edge-pattern-dotted{stroke-dasharray:2;}#mermaid-8 .marker{fill:lightgrey;stroke:lightgrey;}#mermaid-8 .marker.cross{stroke:lightgrey;}#mermaid-8 svg{font-family:"trebuchet ms",verdana,arial,sans-serif;font-size:16px;}#mermaid-8 .commit-id,#mermaid-8 .commit-msg,#mermaid-8 .branch-label{fill:lightgrey;color:lightgrey;font-family:'trebuchet ms',verdana,arial,sans-serif;font-family:var(--mermaid-font-family);}#mermaid-8 .branch-label0{fill:#2c2c2c;}#mermaid-8 .commit0{stroke:hsl(180, 1.5873015873%, 48.3529411765%);fill:hsl(180, 1.5873015873%, 48.3529411765%);}#mermaid-8 .commit-highlight0{stroke:rgb(133.6571428571, 129.7428571428, 129.7428571428);fill:rgb(133.6571428571, 129.7428571428, 129.7428571428);}#mermaid-8 .label0{fill:hsl(180, 1.5873015873%, 48.3529411765%);}#mermaid-8 .arrow0{stroke:hsl(180, 1.5873015873%, 48.3529411765%);}#mermaid-8 .branch-label1{fill:lightgrey;}#mermaid-8 .commit1{stroke:hsl(321.6393442623, 65.5913978495%, 38.2352941176%);fill:hsl(321.6393442623, 65.5913978495%, 38.2352941176%);}#mermaid-8 .commit-highlight1{stroke:rgb(93.5483870969, 221.4516129033, 139.677419355);fill:rgb(93.5483870969, 221.4516129033, 139.677419355);}#mermaid-8 .label1{fill:hsl(321.6393442623, 65.5913978495%, 38.2352941176%);}#mermaid-8 .arrow1{stroke:hsl(321.6393442623, 65.5913978495%, 38.2352941176%);}#mermaid-8 .branch-label2{fill:lightgrey;}#mermaid-8 .commit2{stroke:hsl(194.4, 16.5562913907%, 49.6078431373%);fill:hsl(194.4, 16.5562913907%, 49.6078431373%);}#mermaid-8 .commit-highlight2{stroke:rgb(149.4437086091, 117.6092715231, 107.5562913906);fill:rgb(149.4437086091, 117.6092715231, 107.5562913906);}#mermaid-8 .label2{fill:hsl(194.4, 16.5562913907%, 49.6078431373%);}#mermaid-8 .arrow2{stroke:hsl(194.4, 16.5562913907%, 49.6078431373%);}#mermaid-8 .branch-label3{fill:#2c2c2c;}#mermaid-8 .commit3{stroke:hsl(23.0769230769, 49.0566037736%, 40.7843137255%);fill:hsl(23.0769230769, 49.0566037736%, 40.7843137255%);}#mermaid-8 .commit-highlight3{stroke:rgb(99.9811320754, 162.7735849057, 202.0188679245);fill:rgb(99.9811320754, 162.7735849057, 202.0188679245);}#mermaid-8 .label3{fill:hsl(23.0769230769, 49.0566037736%, 40.7843137255%);}#mermaid-8 .arrow3{stroke:hsl(23.0769230769, 49.0566037736%, 40.7843137255%);}#mermaid-8 .branch-label4{fill:lightgrey;}#mermaid-8 .commit4{stroke:hsl(0, 83.3333333333%, 43.5294117647%);fill:hsl(0, 83.3333333333%, 43.5294117647%);}#mermaid-8 .commit-highlight4{stroke:rgb(51.5000000001, 236.5, 236.5);fill:rgb(51.5000000001, 236.5, 236.5);}#mermaid-8 .label4{fill:hsl(0, 83.3333333333%, 43.5294117647%);}#mermaid-8 .arrow4{stroke:hsl(0, 83.3333333333%, 43.5294117647%);}#mermaid-8 .branch-label5{fill:lightgrey;}#mermaid-8 .commit5{stroke:hsl(289.1666666667, 100%, 24.1176470588%);fill:hsl(289.1666666667, 100%, 24.1176470588%);}#mermaid-8 .commit-highlight5{stroke:rgb(154.2083333334, 255, 132.0000000001);fill:rgb(154.2083333334, 255, 132.0000000001);}#mermaid-8 .label5{fill:hsl(289.1666666667, 100%, 24.1176470588%);}#mermaid-8 .arrow5{stroke:hsl(289.1666666667, 100%, 24.1176470588%);}#mermaid-8 .branch-label6{fill:lightgrey;}#mermaid-8 .commit6{stroke:hsl(35.1315789474, 98.7012987013%, 40.1960784314%);fill:hsl(35.1315789474, 98.7012987013%, 40.1960784314%);}#mermaid-8 .commit-highlight6{stroke:rgb(51.331168831, 135.1948051946, 253.6688311688);fill:rgb(51.331168831, 135.1948051946, 253.6688311688);}#mermaid-8 .label6{fill:hsl(35.1315789474, 98.7012987013%, 40.1960784314%);}#mermaid-8 .arrow6{stroke:hsl(35.1315789474, 98.7012987013%, 40.1960784314%);}#mermaid-8 .branch-label7{fill:lightgrey;}#mermaid-8 .commit7{stroke:hsl(106.1538461538, 84.4155844156%, 35.0980392157%);fill:hsl(106.1538461538, 84.4155844156%, 35.0980392157%);}#mermaid-8 .commit-highlight7{stroke:rgb(206.1818181817, 89.948051948, 241.051948052);fill:rgb(206.1818181817, 89.948051948, 241.051948052);}#mermaid-8 .label7{fill:hsl(106.1538461538, 84.4155844156%, 35.0980392157%);}#mermaid-8 .arrow7{stroke:hsl(106.1538461538, 84.4155844156%, 35.0980392157%);}#mermaid-8 .branch{stroke-width:1;stroke:lightgrey;stroke-dasharray:2;}#mermaid-8 .commit-label{font-size:10px;fill:rgb(183.8476190475, 181.5523809523, 181.5523809523);}#mermaid-8 .commit-label-bkg{font-size:10px;fill:hsl(180, 1.5873015873%, 28.3529411765%);opacity:0.5;}#mermaid-8 .tag-label{font-size:10px;fill:#e0dfdf;}#mermaid-8 .tag-label-bkg{fill:#1f2020;stroke:#cccccc;}#mermaid-8 .tag-hole{fill:#ccc;}#mermaid-8 .commit-merge{stroke:#1f2020;fill:#1f2020;}#mermaid-8 .commit-reverse{stroke:#1f2020;fill:#1f2020;stroke-width:3;}#mermaid-8 .commit-highlight-inner{stroke:#1f2020;fill:#1f2020;}#mermaid-8 .arrow{stroke-width:8;stroke-linecap:round;fill:none;}#mermaid-8 .gitTitleText{text-anchor:middle;font-size:18px;fill:#ccc;}#mermaid-8 :root{--mermaid-font-family:"trebuchet ms",verdana,arial,sans-serif;}mainfea-networkfix-log0-14645aa1-6fbbbf7V0.12-dfad1603-a9e8c90V0.25-dda00826-f304b18 5.3.4 Bibliothèques (zod, react-hook-form, react-query, react-router, orval, react-testing-library, vitest) 5.3.4.1 Routing Définition Modules Gestion “plug-in” multi-app 5.3.5 Gestion des erreurs et des chargements 5.3.6 Authentification et autorisation Utilisateurs et rôles principes Note Existant HTTP / Headers auth Problème Solutions HTTPS/JWT choix reauthentification flow dashboard lucas 5.3.6.1 RBAC 5.3.7 Application Primitives/Core Routing Navigation Layout Page d’accueil QtReact Empty Previous Next 5.3.8 Modules J’ai étudié l’application Qt pour comprendre son fonctionnement et son architecture. J’ai ensuite identifié les modules qui la composent et les fonctionnalités qu’ils offrent. J’ai regroupé certains modules qui partagent des fonctionnalités similaires. J’ai ensuite créé une liste permettant de prioriser les modules en fonction de leur importance. Cette liste est présentée dans le tableau ci-dessous. Table 4: Ordre de priorité des modules Priorité Module Qt Module React 0 IDENTIFICATION Shell REDEMARRAGE CALCULATEUR Shell DECONNEXION Shell 1 ETHERNET ADRESSES IP Réseau NODES CARTES ETAT DES CARTES Cartes ETAT DES E/S DATE/HEURE ETAT SYNCHRO Date et heure MODE SYNCHRO FUSEAU HORAIRE DATE / HEURE JOURNAL JOURNAL Journal RAZ DU JOURNAL 2 CONFIGURATION IMPORT CONFIGURATION Configuration EXPORT CONFIGURATION MOTS DE PASSE Utilisateurs CERT. EQUIPEMENT Certificats CERTIFICATS CA Autorités de certification 3 LISTE MNEMOS Mnémoniques SYS LOG SYSLOG SNMP SNMP 4 SERVEUR IEC 60870 IEC 61850 CLIENT IEC 61850 SERVEUR MODBUS MODBUS MAITRE MODBUS MODBUS ESCLAVE MQTT MQTT 5 FICHIER Fichier IMPRIMANTES Imprimantes DEBOGAGE Débogage 5.3.8.1 Réseau QtReact Previous Next Previous Next Le premier module sur lequel j’ai travaillé est le module réseau. Ce module permet de configurer les adresses IP des interfaces réseau de l’équipement. Il permet également de configurer les nœuds du réseau. Les nœuds sont des équipements distants qui peuvent être connectés à l’équipement principal. Ce module est essentiel pour la configuration de l’équipement et pour assurer la communication avec les autres équipements du réseau. HSR et PRP TODO : Expliquer HSR et PRP https://fr.belden.com/solutions/high-availability-seamless-redundancy 5.3.8.2 Journal QtReact Previous Next Previous Next 5.3.8.3 Date et heure QtReact Previous Next Previous Next 5.3.8.4 Cartes QtReact Previous Next Previous Next 5.3.8.5 Configuration QtReact Previous Next Previous Next 5.3.8.6 Utilisateurs QtReact Previous Next 5.3.8.7 Certificats QtReact Previous Next 5.3.8.8 Autorités de certification QtReact Empty Previous Next 5.3.8.9 Mnémoniques QtReact Previous Next 5.3.9 Tests 5.3.10 Performances Les performance des deux applications ont été évaluée avec l’outil lighthouse de chrome. C’est un outil automatisé open source permettant de mesurer la qualité des pages Web. Les rapports générés par cet outil permettent d’identifier les problèmes de performance, d’accessibilité et de compatibilité. Rapport application Qt Rapport application React Également, l’outil ne calcule pas correctement les temps de chargement initiaux au démarrage des application WebAssembly (Qt). Le temps de chargement de l’application Qt est d’environ 15 secondes tandis que le temps de chargement de l’application react est quasiment instantané (autour de 1 seconde). Finalement l’outil ne rends pas compte de l’expérience utilisateur. L’application React est plus fluide et plus réactive. L’intégration d’une couche de cache limite aussi le rechargement des données. L’interface de l’application Qt est plus simple mais ne permets pas de naviguer rapidement entre les pages. 5.3.11 Application 5.3.11.1 Sprint 1 Autoformation Développement de l’application Shell et layout Module réseau Recherche d’une suite de tests Intégration avec des bibliothèques pertinentes Routage et navigation : react-router Cache client : react-query 5.3.11.2 Sprint 2 Développement de l’application Module journal Module date et heure Module cartes Refactorisation des modules développés modules 1 à 4 Délégation de la gestion des formulaire Validation des données : zod Gestion des formulaires : react-hook-form Mise en place de la gestion multi-application Mise en place des principes d’authentification et d’autorisation avec la gestion multi-utilisateur Mise en place des tests de composants avec cypress Conformité Non régression 5.3.11.3 Sprint 3 Écriture des tests des modules 1 à 4 Test de déploiement sur cible avec un nouveau serveur web Problématiques liées au développement d’ihm Layout Front/back Authentification Echange de données Technologies Evolution rapide Cest quoi react ? fonctionnement => une seule responsabilité => affichage ecosystème très large, nécessité d’adopter d’autres bibliothèques pour étoffer les fonctionnalités et garantir la qualité du code 6 Conclusion En conclusion, le projet entamé en juin 2023, a permis de développer une application de configuration des automates SDEL plus ergonomique et sécurisée. Les nouvelles fonctionnalités et interfaces ont été conçues en collaboration avec les utilisateurs finaux. Les retours positifs attestent de l’efficacité de cette approche centrée utilisateur. Cette nouvelle application vient étoffer notre offre de services et renforce notre positionnement sur le marché de l’énergie qui nécessite l’intégration de contraintes de cybersécurité de plus en plus importantes au fils des années. La vigilance demeure de mise, et une gestion proactive post-implémentation est recommandée pour assurer une adaptation continue aux évolutions technologiques et aux nouvelles menaces. Ce projet constitue ainsi un jalon important dans notre trajectoire vers l’innovation et la pérennité de nos systèmes. Aujourd’hui, l’application est utilisée quotidiennement par les opérateurs de maintenance de nos clients et les équipes de test en interne. D’un point de vue personnel, ce projet vient clôturer ma formation d’ingénieur. Il marque un terme à mes études et m’a permis de mettre en pratique les connaissances acquises durant ces trois années d’apprentissage. Sur le plan technique et scientifique, le projet m’a permis de consolider mes compétences en informatique. J’ai pu découvrir de nouvelles technologies et approfondir mes connaissances en développement WEB. En termes de montée en compétences, j’ai constaté une nette amélioration de ma compréhension des applications WEB ainsi que les contraintes appliqués au domaine de l’embarqué et des réseaux privés. Du point de vue économique, le chiffrage du projet et la mise en place d’un feuille de route ont constitué une expérience pratique sur le fonctionnement et la nécessité des ces processus en entreprise. Cela m’a sensibilisé aux implications financières des choix effectués en tant que responsable et aiguillé ma vision vers une approche plus stratégique et pragmatique. Sur le plan organisationnel, la gestion du projet m’a confronté à certains défis en termes de coordination, de planification et de suivi des tâches. J’ai pu mettre en pratique les méthodes agiles apprises en cours. J’ai également pu progresser en autonomie. Au début parfois trop distant et au fur et à mesure de l’avancement du projet, j’ai amélioré la transmission d’informations vers les différentes parties du projet. Ces compétences organisationnelles acquises seront indéniablement bénéfiques dans ma future carrière professionnelle. Quant à mon projet professionnel, cette expérience a confirmé ma passion pour le développement logiciel et a éclairé les prochaines étapes de ma carrière. Mon objectif est de m’orienter vers le développement logiciel, en particulier dans le domaine WEB et fullstack. Pour finir, cette expérience a été formatrice et enrichissante à bien des égards. Elle a façonné mon identité professionnelle et renforcé ma détermination à exceller dans le domaine de l’informatique. Je suis prêt à relever de nouveaux défis et à contribuer de manière significative à des projets futurs. 6.1 Bilan Le projet a débuté par la mise en place de l’environnement de travail et l’apprentissage autonome des technologies, avec peu de références internes dans l’entreprise. Environ 600 à 700 heures ont été consacrées au projet, avec une moitié du projet déjà réalisée. La plupart des objectifs des premières phases du projet ont été atteints. L’authentification reste à valider lors du prochain sprint. Des refactorisations ont été nécessaires au fur et à mesure de l’apprentissage. Le projet a donc atteint des jalons significatifs, mais le chemin vers la finalisation nécessite encore des efforts substantiels. Il reste encore environ 600 heures de travail pour finir le développement des modules restant. 7 Annexes 7.1 Références 7.2 Grille des compétences 7.3 Acknowledgments I am grateful for the insightful comments offered by the anonymous peer reviewers at Books & Texts. The generosity and expertise of one and all have improved this study in innumerable ways and saved me from many errors; those that inevitably remain are entirely my own responsibility

      pas fou

    Annotators

    1. Author response:

      The following is the authors’ response to the original reviews.

      Reviewer #1 (Public Reviews):

      Summary:

      This paper by Schommartz and colleagues investigates the neural basis of memory reinstatement as a function of both how recently the memory was formed (recent, remote) and its development (children, young adults). The core question is whether memory consolidation processes as well as the specificity of memory reinstatement differ with development. A number of brain regions showed a greater activation difference for recent vs. remote memories at the long versus shorter delay specifically in adults (cerebellum, parahippocampal gyrus, LOC). A different set showed decreases in the same comparison, but only in children (precuneus, RSC). The authors also used neural pattern similarity analysis to characterize reinstatement, though I have substantive concerns about how this analysis was performed and as such will not summarize the results. Broadly, the behavioural and univariate findings are consistent with the idea that memory consolidation differs between children and adults in important ways, and takes a step towards characterizing how.

      Strengths:

      The topic and goals of this paper are very interesting. As the authors note, there is little work on memory consolidation over development, and as such this will be an important data point in helping us begin to understand these important differences. The sample size is great, particularly given this is an onerous, multi-day experiment; the authors are to be commended for that. The task design is also generally well controlled, for example as the authors include new recently learned pairs during each session.

      Weaknesses:

      As noted above, the pattern similarity analysis for both item and category-level reinstatement was performed in a way that is not interpretable given concerns about temporal autocorrelation within the scanning run. Below, I focus my review on this analytic issue, though I also outline additional concerns.

      We thank the reviewer for both the positive and critical appraisal of our paper.

      (1) The pattern similarity analyses were not done correctly, rendering the results uninterpretable (assuming my understanding of the authors' approach is correct).

      a. First, the scene-specific reinstatement index: The authors have correlated a neural pattern during a fixation cross (delay period) with a neural pattern associated with viewing a scene as their measure of reinstatement. The main issue with this is that these events always occurred back-to-back in time. As such, the two patterns will be similar due simply to the temporal autocorrelation in the BOLD signal. Because of the issues with temporal autocorrelation within the scanning run, it is always recommended to perform such correlations only across different runs. In this case, the authors always correlated patterns extracted from the same run, which moreover have temporal lags that are perfectly confounded with their comparison of interest (i.e., from Fig 4A, the "scene-specific" comparisons will always be back-to-back, having a very short temporal lag; "set-based" comparisons will be dispersed across the run, and therefore have a much higher lag). The authors' within-run correlation approach also yields correlation values that are extremely high - much higher than would be expected if this analysis was done appropriately. The way to fix this would be to restrict the analysis to only cross-run comparisons, but I don't believe this is possible unfortunately given the authors' design; I believe the target (presumably reinstated) scene only appears once during scanning, so there is no separate neural pattern during the presentation of this picture that they can use. For these reasons, any evidence for "significant scene-specific reinstatement" and the like is completely uninterpretable and would need to be removed from the paper.

      We thank the reviewer for this important input. We acknowledge that our study design leads to temporal autocorrelation in the BOLD signal when calculating RSA between fixation and scene time windows. We also recognize that we cannot interpret the significance of scene-specific reinstatement compared to zero and have accordingly removed this information. Nevertheless, our primary objective was to investigate changes in scene-specific reinstatement in relation to the different time delays of retrieval. Given that the retrieval procedure is the same over time and presumably similarly influenced by temporal autocorrelations, we argue that our results must be attributed to the relative differences in reinstatement across recent and remote trials. Bearing this in mind, we argue that our results can be interpreted in terms of delay-related changes in reinstatement. This information is discussed in pp. 21, 40 of the manuscript.

      We agree with the reviewer that cross-run comparisons would be extremely interesting. This could be achieved by introducing the same items repeatedly across different runs, which was not possible in our current setup since we were interested in single exposure retrieval and practical time restriction in scanning children. We have  introduced this idea in Limitations and Discussion sections (pp. 40, 44) of the manuscript to inform future studies.

      Finally, thanks to the reviewer’s comment, we identified a bug in the final steps of our RSA calculation. Fischer’s z-transformation was incorrectly applied to r-1 values, resulting in abnormally high values. We apologize for this error. We have revised the scripts and rectified the bug by correctly applying Fischer’s z-transformation to the r similarity values. We also adjusted the methods description figure accordingly (Figure 5, p. 22). This adjustment led to slightly altered reinstatement indices. Nevertheless, the overall pattern of delay-related attenuation in the scene-specific reinstatement index, observed in both children and adults, remains consistent. Similarly, we observed gist-like reinstatement uniquely in children.

      b. From a theoretical standpoint, I believe the way this analysis was performed considering the fixation and the immediately following scene also means that the differences between recent and remote could have to do with either the reactivation (processes happening during the fixation, presumably) or differences in the processing of the stimulus itself (happening during the scene presentation). For example, people might be more engaged with the more novel scenes (recent) and therefore process those scenes more; such a difference would be interpreted in this analysis as having to do with reinstatement, but in fact could be just related to the differential scene processing/recognition, etc.

      Thank you for your insightful comments. We acknowledge the theoretical concerns raised about distinguishing between the effects of reactivation processes occurring during fixation and differential processing of the stimulus itself during scene presentation. Specifically, the notion that engagement levels with recent scenes could result in enhanced processing, which might be misattributed to memory reinstatement mechanisms.

      We argue, however, that during scene presentation, scenes are processed more “memory-wise” rather than “perception-wise”, since both recent and remote memories are well-learned, as we included only correctly recalled memories in the analysis.

      We concur that scene presentations entail perceptual processing; however, such processing would be consistent across all items, given that they were presented with the same repeated learning procedure, rendering them equally familiar to participants. In addition, we would argue that distinct activation patterns elicited during varying delays are more likely attributable to memory-related processing, since participants actively engaged in a memory-based decision-making task during these intervals. We have incorporated this rationale into the discussion section of our manuscript (p. 40).

      With this in mind, we hypothesized that in case of “memory-wise” processing, the neural engagement during the scene time window should be higher for remote compared to recent  items, and this increases with passing time as more control and effort should be exhibited during retrieval due to reorganized and distributed nature of memories. If the scenes are processed more “perception-wise”, we would expect higher neural engagement during the retrieval of recent compared to remote items. Our exploratory analysis (detailed overview in supplementary materials, Figure S3, Table S9) revealed a higher neural activation for remote compared to recent items in medial temporal, prefrontal, occipital and cerebellar brain regions, supporting the notion of “memory-wise” processes during scene time window. However, this exploratory analysis cannot provide a direct solution to the reviewer’s concern as our paradigm per se cannot arbitrate between “memory-wise” and “perception-wise” nature of retrieval. We added the point to the discussion (see p. 40).

      c. For the category-based neural reinstatement:

      (1) This suffers from the same issue of correlations being performed within the run. Again, to correct this the authors would need to restrict comparisons to only across runs (i.e., patterns from run 1 correlated with patterns for run 2 and so on). With this restriction, it may or may not be possible to perform this analysis, depending upon how the same-category scenes are distributed across runs. However, there are other issues with this analysis, as well.

      (2) This analysis uses a different approach of comparing fixations to one another, rather than fixations to scenes. The authors do not motivate the reason for this switch. Please provide reasoning as to why fixation-fixation is more appropriate than fixation-scene similarity for category-level reinstatement, particularly given the opposite was used for item-level reinstatement. Even if the analyses were done properly, it would remain hard to compare them given this difference in approach.

      (3) I believe the fixation cross with itself is included in the "within category" score  Is this not a single neural pattern correlated with itself, which will yield maximal similarity (pearson r=1) or minimal dissimilarity (1-pearson r=0)? Including these comparisons in the averages for the within-category score will inflate the difference between the "within-category" and "between-category" comparisons. These (e.g., forest1-forest1) should not be included in the within-category comparisons considered; rather, they should be excluded, so the fixations are always different but sometimes the comparisons are two retrievals of the same scene type (forest1-forest2), and other times different scene types (forest1-field1)

      (4) It is troubling that the results from the category reinstatement metric do not seem to conceptually align with past work; for example, a lot of work has shown category-level reinstatement in adults. Here the authors do not show any category-level reinstatement in adults (yet they do in children), which generally seems extremely unexpected given past work and I would guess has to do with the operationalization of the metric.

      Thank you for this important input regarding category-based reinstatement.

      (1) The distribution of within-category items across runs was approximately similar and balanced. Additionally, within runs, they were presented randomly without close temporal proximity. Based on this arrangement, we believe that the issue of close temporal autocorrelation, as pointed out by the reviewer in the context of scene-specific reinstatement, may not apply to the same extent here. Again, our focus is not on the absolute level of category-based reinstatement, but the relative difference across conditions (recent vs. remote short delay vs. remote long delay) which are equally impacted by the autocorrelations.  

      (2) We apologize for not motivating this analysis further. Whereas the scene-reinstatement index (i.e., fixation to scene correlation) gives us a measure of the pre-activation of a concrete scene (e.g., a yellow forest in autumn), the gist-like reinstatement gives us a measure of the pre-activation of a whole category of scenes (e.g., forests). Critically, our window of interest is the fixation period for both sets of analysis (in the absence of any significant visual input). The scene-specific reinstatement uses the scene window as a neural template against which the fixation period can be compared, while the gist-like reinstatement compares similarity of reactivation pattern for trials from the same category but differ in the exact memory content. The reinstatement of more generic, gist-like memory (e.g., forest) across multiple trials should yield more similar neural activation patterns. Significant gist-like reinstatement would suggest that neural patterns for scenes within the same category are more generic, as indicated by higher similarity among them. On the other hand, a more detailed reinstatement of specific types of forests (e.g., a yellow forest in autumn, green pine trees, a bare-leaved forest in spring, etc.) that differ in various dimensions could result in neural activation patterns that are as dissimilar as those seen in the reinstatement of scenes from entirely different categories. Through this methodology, we could distinguish between more generic, gist-like reinstatement and more specific, detailed reinstatement. This is now clarified in the manuscript, see p. 25.

      (3) We apologize for the confusion caused by the figure and analysis description. In our analysis, we indeed excluded the correlation of the fixation cross with itself. Consequently, the diagonal in the figure should be blank to indicate this. This is now revised in the manuscript (Figure 7B and in Methods).

      (4) We appreciate your concern and recognize that the terminology we used might not align perfectly with the conventional understanding of category-based reinstatement. Typically, category-level neural representations (as discussed in Polyn et al., 2005; Jafarpour et al., 2014; among others) are investigated to identify specific brain areas associated with encoding/perception of scenes or faces. Our aim, however, was to explore the mnemonic reinstatement of highly detailed scenes that were elaborately encoded, with the hypothesis that substantial representational transformations would occur over time and vary with age. This hypothesis is based on the memory literature, including the Fuzzy-Trace Theory, the Contextual Binding Theory, and the Trace Transformation Theory (Brainerd & Reyna, 1998; Yonelinas, 2019; Moscovitch & Gilboa, 2023). Therefore, we renamed 'category-based' reinstatement to 'gist-like' reinstatement, which clarifies our concept and better aligns it with existing literature.

      We anticipated that young adults, having the ability to retain detailed narratives post-encoding, would demonstrate a reinstatement of scenes with distinct details, making these scenes dissimilar from each other (see similar findings in Sommer et al., 2021). In contrast, given the anticipated lesser strategic elaboration during learning in children, we hypothesized that they would demonstrate a shallower, more gist-like reinstatement (for instance, children recalling a forest or a field in a general sense without specific details or vivid imagery). This could result in higher category-based similarity, as children might reinstate a more generic forest concept.

      We did not gather additional data on the verbal quality of reinstatement due to the limited scanning time available for children, so these assumptions remain unverified. However, anecdotal observations post-retrieval indicated that adults often reported very vivid scenes associated with clear narrative recall. In contrast, children frequently described more vague memories (e.g., “I know it was a forest”) without specific details. Future studies should include measures to assess the quality of reinstatement, potentially outside the scanning environment.

      (2) I did not see any compelling statistical evidence for the claim of less robust consolidation in children.

      Specifically in terms of the behavioral results of retention of the remote items at 1 vs 14 days, shown in Figure 2B, the authors conclude that memory consolidation is less robust in children (line 246). Yet they do not report statistical evidence for this point, as there was no interaction of this effect with the age group. Children had worse memory than adults overall (in terms of a main effect - i.e. across recent and remote items). If it were consolidation-specific, one would expect that the age differences are bigger for the remote items, and perhaps even most exaggerated for the 14-day-old memories. Yet this does not appear to be the case based on the data the authors report. Therefore, the behavioral differences in retention do not seem to be consolidation specific, and therefore might have more to do with differences in encoding fidelity or retrieval processes more generally across the groups. This should be considered when interpreting the findings.

      Thank you for highlighting this important issue. We acknowledge that our initial description and depiction of our behavioral findings may not have effectively conveyed the main message about memory consolidation. Therefore, we have revised the behavioral results section (see pp. 12-14) to communicate our message more clearly.

      As detailed in the methods section, we reported retention rates only for those items that were correctly (100%) learned on day 0, day 1, and day 14. This approach meant that different participants had varying numbers of items learned correctly. However, this strategy allowed us to address our primary question: whether memory consolidation, based on all items initially encoded successfully, is comparably robust between groups.

      To illustrate the change in retention rate slopes over time for recently learned items (i.e., immediately 30 minutes after learning), short delay remote, and long delay remote items, relative to the initially correctly learned items more clearly and straightforward, we conducted the following analysis: after observing no differences between sessions in both age groups for recent items on days 1 and 14, we combined the recent items. This approach enabled us to investigate how the slope of memory retention for initially correctly learned items (with a baseline of 100%) changes over time. We observed a significant interaction between item type (recent, short delay remote, long delay remote) and group (F(3,250) = 17.35, p < .001, w2 = .16). The follow up of this interaction revealed significantly less robust memory consolidation across all delay times in children compared to young adults. This information is added in the manuscript in pp. 12-14. We have also updated the figures, incorporating the baseline of 100% correct performance.

      (3) Please clarify which analyses were restricted to correct retrievals only. The univariate analyses states that correct and incorrect trials were modelled separately but does not say which were considered in the main contrast (I assume correct only?). The item specific reinstatement analysis states that only correct trials were considered, but the category-level reinstatement analysis does not say. Please include this detail.

      Thank you for bringing this to our attention. We indeed limited our analysis – including univariate, specific reinstatement, and gist-like analyses – to only correctly remembered items. This decision was made because our goal was to observe delay-related changes in the neural correlates of correct memories, which are potentially stronger. We have incorporated this information into the manuscript.

      (4) To what extent could performance differences be impacting the differences observed across age groups? I think (see prior comment) that the analyses were probably limited to correct trials, which is helpful, but still yields pretty big differences across groups in terms of the amount of data going into each analysis. In general, children showed more attenuated neural effects (e.g., recent/remote or session effects); could this be explained by their weaker memory? Specifically, if only correct trials are considered that means that fewer trials would be going into the analysis for kids, especially for the 14-day remote memories, and perhaps pushing the remove > recent difference for this condition towards 0. The authors might be able to address this analytically; for example, does the remote > recent difference in the univariate data at day 14 correlate with day 14 memory?

      Thank you for pointing this out. Indeed, there was a significant relationship between remote > recent difference in the univariate data and memory performance at day 14 across both age group (see Figure 4C-D). The performance of all participants including children was above chance level for remote trial on day 14. In addition, although number of remote trials was lower in children (18 trials on average) in comparison to adults (22 trials on average), we believe that the number of remote trials was not too low or different across groups for the contrast.

      (5) Some of the univariate results reporting is a bit strange, as they are relying upon differences between retrieval of 1- vs. 14-day memories in terms of the recent vs. report difference, and yet don't report whether the regions are differently active for recent and remote retrieval. For example, in Figure 3A, neither anterior nor posterior hippocampus seem to be differentially active for recent vs. remote memories for either age group (i.e., all data is around 0). This difference from zero or lack thereof seems important to the message - is that correct? If so, can the authors incorporate descriptions of these findings?

      Thank you for this valuable input. When examining recent and remote retrieval separately, indeed both the anterior and posterior regions of the hippocampus exhibited significant activation from zero in adults (all p < .0003FDRcorr) and children (all p < .014FDRcorr, except for recent posterior hippocampus) during all delays. We include this information in the manuscript (see p. 17) and add it to the supplementary materials (Figure S2, Table S7).

      (6) Please provide more details about the choices available for locations in the 3AFC task. (1) Were they different each time, or always the same? If they are always the same, could this be a motor or stimulus/response learning task? (2) Do the options in the 3AFC always come from the same area - in which case the participant is given a clue as to the gist of the location/memory? Or are they sometimes randomly scattered across the image (in which case gist memory, like at a delay, would be sufficient for picking the right option)? Please clarify these points and discuss the logic/impact of these choices on the interpretation of the results.Response: Thank you for pointing this out. During learning and retrieval, we employed the 3AFC (Three-Alternative Forced Choice) task.

      The choices for locations varied across scenes while remained the same across time within individuals. There were 18 different key locations for the objects, distributed across the stimulus set. This means the locations of the objects were quite heterogeneous and differed between objects. The location of the object within the task was presented once during encoding and remained consistent throughout learning. Given the location heterogeneity, we believe our task cannot be reduced to a mere “stimulus/response learning task” but is more accurately described as an object-location associations task.

      Similar to the previous description, the options for the 3AFC task did not originate from the same area, as there were 18 different areas in total. The three choice options were distributed equally: so sometimes the “correct” answer was the left option, sometimes in the middle option, or sometimes the right option. Therefore, we believe that the 3AFC task did not provide clues to the location but required detailed and precise memory of the location. Moreover, the options were not randomly scattered but rather presented close together in the scene, demanding a high level of differentiation between choices.

      Taking all the above into consideration, we assert that precise object-location associative memory is necessary for a correct answer. We have added this information to the manuscript (p. 9).

      (7) Often p values are provided but test statistics, effect sizes, etc. are not - please include this information. It is at times hard to tell whether the authors are reporting main effects, interactions, pairwise comparisons, etc.

      Thank you for bringing this to our attention. We realize that including this information in the Tables may not be the most straightforward approach. Therefore, we have incorporated the test statistics, effect sizes, and related details into the text of the results section for clarity.

      (8) There are not enough methodological details in the main paper to make sense of the results. For example, it is not clear from reading the text that there are new object-location pairs learned each day.

      Thank you for pointing this out. We have added this information to the main manuscript. Additionally, we have emphasized this information in the text referring to Figure 1B.

      (9) The retrieval task does not seem to require retrieval of the scene itself, and as such it would be helpful for the authors to both explain their reasoning for this task to measure reinstatement. Strictly speaking, participants could just remember the location of the object on the screen. Was it verified that children and adults were recalling the actual scene rather than just the location (e.g. via self-report)? It's possible that there may be developmental differences in the tendency to reinstate the scene depending on e.g., their strategy.

      Thank you for highlighting this important point. Indeed, the retrieval task included explicit instructions for participants to recall and visualize the scene associated with the object presented during the fixation time window. Participants were also instructed to recollect the location of the object within the scene. Since the location was contextually bound to the scene and each object had a unique location in each scene, the location of the object was always embedded in the specific scene context. We have added this information to both the Methods and Results sections.

      From the self-reports of the participants (which unfortunately were not systematically collected on all occasions), they indicated that when they could recall the scene and the location due to the memory of stories created during strategic encoding, it aided their memory for the scene and location immensely. We also concur with your observation that children and young adults may differ in their ability to reinstate scenes, depending on the success of their employed recall strategies. This task was conducted with an awareness of potential developmental differences in the ability to form complex contextual memories. Our elaborative learning procedure was designed to minimize these differences. It is important to note though we did not expect children to achieve performance levels fully comparable to adults. There may indeed be developmental differences in reinstatement, such as due to differences in knowledge availability and accessibility (Brod, Werkle-Bergner, & Shing, 2013). We think that these differences may underlie our findings of neural reinstatement. This is now discussed in p. 34-35, 39-43 of the manuscript.

      (10) In general I found the Introduction a bit difficult to follow. Below are a few specific questions I had.

      a. At points findings are presented but the broader picture or take-home point is not expressed directly. For example, lines 112-127, these findings can all be conceptualized within many theories of consolidation, and yet those overarching frameworks are not directly discussed (e.g., that memory traces go from being more reliant on the hippocampus to more on the neocortex). Making these connections directly would likely be helpful for many readers.

      Thank you for bringing this to our attention. We have incorporated a summary of the general frameworks of memory consolidation into the introduction. This addition outlines how our summarized findings, particularly those related to memory consolidation for repeatedly learned information, align with these frameworks (see lines 126-138, 146-150).

      b. Lines 143-153 - The comparison of the Tompary & Davachi (2017) paper with the Oedekoven et al. (2017) reads like the two analyses are directly comparable, but the authors were looking at different things. The Tompary paper is looking at organization (not reinstatement); while the Oedekoven et al. paper is measuring reinstatement (not organization). The authors should clarify how to reconcile these findings.

      Thank you for highlighting this aspect. We have revised how we present the results from Tompary & Davachi (2017). This study examined memory reorganization for memories both with and without overlapping features, and it observed higher neural similarity for memories with overlapping features over time. The authors also explored item-specific reinstatement for recent and remote memories by assessing encoding-retrieval similarity. Since Oedekoven et al. (2017) utilized a similar approach, their results are comparable in terms of reinstatement. We have updated and expanded our manuscript to clarify the parallels between these studies (see lines 157-162).

      c. Line 195-6: I was confused by the prediction of "stable involvement of HC over time" given the work reviewed in the Introduction that HC contribution to memory tends to decrease with consolidation. Please clarify or rephrase.

      Drawing on the Contextual Binding Theory (Yonelinas et al., 2019), as well as the Multiple Trace Theory (Nadel et al., 2000) and supported for instance by evidence from Sekeres et al. (2018), we hypothesized that detailed contextual memories formed through repeated and strategic learning would strengthen the specificity of these memories, resulting in consistent hippocampal involvement for successfully recalled contextualized detailed memories. We have included additional explanatory information in the manuscript to clarify this hypothesis (see lines 217-219).

      d. Lines 200-202: I was a bit confused about this prediction. Firstly, please clarify whether immediate reinstatement has been characterized in this way for kids versus adults. Secondly, don't adults retain gist more over long delays (with specific information getting lost), at least behaviourally? This prediction seems to go against that; please clarify.

      Thank you for raising this important point. Indeed, there are no prior studies that examined memory reinstatement over extended durations in children. The primary existing evidence suggests that neural specificity or patterns of neural representations in children can be robustly observed, while neural selectivity or univariate activation in response to the same stimuli tends to mature later (i.e., Fandakova et al., 2019). Bearing this in mind and recognizing that such neural patterns can be observed in both children and adults, we hypothesized that adults may form stronger detailed contextual memories compared to children. By employing strategies such as creating stories, adults might more easily recall scenes without the need to resort to forming generic or gist-like memories (for example, 'a red fox was near the second left pine tree in a spring green forest'). This assumption aligns with the Fuzzy Trace Theory (Reyna & Brainerd, 1995), which posits that verbatim memories can be created without the extraction of a gist.

      Conversely, we hypothesized that children, due to the ongoing maturation of associative and strategic memory components (as discussed in Shing et al., 2008 and 2010), which are dependent respectively on the hippocampus (HC) and the prefrontal cortex (PFC), would be less adept at creating, retaining, and extracting stories to aid their retrieval process. This could result in them remembering more generic integrated information, like the relationship between a fox and some generic image of a forest. We have added explanatory information to the manuscript to elucidate these points (see lines 225-230).

      Reviewer #1 (Recommendations For The Authors):

      (1) For Figure 3, I would highly recommend changing the aesthetics for the univariate data - at least on my screen they appear to be open boxes with solid vs. dashed lines, and as such look identical to the recent vs. remove distinction in Figure 2B. It also doesn't match the legend for me, which shows the age groups having purple vs. yellow coloring.

      Thank you for this observation. We have adjusted Figure 2 (now Figure 3) (please refer to p. 14) accordingly, now utilizing purple and yellow colors to distinguish between the age groups.

      (2) Lines 329-330, it is not true that "all" indices were significant from zero but this is only apparent if you read the next sentence. Please rephrase to clarify. e.g., "All ... indices with a few exceptions ... were significantly..."?

      Based on the above suggestions and considering our primary focus on time-related changes in scene-specific reinstatement, we will refrain from further interpreting the relative expression of individual scene-specific indices against 0. Consequently, we have removed this information from our analysis.

      (3) It is challenging to interpret some of the significance markers, such as those in Figure 3. For example what effects are being denoted by the asterisks and bars above vs. below the data on panel D? Please clarify and/or note in the legend.

      We have included a note in the legend to clarify the meaning of all significance markers. In addition, we decided to state any significant main and interaction effects in the figure rather that to use significance markers.

      (4) For Figures 2 and 3, only the meaning of error bars is described in the caption. It is not explained in the caption what the boxes, lines, and points denote. Please clarify.

      Thank you for highlighting this. We have added explanations to the figure's annotation for clarity. Please note, that considering other review’s suggestions figure plots may have been adjusted or changed, resulting in adjustment of the explanations in the figure annotation.

      (5) How were recent and remote interspersed relative to one another? The text says that each run had 10 recent and 10 remote pairs, presented in a "pseudo-random order" - not clear what that (pseudo) means in this case. Please clarify.

      Thank you for raising this point. We provide this information in the Methods section “Materials and Procedure”: 'The jitters and the order of presentation for recent and remote items were determined using OptimizeXGUI (Spunt, 2016), following an exponential distribution (Dale, 1999). Ten unique recently learned pairs (from the same testing day) and ten unique remotely learned items (from Day 0) were distributed within each run (in total three runs) in the order as suggested by the software as the most optimal. There were three runs with unique sets of stimuli each resulting in thirty unique recent and thirty unique remote stimuli overall.'

      (6) Figure 1A, second to last screen on the learning cycles row - what would be presented to participants here, one of these three emojis? What does the sleepy face represent? I see some of these points were mentioned in the methods, but additional clarification in the caption would be helpful.

      Thank you for highlighting this. We have included this information in the figure caption. Specifically, the sleepy face symbol in the figure denotes a 'missed response'.

      (7) Not clear how the jittered fixation time between object presentation and scene test is dealt with in representational similarity analyses.

      Thank you for pointing this out. Beta estimates were obtained from a Least Square Separate (LSS) regression model. Each event was modeled with their respective onset and duration and, as such, one beta value was estimated per event (with the lags between events differing from trial to trial). We have edited the corresponding section (see p. 53).  

      (8) It was a little bit strange to have used anterior vs posterior HPC ROIs separately in univariate analysis but then combined them for multivariate. There are many empirical and theoretical motivations for looking at item-specific and category reinstatement in anterior and posterior HPC separately, so I was surprised not to see this. Please explain this reasoning.

      Thank you for pointing this out. We agree with the reviewer and included the anterior and posterior HC ROIs into the multivariate analysis. Please see the revised results section (pp. 13-15).

      (9) The term "neural specificity" is introduced (line 164) without explanation; please clarify.

      Thank you for bringing this to our attention. The term ‘neural specificity’ refers to the neural representational distinctiveness of information. In other words, ‘neural specificity,’ as defined by Fandakova et al. (2019), refers to the distinctiveness of neural representations in the regions that process that sensory input. We decided, however to refrain from using this term and instead to use neural representational distinctiveness, which is more self-explaining and was also introduced in the manuscript.

      (10) Age range is specified as 5-7 years initially (line 187) and then 6-7 years (line 188).

      We have corrected the age range in line 188 to '5 to 7 years.'

      Reviewer #2 (Public Reviews):

      Schommartz et al. present a manuscript characterizing neural signatures of reinstatement during cued retrieval of middle-aged children compared to adults. The authors utilize a paradigm where participants learn the spatial location of semantically related item-scene memoranda which they retrieve after short or long delays. The paradigm is especially strong as the authors include novel memoranda at each delayed time point to make comparisons across new and old learning. In brief, the authors find that children show more forgetting than adults, and adults show greater engagement of cortical networks after longer delays as well as stronger item-specific reinstatement. Interestingly, children show more category-based reinstatement, however, evidence supports that this marker may be maladaptive for retrieving episodic details. The question is extremely timely both given the boom in neurocognitive research on the neural development of memory, and the dearth of research on consolidation in this age group. Also, the results provide novel insights into why consolidation processes may be disrupted in children. Despite these strengths, there are quite a few important design and analytical choices that derail my enthusiasm for the paper. If the authors could address these concerns, this manuscript would provide a solid foundation to better understand memory consolidation in children.

      We thank the reviewer for both the positive and critical appraisal of our paper.

      Reviewer #2 (Recommendations For The Authors):

      (1) My greatest concern is the difference in memory accuracy that emerges as soon as immediate learning, which undermines the interpretation of any consolidation-related differences. This concern is two-fold. The authors utilize an adaptive learning approach in which participants learn to criteria or stop after 4 repetitions. This type of approach leads to children seeing the stimuli more often during learning compared to adults, which on its own could have consequences for consolidation-related neural markers. Specifically, within adults theoretical and empirical work this shows that repeating information can actually lead to more gist-like representations, which is the exact profile the children are showing. While there could be a strength to this approach because it allows for equivocal memory, the decision to stop repetitions before criteria means that memory performance is significantly lower in the children, which again could have consequences to consolidation-related neural markers. First, the authors do not show any of the learning-related data which would be critical to assess the impact of this design choice. Second, there are likely differences in memory strength at the delay, making it extremely difficult to determine if the neural markers reflect development, worse memory strength, or both. This issue is compounded by the use of a 3-AFC paradigm, wherein "correct responses" included in the analysis could contain a significant amount of guessing responses. I think a partial solution to this problem is to analyze the RT data and include them in the analyses or use a drift-diffusion modeling approach to get more precise estimates of memory strength to control for this feature. An alternative is to sub-select participants in each group to have a sample matched on performance (including # of repetitions) and re-run all the analyses in this sub-sample. Without addressing these concerns it is near impossible to interpret the presented data.

      Thank you for highlighting this point.

      Firstly, we believe that our approach, involving strategic and repeated learning coupled with feedback, enhances the formation of detailed contextual memories. The retrieval procedure also emphasized the need for detailed memory for location. These are critical differences in experimental procedure from previous studies, which enhanced the importance of detailed representations and likely reduced the likelihood of forming gist-like memories.

      Indeed, we ceased further learning after the fourth repetition. Extensive piloting, where we initially stopped after the seventh repetition, showed no improvement beyond the fourth repetition. In fact, performance tended to decline due to fatigue. Therefore, we limited the number of repetition cycles to the point where an improvement of performance was still feasible. Even though children exhibited lower final learning performance overall, we believe our procedure facilitated them to reach their maximal performance within the experimental setup.

      To address the reviewer’s concern, we included learning data to illustrate the progression of learning (see Fig. 1C, pp. 9-10 in Results).

      When interpreting the retention rates, it is important to note that we reported retention rates only for items that were correctly learned (100%) on day 0, day 1, and day 14. This approach meant that different participants had varying numbers of items learned correctly. However, this method enabled us to address our primary question: whether memory consolidation, based on all items initially encoded successfully, is comparably robust between the groups. To simultaneously examine the change in retention rate slopes over time for recent (30 minutes after learning), short delay (one night after) remote, and long delay (two weeks after) remote items, we conducted a separate analysis of retention rates for recent items on days 1 and 14. After observing no differences between sessions in both age groups, we combined the data for recent items. This allowed us to investigate how the slope of memory retention for initially correctly learned items (with a baseline of 100%) changes over time. We observed a significant interaction between item type (recent, short delay remote, long delay remote) and group. Analysis of this interaction revealed significantly less robust memory consolidation across all delay times for children compared to young adults. The figures have been adjusted accordingly to incorporate the baseline of 100% correct performance.

      Following your suggestion, we also employed the drift diffusion model approach to characterize memory strength, calculating drift rate, boundary and non-decision time parameters. We added the results to the Supplementary Materials (section S2.1, Figure S1).

      Generally, our findings indicate lower overall drift rate in children when considering all items that had to be learned. We also observed that adults show higher slope of decline in drift rate in short and long delay, which, however, are characterized still by higher memory strength compared to children. Both age groups required similar amount of evidence to make decision, which declined with delay. It may indicate an adaptation of weaker memory. Further, we observed lesser non-decision time in children compared to adults, potentially suggesting less error checking or less thorough processing and memory access through strategy in children.

      Overall, these results indicate weaker memory strength in children as a quantitative measure. It may nevertheless stem from qualitatively different memory representations that children form, as our RSA findings suggest. We believe that our neural effect reflects the effect of interest (i.e., worse memory due to lower memory strength in children). When controlled for, it will take away variance of interest in the neural data. Therefore, we will refrain from including memory strength into the model. However, we will include mean RT as the indicator of general response tendencies.

      Given that the paper is already very complex and long, we opted to add the diffusion model results to the Supplementary Materials (section S2.1, Fig. S1), while discussing the results in the discussion (p. 35).

      (2) More discussion of the behavioral task should be included in the results, in particular the nature of the adaptive learning paradigm including the behavioral results as well as the categorical nature of the memoranda. Without this information, it is difficult for the reader to understand what category-level versus item-level reinstatement reflects.

      Thank you for this valuable input. We have incorporated this information into the results section. Please refer to pp. 9-10, 12, 14, 21, 25-26 for the added details.

      (3) Some of the methods for the reinstatement analysis were unclear to me or warranted further adjustment. I believe the authors compared the scene against all other scenes. I believe it would be more appropriate to only compare this against scenes drawn from the same category as opposed to all scenes. Secondly, from my reading, it seems like the reinstatement was done during the scene presentation, rather than the object presentation in which they would retrieve the scene. I believe the reinstatement results would be much stronger if it was captured during the object presentation rather than the re-presentation of the scene. Or perhaps both sets of analyses should be included.

      We apologize for the confusion regarding the analysis method.

      During the review process we have improved the description of this analysis and hope it is easier to follow now. In short, we used both approaches (within and between categories) to suit different goals (I.e., measuring scene-reinstatement and gist-like reinstatement).

      Both types of reinstatement were assessed during the fixation cross to avoid confounds with the object itself being on the screen. We only used the scene window in one analysis (scene-reinstatement index) as a neural template to track its pre-activation during the fixation. So, as the reviewer suggests, our rationale is that the reinstatement indeed starts taking place at the short object presentation window, but importantly, extends to the fixation window. We added this clarifying information to the results section (see p. 21-27).

      (4) For the univariate results, it was unclear to me when reading the results whether they were focusing on the object presentation portion of the trial or the scene presentation portion of the trial. Again, I think the claims of reinstatement related activity would be stronger if they accounted for the object presentation period.

      Thank you for pointing this out. Indeed, the univariate results were based on the object presentation time window. We added this information to the results section (Fig. 3, pp. 14, 16).

      (5) Further, given the univariate differences shown across age groups, the authors should re-run all analyses for the RSA controlling for mean activation within the ROI.

      Thank you for highlighting this. We re-ran all analysis for the RSA controlling for the mean activation within the ROI. The results remained unchanged. We have added this information to the results section as well as in Table S8 and S11 in the Supplementary Materials for further details.

      (6) The authors should include explicit tests across groups for their brain-behavior analyses if they want to make any developmentally relevant interpretations of the data. Also, It would be helpful to include similar analyses to those using the univariate signals, and not just the RSA results.

      Following reviewer’s suggestion, we included brain-behavior analyses for univariate data as well as RSA data with explicit tests across groups. These can be found in the Results Section pp. 18-20, 28-32. Due to the interdependence of predefined ROIs and to avoid running a high number of correlation tests, we employed the partial least square correlation analysis for this purpose. This approach focuses on multivariate links between specified Regions of Interest (ROIs) and fluctuations in memory performance over short and long delays across different age cohorts. We argue that this multivariate strategy offers a more comprehensive understanding of the relationships between brain metrics across various ROIs and memory performance, given their mutual dependence and connectivity (refer to Genon et al. (2022) for similar discussions).

      (7) There could be dramatic differences in memory processing across 5-7 year olds. I know the sample is a little small for this, but I would like to see regressions done within the middle childhood group in addition to the across-group comparisons.

      We have included information detailing the relationship between memory retention rate and age within the child group (refer to p. 13). In the child group, both recent and short delay remote memory improved with age. However, the retention rate for long-delayed memory did not show a significant improvement with increasing age in children.

      (8) I am concerned that the authors used global-signal as a regressor in their first-level analyses, given that there could be large changes in the amount of univariate activation that occurs across groups. This approach can lead to false positives and negatives that obscure localized differences. The authors should remove this term, and perhaps use the mean sum of the white matter or CSF to achieve the noise regressor they wanted to include.

      We understand the reviewers' concerns. However, we believe that our approach is recommended for the pediatric population. Specifically, Graff et al., 2021, found that global signal regression is a highly efficacious denoising technique in their study of 4 to 8-year-old children. This technique was previously suggested for adults by Ciric et al., 2017, and the benefits in terms of motion and physiological noise removal outweigh the potential costs of removing some signal of interest, as indicated by Behzadi et al., 2007. Additionally, we incorporated the six anatomic component-based noise correction (CompCor) to account for WM and CSF signals, as recommended in the pediatric literature.

      (9) The authors discuss the relationship between hippocampal reactivation and worse memory through the lens of Schapiro et al., but a new paper by Tanriverdi et al came out in JOCN recently that is more similar to the authors' findings.

      Thank you for highlighting the recent paper by Tanriverdi et al. in JOCN, which aligns closely with our findings. We appreciate the suggestion and agree that exploring this alignment could further enrich our discussion on the relationship between hippocampal reactivation and memory retention. We incorporated this work in our revised manuscript .

      Minor Comments

      - I was surprised that the authors did not see any differences in univariate signals for memory retrieval as a function of development, as much of the prior work has shown differences (for example work by Tracy Riggins). I believe this contrast should be highlighted in the discussion.

      - Given the robust differences in sleep patterns across childhood and the role of sleep in systems consolidation framework, I think this feature should be highlighted in either the introduction or discussion.

      - Could the authors report on differences (or lack of differences) in head motion across the groups, and if they are different whether they could include them as a confounding variable.

      I believe we included six motion parameters and their derivatives into the model

      Thank you for your comments.

      First, prior works on univariate signals of memory retrieval focused mostly on remembered vs forgotten contrasts, while in our study we focused on remote vs recent in short and long delay only for correctly remembered items. This can partially explain the results. We highlighted this information in the discussion session.

      Second, we agree with the reviewer that sleep patterns across childhood should be addressed in the analysis. Therefore, we incorporated them in the discussion section.

      Third, indeed head motion were included in the analysis as confounding variables, as adding them is highly recommended for the developmental population (e.g., Graff et al. 2021). As an example, we observed higher framewise displacement in children compared to adults, t = -16(218), p <. 001, as well as in translational y, t = -2.33(288), p = .02.

      Reviewer #3 (Public Reviews):

      Summary:

      This study aimed to understand the neural correlates of memory recall over short (1-day) and long (14-days) intervals in children (5-7 years old) relative to young adults. The results show that children recall less than young adults and that this is accompanied by less activation (relative to young adults) in brain networks associated with memory retrieval.

      Strengths:

      This paper is one of few investigating long-term memory (multiple days) in a developmental population, an important gap in the field. Also, the authors apply a representational similarity analysis to understand how specific memories evolve over time. This analysis shows how the specificity of memories decreases over time in children relative to adults. This is an interesting finding.

      We thank the reviewer for the appraisal of our manuscript.

      Weaknesses:

      Overall, these results are consistent with what we already know: recall is worse in children relative to adults (e.g., Cycowicz et al., 2001) and children activate memory retrieval networks to a lesser extent than adults (Bauer et al, 2017).

      It seems that the reduced activation in memory recall networks is likely associated with less depth of memory encoding in children due to inattentiveness, reduced motivation, and documented differences in memory strategies. In regard to this, there was consideration of IQ, sex, and handedness but these were not included as covariates as they were not significant although I note p<.16 suggests there was some level of association nonetheless. Also, IQ is measured differently for the children and adults so it's not clear these can be directly contrasted. The authors suggest the instructed elaborative encoding strategy is effective for children and adults but the reference in support of this (Craik & Tulving, 1975) does not seem to support this point.

      Thank you for your review, and we appreciate your valuable feedback. Here are our responses and clarifications:

      Regarding the novelty of the results in terms of mentioned existent literature, we believe that in contrast to Cycowicz et al. (2001) and Bauer et al (2017), etc, we assess not only immediate memory after encoding with semantic judgement of abstract associations, but add to these findings investigating consolidation-related changes in complex associative and contextual information in much under investigated sample of 5-to-7-year-old preschoolers. With this we are able to infer also how neural representations of children change over time, providing invaluable insights into knowledge formation in this developmental cohort.

      With this, the observed age differences are not so of primary importance, as time-related changes in mnemonic representations observed in children.

      Regarding the assumption of inattentiveness in children, we want to emphasize that the experimenter was present throughout the learning process, closely supervising the children. We observed prompt responses to every trial in children and noted an increase in accuracy over the encoding-learning cycles, leading us to conclude that the children were indeed attentive to the task. The observed accuracy improvement across learning cycles  indicates increase in remembered information. Furthermore, we took measures to ensure their engagement, including extensive training in both verbal and computerized versions to ensure that they understood and actively created stories to support their learning.

      We collected motivation data after each task execution in children, and the results indicated that they scored high in motivation. Children not only completed the tasks but also expressed their willingness to participate in subsequent appointments, highlighting their active involvement in the study.

      The observed differences in the efficiency of strategy utilization were expected, given developmental differences in the associative and strategic components of memory in children, as noted in prior research (Shing, 2008, 2010).

      We appreciate your point about IQ, sex, and handedness. These variables were indeed included in the behavioral models, and mean brain activation was also included in the brain data models, addressing the potential influence of these factors on our results.

      While it's true that we applied different tests to measure IQ in children and adults, these tests targeted comparable subtests that addressed similar cognitive constructs. As the final IQ values are standardized, we believe it is appropriate to compare them between the two groups.

      Lastly, we agree that the citation Craik & Tulving, 1975 supports the notion of effectiveness of instructed elaborative learning only in adults, but not in children. For this purpose, we added relevant literature for the child cohort (i.e., Pressley, 1982; Pressley et al., 1981; Shing et al., 2008).

      Reviewer #3 (Recommendations For The Authors):

      An additional point for the authors to consider is that the hypotheses were uncertain. The first is that prefrontal, parietal, cerebellar, occipital, and PHG brain regions would have greater activation over time in adults and not children - which is very imprecise as this is basically the whole brain. Moreover, brain imaging data may be in opposition to this prediction: e.g., the hippocampus has a delayed maturational pattern beyond 5-yrs (e.ge., Canada 2019; Uematsu 2012) and some cortical data predicts earlier development in these regions.

      Thank you for your feedback, and we appreciate your insights regarding our hypotheses.

      The selection of our regions of interest (ROIs) was guided by prior literature that has demonstrated the interactive involvement of multiple brain areas in memory retrieval and consolidation processes. Additionally, our recent work utilizing multivariate partial least square correlation analysis (Schommartz, 2022, Developmental Cognitive Neuroscience) has indicated that unique profiles derived from the structural integrity of multiple brain regions are differentially related to short and long-delay memory consolidation.

      Indeed, the literature suggests that the hippocampus may exhibit a more delayed maturational pattern extending into adolescence, as supported by studies such as Canada (2019) and Uematsu (2012), etc. We added this information as well as findings from the literature on cortical development to be more balanced in our review of the literature.

      Given this complexity, we believe it is important to emphasize in our discussion that both the medial temporal lobe, including the hippocampus, and cortical structures, as well as the cerebellum, undergo profound neural maturation. We highlight these nuances in our revised manuscript to provide a more comprehensive perspective on the developmental differences in memory retention over time.

      The writing was challenging to follow - consider as an example on page 9 the sentence that spans 10 lines of text.

      Thank you for bringing this to our attention. We have carefully reviewed the manuscript and have made efforts to streamline the text, ensuring that sentences are not overly long or complex to improve readability and comprehension.

      I found the analysis (and accompanying figures) a bit of a data mine - there are so many results that are hard to digest and in other cases highly redundant one from the other. This may be resolved in part by moving redundant findings to the supplemental. Some were hard to follow - so when there is a line between recent and recent data, that seems confusing to connect data that, I believe, are different sets of items. Later scatterplots (Fig 7) have pale yellow dots that I had a hard time seeing.

      Thank you for bringing up your concerns regarding the analysis and figures in our manuscript. We have carefully considered your feedback and made several improvements to address these issues.

      To alleviate the challenge of digesting numerous results, we have taken steps to enhance clarity and reduce redundancy. Specifically, we have moved some of the redundant findings to the supplementary sections, which should help streamline the main manuscript and make it more reader friendly.

      Regarding the line between 'recent' and 'recent data,' figure were transformed to a clearer version. Furthermore, we have improved the visibility of certain elements, such as the pale-yellow dots in the scatterplots (Fig 1, 2, 4, etc. ), to ensure that readers can better discern the data points.

    1. nearly 75 years — waiting and watching the outside world from behind the thick panes of hospital safety windows because DCFS has too few beds in other facilities where they could be more appropriately treated.

      Did those working with these children not remember or realize that they are still growing and their brains are still developing which makes it even more important for them to not be locked in a hospital for what feels like years to them?

    1. Reviewer #3 (Public Review):

      Summary:

      The study demonstrates differential patterns of entrainment to biological motion (BM). At a basic, sensory level, the authors demonstrate entrainment to faster rhythms that make up BM (step-cycle) which seems to be separate from its audio aspects and its visual aspects (though to a much lesser degree). Ultimately this temporal scale seems to reside in a manner that does not indicate much multi-modal integration. At a higher-order, emergent rhythms in motion that are biologically relevant (gait-cycle) seem to be the result of multisensory integration. The work sheds light on the perceptual processes that are engaged in perceiving BM as well as the role of multisensory integration in these processes. Moreover, the work also outlines interesting links between shorter and longer integration windows along the sensory and multisensory processing stages.

      In a series of experiments, the authors sought to investigate the role of multisensory integration in the processing of biological motion (BM). Specifically, they study neural entrainment in BM light-point walkers. Visual-only, auditory-only, and audio-visual (AV) displays were compared under different conditions.

      Experiments 1a and b mainly characterized entrainment to these stimuli. Here, entrainment to step cycle (at different scales for 1a and 1b) was found to entrain in the presence of the auditory rhythm and to a certain degree also for the visual stimulus (though barely beyond the noise floor in 1b). The AV condition for this temporal scale seemed to follow an additive rule whereby the combined stimulation resulted in entrainment more or less equal to the sum of the unimodal effects. At the slower, gait cycle a slightly different pattern emerges whereby neither unimodal stimulation conditions result in entrainment however the AV condition does.

      This finding was further explored in Experiment 2 where two extra manipulations were added. Point-light walkers could generally be either congruently paired with AV or incongruently. In addition, the visual BM stimulus was matched with a control consisting of an inverted BM and thus non-BM movement. This study enabled further discerning among the step- and gait-cycle findings seeing that the pattern that emerged suggested that step-cycle entrainment was consistent with a low-level process that is not selective to BM whilst gait-cycle entrainment was only found for BM. This generally replicated the findings in Experiment 1 and extended them further suggesting that entrainment seen for uni- and multisensory step cycles is reflects a different process than that captured in the gait-cycle multi-modal entrainment. The selective BM finding seemed to demonstrate a link to autistic traits within a sample of 24 participants informing a hypothesis that sensitivity to biological motion might be related to social cognition.

      Strengths:

      The main strengths of the paper relate to the conceptualization of BM and the way it is operationalized in the experimental design and analyses. The use of entrainment, and the tracking of different, nested aspects of BM result in seemingly clean data that demonstrate the basic pattern. The first experiments essentially provide the basic utility of the methodological innovation and the second experiment further hones in on the relevant interpretation of the findings by the inclusion of better control stimuli sets.

      Another strength of the work is that it includes at a conceptual level two replications.

      Weaknesses:

      The statistical analysis is misleading and inadequate at times. The inclusion of the autism trait is not foreshadowed and adequately motivated and is likely underpowered. Finally, a broader discussion over other nested frequencies that might reside in the point-light walker stimuli would also be important to fully interpret the different peaks in the spectra.

    1. Author response:

      Reviewer #1 (Public Review):

      Using the UK Biobank, this study assessed the value of nuclear magnetic resonance measured metabolites as predictors of progression to diabetes. The authors identified a panel of 9 circulating metabolites that improved the ability in risk prediction of progression from prediabetes to diabetes. In general, this is a well-performed study, and the findings may provide a new approach to identifying those at high risk of developing diabetes. I have some comments that may improve the importance of this study.

      We deeply appreciate the reviewer's invaluable time dedicated to the review of this manuscript and the insightful comments to enhance its overall quality.

      (1) It is unclear why the authors only considered the top 20 variables in the metabolite selection and why they did not set a wider threshold.

      Thank you for the comment. We set the top 20 variables in the metabolite selection balancing the performance of the final diabetes risk prediction model and the clinical applicability due to measurement costs. We have added this explanation in the “Methods” section.

      “We chose the intersection set of the top 20 most important variables selected by the three machine learning models, after balancing the performance of the final diabetes risk prediction model and the clinical applicability associated with measurement costs of metabolites.”

      (2) The methods section would benefit from a more detailed exposition of how parameter tuning was conducted and the range of parameters explored during the training of the RSF model.

      According to the reviewer’s suggestion, we have added a more detailed description of parameters tunning and the range of parameters explored during the training of the RSF model in the “Method S2” section in the Supplementary material.

      “The RSF model was fitted using the “randomForestSRC” package and the grid search method was used for hyperparameter tuning. Specifically, the grid search method was used to tune hyperparameters among the RSF model, through minimizing out-of-sample or out-of-bag error1. Each tree in the RSF is constructed from a random sample of the data, typically a bootstrap sample or 63.2% of the sample size (as in the present study). Consequently, not all observations are used to construct each tree. The observations that are not used in the construction of a tree are referred to as out-of-bag observations. In an RSF model, each tree is built from a different sample of the original data, so each observation is “out-of-bag” for some of the trees. The prediction for an observation can then be obtained using only those trees for which the observation was not used for the construction. A classification for each observation is obtained in this way and the error rate can be estimated from these predictions. The resulting error rate is referred to as the out-of-bag error. Through calculating the out-of-bag error in each iteration, the best hyperparameters were finally determined.

      The hyperparameters to be tuned and range of grid search in the present study were below: number of trees (50-1000, by 50), number of variables to possibly split at each node (3-6, by 1), and minimum size of terminal node (1-20, by 1)2.”

      (3) It is hard to understand the meaning of the decision curve analysis and the clinical implications behind the net benefit, which are required to clarify the application values of models.

      Thank you for the comment. We have added more description and discussion about the decision curve analysis in the “Methods” and “Discussion” sections.

      “Furthermore, we used decision curve analysis (DCA) to assess the clinical usefulness of prediction model-based guidance for prediabetes management, which calculates a clinical “net benefit” for one or more prediction models in comparison to default strategies of treating all or no patients3.”

      “Most importantly, a model with good discrimination does not necessarily have high clinical value. Hence, DCA was used to compare the clinical utility of the model before and after adding the metabolites, and this showed a higher net benefit for the latter than the basic model, suggesting the addition of the metabolites increased the clinical value of prediction, i.e., the potential benefit of guiding management in individuals with prediabetes3,4. These results provided novel evidence supporting the value of metabolic biomarkers in risk prediction and stratification for the progression from prediabetes to diabetes.”

      (4) Notably, the NMR platform utilized within the UK Biobank primarily focused on lipid species. This limitation should be discussed in the manuscript to provide context for interpreting the results and acknowledge the potential bias from the measuring platform.

      Thank you for the comment. We acknowledged this limitation that NMR platform within the UK Biobank primarily focused on lipid species and the potential bias from the measuring platform and have added this in “Discussion” section.

      “Third, the Nightingale metabolomics platform primarily focused on lipids and lipoprotein sub-fractions, and thus the predictive value of other metabolites in the progression from prediabetes to diabetes warranted further research using an untargeted metabolomics approach.”

      (5) The manuscript should explain the potential influence of non-fasting status on the findings, particularly concerning lipoprotein particles and composition. There should be a detailed discussion of how non-fasting status may impact the measurement and the findings.

      According to the reviewer’s suggestion, we have added more details to explain the potential influence of non-fasting status on our findings in the “Discussion” section.

      “Additionally, the use of non-fasting blood samples might increase inter-individual variation in metabolic biomarker concentrations, however, fasting duration has been reported to account for only a small proportion of variation in plasma metabolic biomarker concentrations5. Therefore, we believe the impact of non-fasting samples on our findings would be minor.”

      (6) Cross-platform standardization is an issue in metabolism, and further descriptions of quality control are recommended.

      Thank you for the comment. We have added more description of quality control in the “Method S1” section in the Supplementary material.

      “Metabolic biomarker profiling by Nightingale Health’s NMR platform provides consistent results over time and across spectrometers. Furthermore, the sample preparation is minimal in the Nightingale Health’s metabolic biomarker platform, circumventing all extraction steps. These aspects result in highly repeatable biomarker measurements. Pre-specified quality metrics were agreed between UK Biobank and Nightingale Health to ensure consistent results across the samples, and pilot measurements were conducted. Nightingale Health performed real-time monitoring of the measurement consistency within and between spectrometers throughout the UK Biobank samples. Two control samples provided by Nightingale Health were included in each 96-well plate for tracking the consistency across multiple spectrometers. Furthermore, two blind duplicate samples provided by the UK Biobank were included in each well plate, with the position information unlocked only after results delivery. Coefficient of variation (CV) targets across the metabolic biomarker profile were pre-specified for both Nightingale Health’s internal control samples and UK Biobank’s blind duplicates. The targets were met for each consecutively measured batch of ~25,000 samples. For the majority of the metabolic biomarkers, the CVs were below 5% (https://biobank.ndph.ox.ac.uk/showcase/refer.cgi?id=3000). Further, the distributions of measured biomarkers from 5 sample batches indicated absence of batch effects (https://biobank.ctsu.ox.ac.uk/ukb/ukb/docs/nmrm_app1).”

      Reviewer #2 (Public Review):

      Deciphering the metabolic alterations characterizing the prediabetes-diabetes spectrum could provide early time windows for targeted preventive measures to extend precision medicine while avoiding disproportionate healthcare costs. The authors identified a panel of 9 circulating metabolites combined with basic clinical variables that significantly improved the prediction from prediabetes to diabetes. These findings provided insights into the integration of these metabolites into clinical and public health practice. However, the interpretation of these findings should take account of the following limitations.

      We appreciate the reviewer’s positive comments and encouragement.

      (1) First, the causal relationship between identified metabolites and diabetes or prediabetes deserves to be further examined particularly when the prediabetic status was partially defined. Some metabolites might be the results of prediabetes rather than the casual factors for progression to diabetes.

      Thank you for your insightful comments. We agree with you that the panel of metabolites in this study might not be the causal factor for progression from prediabetes to diabetes, which needs further validation in experimental studies. We have added this limitation in the “Discussion” section.

      “Fifth, we could not draw any conclusion about the causality between the identified metabolites and the risk for progression to diabetes due to the observational nature, which remained to be validated in further experimental studies.”

      (2) The blood samples were taken at random (not all in a non-fasting state) and so the findings were subjected to greater variability. This should be discussed in the limitations.

      According to the reviewer’s suggestion, we have added more details to explain the potential influence of non-fasting status on our findings in the “Discussion” section.

      “Additionally, the use of non-fasting blood samples might increase inter-individual variation in metabolic biomarker concentrations, however, fasting duration has been reported to account for only a small proportion of variation in plasma metabolic biomarker concentrations5. Therefore, we believe the impact of non-fasting samples on our findings would be minor.”

      (3) The strength of NMR in metabolic profiling compared to other techniques (i.e., mass spectrometry [MS], another commonly used metabolic profiling method) could be added in the Discussion section.

      According to the reviewer’s suggestion, we have added the strength of NMR in metabolic profiling compared to other techniques in the “Discussion” section.

      “Circulating metabolites were quantified via NMR-based metabolome profiling within the UK Biobank, which offers metabolite qualification with relatively lower costs and better reproducibility6.”

      (4) Fourth, the applied platform focuses mostly on lipid species which may be a limitation as well.

      Thank you for the comment. We acknowledged this limitation that NMR platform within the UK Biobank primarily focused on lipid species and the potential bias from the measuring platform and have added this in the “Discussion” section.

      “Third, the Nightingale metabolomics platform primarily focused on lipids and lipoprotein sub-fractions, and thus the predictive value of other metabolites in the progression from prediabetes to diabetes warranted further research using an untargeted metabolomics approach.”

      (5) it is a very large group with pre-diabetes, but the results only apply to prediabetes and not to the general population. This should be clear, although the authors have also validated the predictive value of these metabolites in the general population.

      Thank you for the comment. We agree with you that the results only apply to prediabetes and not to the general population, though they also showed potential predictive value among participants with normoglycemia. We have accordingly modified the relevant expressions in the “Conclusion” section to restrict these findings to participants with prediabetes.

      “In this large prospective study among individuals with prediabetes, we detected a panel of circulating metabolites that were associated with an increased risk of progressing to diabetes.”

      References

      (1) Janitza S, Hornung R. On the overestimation of random forest's out-of-bag error. PLoS One. 2018;13(8):e0201904.

      (2) Tian D, Yan HJ, Huang H, et al. Machine Learning-Based Prognostic Model for Patients After Lung Transplantation. JAMA Netw Open. 2023;6(5):e2312022.

      (3) Vickers AJ, van Calster B, Steyerberg EW. A simple, step-by-step guide to interpreting decision curve analysis. Diagn Progn Res. 2019;3:18.

      (4) Li J, Xi F, Yu W, Sun C, Wang X. Real-Time Prediction of Sepsis in Critical Trauma Patients: Machine Learning-Based Modeling Study. JMIR Form Res. 2023;7:e42452.

      (5) Li-Gao R, Hughes DA, le Cessie S, et al. Assessment of reproducibility and biological variability of fasting and postprandial plasma metabolite concentrations using 1H NMR spectroscopy. PLoS One. 2019;14(6):e0218549.

      (6) Geng T-T, Chen J-X, Lu Q, et al. Nuclear Magnetic Resonance–Based Metabolomics and Risk of CKD. American Journal of Kidney Diseases. 2023.

    2. Reviewer #2 (Public Review):

      Deciphering the metabolic alterations characterizing the prediabetes-diabetes spectrum could provide early time windows for targeted preventive measures to extend precision medicine while avoiding disproportionate healthcare costs. The authors identified a panel of 9 circulating metabolites combined with basic clinical variables that significantly improved the prediction from prediabetes to diabetes. These findings provided insights into the integration of these metabolites into clinical and public health practice. However, the interpretation of these findings should take account of the following limitations.

      First, the causal relationship between identified metabolites and diabetes or prediabetes deserves to be further examined particularly when the prediabetic status was partially defined. Some metabolites might be the results of prediabetes rather than the casual factors for progression to diabetes.

      Second, the blood samples were taken at random (not all in a non-fasting state) and so the findings were subjected to greater variability. This should be discussed in the limitations.

      Third, the strength of NMR in metabolic profiling compared to other techniques (i.e., mass spectrometry [MS], another commonly used metabolic profiling method) could be added in the Discussion section.

      Fourth, the applied platform focuses mostly on lipid species which may be a limitation as well.

      Fifth, it is a very large group with pre-diabetes, but the results only apply to prediabetes and not to the general population. This should be clear, although the authors have also validated the predictive value of these metabolites in the general population.

    1. Author response:

      The following is the authors’ response to the previous reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      The manuscript addresses a fundamental question about how different types of communication signals differentially affect brain state and neurochemistry. In addition, their manuscript  highlights the various processes that modulate brain responses to communication signals, including prior experience, sex, and hormonal status. Overall, the manuscript is well-written and the research is appropriately contextualized.

      That being said, it remains important for the authors to think more about their analytical approaches. In particular, the effect of normalization and the explicit outlining and interpretations of statistical models. As mentioned in the original review, the normalization of neurochemical data seems unnecessary given the repeated-measures design of their analysis and by normalizing all data to the baseline data and including this baseline data in the repeated measures analysis,   one artificially creates a baseline period with minimal variation that dramatically differs in variance from other periods (akin to heteroscedasticity). If the authors want to analyze how a stimulus changes neurochemical concentrations, they could analyze the raw data but depict normalized data in their figures (similar to other papers). Or they could analyze group differences in the normalized data of the two stimulus periods (i.e., excluding the baseline period used for normalization).

      We appreciate the reviewer’s point on the difference in variance caused by including the 100% baseline values in the analysis. After consulting with our statistician, we chose the latter of the two approaches suggested by the reviewer. Specifically, we reran the analysis to exclude the baseline and focus only on the playback windows and the group differences. The text in the results, the significance signs in the figures, and the discussion are corrected accordingly. Despite these changes, our major conclusions remains as before.

      We also followed this reviewer’s suggestions to clarify the statistical model in studying the experience effect. After further consultation with our statistician, we reran the analysis on experience effect, including all the groups of EXP and INEXP animals together. We have corrected text in the figure captions, results, discussion, and data analysis sections of the manuscript related to the effect of experience and its interactions. This has not changed the conclusion made related to the experience effect in the dataset.

      It would also be useful for the authors to provide further discussion of the potential contributions of different types of experiences (mating vs. restraint) to the change in behavior and neurochemical responses to the vocalization playbacks and to try to disentangle sensory and  motor contributions to neurochemical changes.

      We have acknowledged in the Discussion that previous studies suggest that the effect of experience involving stress could be generalized. We believe that this is an important area of future research. Our Discussion acknowledges that the relationship between sensory and motor contributions to neurochemical changes remains an area of interest. We further point out that the time resolution of microdialysis data renders the suggested discussion highly speculative. We plan to use other methods to assess this in future experiments.

      Reviewer #3 (Public Review):

      The work by Ghasemahmad et al. has the potential to significantly advance our understanding of how neuromodulators provide internal-state signals to the basolateral amygdala (BLA) while an animal listens to social vocalizations.

      Ghasemahmad et al. made changes to the manuscript that have significantly improved the work. In particular, the transparency in showing the underlying levels of Ach, DA, and 5HIAA is excellent. My previous concerns have been adequately addressed.

      Recommendations for the authors:

      Reviewer #1 (Recommendations For The Authors):

      I appreciate the authors responses to my previous queries (and to the comments by other reviewers). The introduction does a better job contextualizing the data, and the additional details in the results and Methods sections help readers digest the material. I continue to think the topic  is interesting and the manuscript is potentially impactful. However, I continue to be concerned about their analytical approaches and other aspects of the revised manuscript.

      (a) Normalization

      In my original review I wrote: "The normalization of neurochemical data seems unnecessary   given the repeated-measures design of their analysis and could be problematic; by normalizing     all data to the baseline data (p. 24), one artificially creates a baseline period with minimal   variation (all are "0"; Figures 2, 3 & 5) that could inflate statistical power." I continue to feel that an analysis of normalized data that includes the baseline data is inappropriate because of the minimal variation in the normalized data for the baseline period. When the normalized data for   the baseline period is included in the analysis, there is clearly variation in the extent of variability within each of the time periods (no variability at baseline, variability during periods 1 & 2; analogous to heteroscedasticity). For example, when analyzing the RAW DATA about the change in ACh release in experienced males listening to restraint vocalizations (thank you for releasing the raw data), there was a non-significant effect of time (baseline, period 1, and period 2; linear mixed effects model; F(2,12)=3.2, p=0.0793). However, when the normalized data for  this dataset was analyzed (with baseline values being set at 100% for each mouse), there was a statistically significant effect (F(2,12)=4.5, p=0.0352). This example is just to illustrate how normalization can affect (e.g., inflate) statistical power.

      That being said, I do think that it is reasonable to analyzed normalized data if the period used for normalization is NOT included in the analysis (see Figure 3 of one of the paper the authors listed in their response to reviewers: Galvez-Marquez et al., 2022). However, from the reading of this manuscript, it does seem like normalized baseline data are analyzed to assess how stimuli affect neurochemical concentrations.

      We appreciate the reviewer’s point on the difference in variance caused by including the 100% baseline values in the analysis. After consulting with our statistician, we chose one of the two approaches suggested by the reviewer. Specifically, we reran the analysis to exclude the baseline and focus only on the playback windows and the group differences. The text in the results, the significance signs in the figures, and the discussion are corrected accordingly. Despite these changes, our major conclusions remains as before. We have included some descriptive statistics in the text because we think these are informative.

      We decided to take this approach because the inter-individual variability in the raw data levels, caused by non-experimental factors, is too great to be useful. As we have stated before, these values are affected by probe placement, collection process, or differences in the HPLC or LC/MS runs. These effects are widely recognized in the field.

      It is worth pointing out a few things about the papers listed by the authors. Li et al. (2023) does depict normalized microanalysis data but it isn't clear that any analysis of the normalized data is conducted. The same can be said about Holly et al. (2016). Further, in Bagley et al (2011), the authors depict normalized data in the figures but conduct analyses on the raw data ("After  chronic morphine treatment, systemic naloxone injection increased GABA outflow in PAG by 41% (from 24.6 {plus minus} 2.9 nM to a peak of 34.8 {plus minus} 3.8 nM, n = 6, P = 0.016), but did not alter GABA levels after vehicle treatment (39.8 {plus minus} 8.3 to 38.6 {plus  minus} 7.4 nM with naloxone at matched peak time, n = 4; Fig. 3a)". This latter approach (analyzing raw data in a repeated-measures manner and depicted normalized data) seems reasonable for the authors of the current study.

      (b) Clarification and modification of statistical models

      When analyzing the effect of experience on neuromodulator release, the authors analyze the experienced and inexperienced mice independently (e.g., figure 3 vs. 6). The ideal way to assess the effects of experience is to create a factorial model. For example, one could analyze a full factorial model with experience (exp vs. inexp), stimulus time (mating vs. restraint) and time  (baseline, period 1 vs period 2, assuming raw data are used). If one wanted to exclude the  baseline period because group differences in baseline are not informative, conducting a factorial analysis of normalized data with just the data from period 1 and 2 seems fine. I believe an analysis like this will help increase the legitimacy of the analysis. For example, when analyzing the normalized data (periods 1 and 2) of experienced and inexperienced males in response to mating or restraint vocalizations, you find a significant interaction between experience and stimulus type. Finding an effect of experience in an analysis that includes both experienced and inexperienced mice is ideal from an analytical framework.

      In Figure 6, it is not clear what the statistical model is and what the interactions mean. For example, in the figure legend for figure 6, the authors report time*context and time*sex interactions. However, in this analysis there are two groups of inexperienced males (males that   are listening to restraint vocalizations, males that are listening to mating vocalizations) and one group of females (females that are listening to mating vocalizations); in other words, this is an unbalanced analysis. So, when the authors indicate a time*context interaction, does that mean  they are comparing the male-restraint group to the combination of males and females listening to mating vocalizations? And when they talk about a time*sex interaction, are they analyzing how males listening to either mating or restraint vocalizations differ from females listening to a   mating vocalization? This all seems peculiar to me.

      - A similar set of questions could be raised about interaction effects depicted in Figure 4.

      Overall, I would like this manuscript to be reviewed by a statistician to provide additional input on how best to analyze the data.

      We followed the reviewer’s suggestions to clarify the statistical model in studying the experience effect. After further consultation with the statistician, we reran the analysis on experience effect, including all the groups of EXP and INEXP animals together.

      Design: Intercept + Sex +Context + Experience+ Sex* Experience + Context* Experience.

      The model is not full factorial as recommended by the statistician, because we don’t have females in the restraint group and that would make an unbalanced design. Therefore, running GLM based on the above model and included factors, as advised by the statistician, is the best way of approaching the analysis for the current dataset.

      We have corrected text in the figure captions, results, discussion, and data analysis sections of the manuscript related to the effect of experience and its interactions. The GLM models are clarified for all the figures in the “data analysis” section of the manuscript. We have clarified that the major effect of experience on neuromodulators was seen in the ACh data.

      (c) Analysis of post-stimulus period

      I agree with Reviewer 3 that analyzing the post-stimulus period would be useful. As mentioned     in the original review, these data could serve as an opportunity to show that the neurochemical levels returned to baseline and add further support for the model described in Figure 6. In   addition, these data could help reveal the link  between  neurochemical  release,  auditory responses, and behavior. If neurochemical changes reflect auditory responses, then these should back to baseline during the post-stimulus period. In addition, if behavioral variation (e.g.,    between mice hearing mating vs. restraint stimuli) persists following the termination of playback, then one could similarly assess whether neurochemical variation persists following playback. If   the latter is the case, then the neurochemical release could be more related to the behavior than to the playback stimulus itself.

      We did not change this analysis. Our response to Reviewer 3’s comment is shown below.

      “We decided not to include analyses of the post-stimulus period because this period is subject to wider individual and neuromodulator-specific effects and because it weakens statistical power in addressing the core question—the change in neuromodulator release DURING vocal playback. We agree that the general question is of interest to the field, but we don’t think our study is best designed to answer that question.”

      This was accepted by Reviewer 3. We also note that release patterns have multiple time courses (e.g., Aitta-aho et al., 2018 for ACh), and thus may not support an assumption that levels should return to baseline shortly after playback offset.

      Minor comments:

      Page 7, line 15: I suggest changing "vocalization-dependent" to "stimulus-dependent" because the former could connote patterns of release related to the animal itself vocalizing.

      Changed to: “There were also distinct patterns of ACh and DA release into the BLA depending on the type of vocalization playback (Fig 3C,D).”

      Discussion section: The authors should point out a few caveats with their experiments in the Discussion section. First, experienced animals received both mating (social) and restraint experiences, and it is not clear to what degree each type of experience affected neural and behavioral responses (i.e., specificity of experience effects). For example, mating experience can lead to a wide range of physiological changes, including a resilience to stress (e.g., Leuner et al., PLoS One, 2010; Arnold et al., Hormones and Behavior, 2019), so it is possible that mating experiences by themselves could have induced these changes. Or it could be that experiencing restraint stress affects responses to mating stimuli. This could be added to the first paragraph in page 16. (The authors could also discuss which aspects of the sexual encounters might be most important for the behavioral and neural plasticity.)

      We have added text to raise this issue, stating that it is unknown wither the experience effects are specific and citing the above references concerning the generalized effects of certain experiences.

      Discussion section: It would also be useful for the authors to discuss the extent to which behavior might be driving the neurochemical changes. Some of the analyses suggest that the release is independent of the behavior (e.g., reflects a sensory responses) but this could be emphasized    more in the Discussion.

      We believe that we have addressed this issue sufficiently in our previous response to related issues raised by this reviewer. As we note, there are limitations in the time resolution of microdialysis data that render the suggested discussion highly speculative. We plan to use other methods to assess this in future experiments.

      Figure 2, legend: Please note that the text above the images describes the stimulus played back to these animals and their hormonal state, and not the type of experienced they underwent (i.e.,  clarify the titles)

      Changed as requested.

      I also agree with Reviewer 3 that "mating experience" is a misnomer for this manuscript. "Social experience with a female" is a more accurate descriptor. If they wanted to specifically provide mating experience, males should have only been tested with estrus (receptive females). I don't think this wording change detracts from their findings.

      We have not changed this term. As noted in our previous response to Reviewer #3, we stated: “In the mating experience, mounting or attempted mounting was required for the animal to be included in subsequent testing.” Due to this requirement, the term “mating behavior” is informative and appropriate. In our view, “Social experience with a female” does not adequately describe our inclusion criterion or the experience.

      Reviewer #3 (Recommendations For The Authors):

      The work by Ghasemahmad et al. has the potential to significantly advance our understanding of how neuromodulators provide internal-state signals to the basolateral amygdala (BLA) while an animal listens to social vocalizations.

      Ghasemahmad et al. made changes to the manuscript that have significantly improved the work. In particular, the transparency in showing the underlying levels of Ach, DA, and 5HIAA is excellent. My previous concerns have been adequately addressed. I only have a few minor suggestions for the text and one figure.

      Minor suggestions:

      Page 2, Ln 9: add adult before male and female mice

      Changed as requested

      Page 4, Ln 10: add a period after Tsukano et al., 2019)

      Changed as requested

      Page 6, Ln 9: what did you mean by "their interaction"? Being more specific, but concise, would help the readers.

      We revised the wording to clarify that the neuromodulatory systems interact in the emission of positive and negative vocalizations.

      Page 6, Ln 17: You mention Stim 1 and Stim 2, but the stimuli are not defined at this point. The clear explanation is provided in the following paragraph. Maybe consider switching the order  and define the stimuli before you describe the liquid chromatography/mass spectrometry technique.

      We have revised and merged these paragraphs so that Stim 1 and Stim 2 are defined on first use. We also revised our description of the depiction and analysis of neurochemical data.

      Page 11, Ln 12: replace well-proven with well-documented

      Changed as requested

      Figure 2: There are two arrows pointing towards a single track. I assume one of the arrows is a duplicate. If so, delete one of the arrows. If not, please explain what the second arrow represents.

      Arrow removed

    1. Reviewer #2 (Public Review):

      Summary:

      This paper aims to test if neural representations of images of objects in the human brain contain a 'pure' dimension of real-world size that is independent of retinal size or perceived depth. To this end, they apply representational similarity analysis on EEG responses in 10 human subjects to a set of 200 images from a publicly available database (THINGS-EEG2), correlating pairwise distinctions in evoked activity between images with pairwise differences in human ratings of real-world size (from THINGS+). By partialling out correlations with metrics of retinal size and perceived depth from the resulting EEG correlation time courses, the paper claims to identify an independent representation of real-world size starting at 170 ms in the EEG signal. Further comparisons with artificial neural networks and language embeddings lead the authors to claim this correlation reflects a relatively 'high-level' and 'stable' neural representation.

      Strengths:

      - The paper features insightful figures/illustrations and clear figures.

      - The limitations of prior work motivating the current study are clearly explained and seem reasonable (although the rationale for why using 'ecological' stimuli with backgrounds matters when studying real-world size could be made clearer; one could also argue the opposite, that to get a 'pure' representation of the real-world size of an 'object concept', one should actually show objects in isolation).

      - The partial correlation analysis convincingly demonstrates how correlations between feature spaces can affect their correlations with EEG responses (and how taking into account these correlations can disentangle them better).

      - The RSA analysis and associated statistical methods appear solid.

      Weaknesses:

      - The claim of methodological novelty is overblown. Comparing image metrics, behavioral measurements, and ANN activations against EEG using RSA is a commonly used approach to study neural object representations. The dataset size (200 test images from THINGS) is not particularly large, and neither is comparing pre-trained DNNs and language models, or using partial correlations.

      - The claims also seem too broad given the fairly small set of RDMs that are used here (3 size metrics, 4 ANN layers, 1 Word2Vec RDM): there are many aspects of object processing not studied here, so it's not correct to say this study provides a 'detailed and clear characterization of the object processing process'.

      - The paper lacks an analysis demonstrating the validity of the real-world depth measure, which is here computed from the other two metrics by simply dividing them. The rationale and logic of this metric is not clearly explained. Is it intended to reflect the hypothesized egocentric distance to the object in the image if the person had in fact been 'inside' the image? How do we know this is valid? It would be helpful if the authors provided a validation of this metric.

      - Given that there is only 1 image/concept here, the factor of real-world size may be confounded with other things, such as semantic category (e.g. buildings vs. tools). While the comparison of the real-world size metric appears to be effectively disentangled from retinal size and (the author's metric of) depth here, there are still many other object properties that are likely correlated with real-world size and therefore will confound identifying a 'pure' representation of real-world size in EEG. This could be addressed by adding more hypothesis RDMs reflecting different aspects of the images that may correlate with real-world size.

      - The choice of ANNs lacks a clear motivation. Why these two particular networks? Why pick only 2 somewhat arbitrary layers? If the goal is to identify more semantic representations using CLIP, the comparison between CLIP and vision-only ResNet should be done with models trained on the same training datasets (to exclude the effect of training dataset size & quality; cf Wang et al., 2023). This is necessary to substantiate the claims on page 19 which attributed the differences between models in terms of their EEG correlations to one of them being a 'visual model' vs. 'visual-semantic model'.

      - The first part of the claim on page 22 based on Figure 4 'The above results reveal that real-world size emerges with later peak neural latencies and in the later layers of ANNs, regardless of image background information' is not valid since no EEG results for images without backgrounds are shown (only ANNs).

      Appraisal of claims:

      While the method shows useful and interesting patterns of results can be obtained by combining contrasting behavioral/image metrics, the lack of additional control models makes the evidence for the claimed unconfounded representation of real-world size in EEG responses incomplete.

      Discussion of likely impact:

      The paper is likely to impact the field by showcasing how using partial correlations in RSA is useful, rather than providing conclusive evidence regarding neural representations of objects and their sizes.

      Additional context important to consider when interpreting this work:

      - Page 20, the authors point out similarities of peak correlations between models ('Interestingly, the peaks of significant time windows for the EEG × HYP RSA also correspond with the peaks of the EEG × ANN RSA timecourse (Figure 3D,F)'. Although not explicitly stated, this seems to imply that they infer from this that the ANN-EEG correlation might be driven by their representation of the hypothesized feature spaces. However this does not follow: in EEG-image metric model comparisons it is very typical to see multiple peaks, for any type of model, this simply reflects specific time points in EEG at which visual inputs (images) yield distinctive EEG amplitudes (perhaps due to stereotypical waves of neural processing?), but one cannot infer the information being processed is the same. To investigate this, one could for example conduct variance partitioning or commonality analysis to see if there is variance at these specific time-points that is shared by a specific combination of the hypothesis and ANN feature spaces.

      - Page 22 mentions 'The significant time-window (90-300ms) of similarity between Word2Vec RDM and EEG RDMs (Figure 5B) contained the significant time-window of EEG x real-world size representational similarity (Figure 3B)'. This is not particularly meaningful given that the Word2Vec correlation is significant for the entire EEG epoch (from the time-point of the signal 'arriving' in visual cortex around ~90 ms) and is thus much less temporally specific than the real-world size EEG correlation. Again a stronger test of whether Word2Vec indeed captures neural representations of real-world size could be to identify EEG time-points at which there are unique Word2Vec correlations that are not explained by either ResNet or CLIP, and see if those time-points share variance with the real-world size hypothesized RDM.

    1. AbstractDefining a multicellular model can be challenging. There may be hundreds of parameters that specify the attributes and behaviors of objects. Hopefully the model will be defined using some format specification, e.g., a markup language, that will provide easy model sharing (and a minimal step toward reproducibility). PhysiCell is an open source, physics-based multicellular simulation framework with an active and growing user community. It uses XML to define a model and, traditionally, users needed to manually edit the XML to modify the model. PhysiCell Studio is a tool to make this task easier. It provides a graphical user interface that allows editing the XML model definition, including the creation and deletion of fundamental objects, e.g., cell types and substrates in the microenvironment. It also lets users build their model by defining initial conditions and biological rules, run simulations, and view results interactively. PhysiCell Studio has evolved over multiple workshops and academic courses in recent years which has led to many improvements. Its design and development has benefited from an active undergraduate and graduate research program. Like PhysiCell, the Studio is open source software and contributions from the community are encouraged.

      This work has been published in GigaByte Journal under a CC-BY 4.0 license (https://doi.org/10.46471/gigabyte.128), and has published the reviews under the same license. This is part of the PhysiCell Ecosystem Series: https://doi.org/10.46471/GIGABYTE_SERIES_0003

      Reviewer 1. Meghna Verma:

      Is installation/deployment sufficiently outlined in the paper and documentation, and does it proceed as outlined?

      The authors have provided links for video descriptions for installation and that is appreciated.

      One overall recommendation is: If all the screenshots (for e.g.: from Fig 1-12 of the main paper and all the subsections in Supplementary) can be combined in one figure that will help enhance the complete overview and the overall flow of the paper.

      Additional comments are available here: https://gigabyte-review.rivervalleytechnologies.comdownload-api-file?ZmlsZV9wYXRoPXVwbG9hZHMvZ3gvVFIvNTA3L1Jldmlld19QaHlzaUNlbGxTdHVkaW9fTVYucGRm

      Reviewer 2. Koert Schreurs and Lin Wouters supervised by Inge Wortel

      Is there a clear statement of need explaining what problems the software is designed to solve and who the target audience is?

      The problem statement is addressed in the introduction, which mentions the need for a GUI tool as a much more accessible way to edit the XML-based model syntax. However, it is somewhat confusing who exactly the intended audience of the paper is. Is the paper targeted at researchers that already use PhysiCell, but might want to switch to the GUI version? Or should it (also) target the potential new user-base of researchers interested in using ABMs, for whom the XML version was not sufficiently accessible and who will now gain access to these models because there is a GUI? Specifying the intended audience might impact some sections of the paper. For example, for users who already use PhysiCell, the step-by-step tutorials might not be useful since they would already know most of the available options; they would just need a quick overview of what info is in which tab. But if the paper is (also) targeted at potential new users, then some additional information could make both the paper and the tool much more accessible, such as:
      
      • A clear comparison to other modeling frameworks and their functionalities. Why should they use PhysiCell instead of one of the other available (GUI) tools? For example, the referenced Morpheus, CC3D and Artistoo all focus on a different model framework (CPMs); this might be worth mentioning. And what about Chaste? Does it represent different types of models, or are there other reasons to consider PhysiCell over Chaste or vice versa? For new users, this would be important information to include. The paper currently also does not mention other frameworks except those that offer a GUI. While the main point of the paper is the addition of the GUI, for completeness sake it might still be good to mention a broader overview of ABM frameworks and how they compare to PhysiCell, or simply to refer to an existing paper that provides such an overview.
      • The current tutorial immediately dives into very specific instructions (what to click and exact values to enter), often without explaining what these options mean or do. New users would probably appreciate to get a rough outline of which types of processes can be modelled, and which steps they would take to do so. This could be as easy as summarising the different main tabs before going into the details. I understand that some of these explanations will overlap with the main PhysiCell software – but considering that the GUI will open up modelling to a different type of community, it might make sense to outline them here to get a self-contained overview of functionality.
      • Indeed, if the above information is provided, the detailed tutorial might fit better as an appendix or in online documentation. That would also leave more space to explain not only which values to enter, but also what these variables do, why choose these values, what other options to consider, etc. Having this information together in one place would be very useful for beginning users.

      Is the source code available, and has an appropriate Open Source Initiative license been assigned to the code?

      The software is available under the GPL v3 licence.

      As Open Source Software are there guidelines on how to contribute, report issues or seek support on the code?

      There is a Github repository, ensuring that it is possible to contribute and report issues, and the paper explicitly invites community contributions. However, although the paper mentions that it is possible to seek support through Github Issues and “Slack channels”, we could find no link to the latter resource. This should probably be added to make this resource usable for the reader (or otherwise the statement should be removed)

      Is installation/deployment sufficiently outlined in the paper and documentation, and does it proceed as outlined?

      Mostly yes, as installation and deployment are outlined in the paper and documentation. However, we did notice a couple of issues: - The studio guide explains how to compile a project in PhysiCell (https://github.com/PhysiCell-Tools/Studio-Guide/blob/main/README.md), but does not mention that Mac users need to specify the g++ version at the top of the Makefile. This is explained in a separate blog (http://www.mathcancer.org/blog/setting-up-gcc-openmp-on-osx-homebrew-edition/) but should be outlined (or at least referenced) here as well. - There are several different resources covering the installation process, referring to e.g. github.com/physicell-training, github.com/PhysiCell-Tools/Studio-Guide, and the abovementioned blog. But this might not be very accessible to all users targeted by the new GUI functionality (especially when command line interventions and manual Makefile edits are involved). While not all of this has to be changed before publication, having all information in one place would already improve accessibility to a larger user-base. - When following the instructions (https://github.com/PhysiCell-Tools/Studio-Guide/blob/main/README.md), “python studio/bin/studio.py -p -e virus-sample” the -p flag gives an error: “Invalid argument(s): [‘-p’]”. We assumed it has to be left out, but perhaps the docs have to be updated.

      Is the documentation provided clear and user friendly?

      Mostly yes, as there is already a lot of documentation available. However, the user-friendliness could be improved with some minor changes. For example, the documentation could be made more user-friendly if resources were available from a central spot. Currently, information can be found in different places: - https://github.com/PhysiCell-Tools/Studio-Guide/blob/main/README.md provides installation instructions and a nice overview of what is where in the GUI, but as mentioned above, does not mention potential issues when installing on MacOS. - The paper provides very detailed examples; these might be nice to include along with the abovementioned overview. - Potentially other places as well. It would be great if the main documentation page could at least link to these other resources with a brief description of what the user will find there. Further, some additions would make the documentation more complete: - It would be good to have an overview somewhere of all the configuration files that can be supplied/loaded (e.g. those for “rules” and for initial configurations). - A clearer instruction/small tutorial on how to use simularium and paraview with physicell studio; especially for paraview there is no instruction on how to use your own data or make your own `.pvsm` file In the longer term, it might be worthwhile to set up a self-contained documentation website (this is relatively easy nowadays using e.g. Github pages), which can outline dependencies, installation instructions, a quick overview, detailed tutorials, example models, links to Github issues/slack communities. This is not a requirement for publication but might be worth looking into in the future as it would be more user-friendly.
      

      Is there a clearly-stated list of dependencies, and is the core functionality of the software documented to a satisfactory level?

      No. The core functionality of the software is nicely outlined in the Github README (https://github.com/PhysiCell-Tools/Studio-Guide/blob/main/README.md), but as mentioned before, this high-level overview is missing in the paper itself. The README and paper recommend installing the Anaconda python distribution to get the required python dependencies. This is fine, but adding a setup file or requirements.txt might still be useful for users who are more familiar with python and want a more minimal installation. Providing a conda environment.yml that allows running the studio along with paraview and/or simularium might also be helpful. Note that running the studio with simularium in anaconda did not work because anaconda did not have the required vtk v9.3.0; instead we had to install simularium without anaconda (“pip3 install simularium”).

      Are there (ideally real world) examples demonstrating use of the software?

      The detail tutorial nicely walks the reader through the tool (although as mentioned before, a high-level overview is missing and the level of detail feels slightly out of place in the paper itself). When walking through the example in the paper and the supplementary, we did run into a few (minor) issues: - It might be good to stress explicitly that after copying the template.xml into tumor_demo.xml, the first step is always to compile using “make”. The paper mentions “Assuming … you have compiled the template project executable (called “project”) …”. But it might not be immediately clear to all users how exactly they should do so (presumably by running “make tumor_demo” after copying the xml file?). - When running “python studio/bin/studio.py -c tumor_demo.xml -e project” as instructed, a warning pops up that “rules0.csv” is not valid (although the tool itself still works). - The instructions for plotting say to press “enter” when changing cmin and cmax, but Mac offers only a return key. Pressing fn+return to get the enter functionality also does not work; it might be good to offer an alternative for Mac. - When reproducing the supplementary tutorial, results were slightly different. It might be good if the example would offer a random seed so that users can verify that they can reproduce these results exactly. In our hands, when reproducing figs 39, 40, 48, 49 yields way more (red) macrophages (even when running multiple times), but we could not be sure if this is due to variation between runs, or a mistake in the settings somewhere.
      
      
      The paper mentions that they have started setting up automated testing, but it does not give an idea of what the current test coverage is. Did they add a few tests here and there, or start to systematically test all parts of the software? I understand the latter might not be achievable immediately, but it would be good if users and/or contributors can at least get a sense of how good the current coverage is. (Note: the framework uses pytest, which seems to offer some functionality to generate coverage reports, see e.g. https://www.lambdatest.com/blog/pytest-code-coverage-report/). The code in studio_for_pytest.py has a comment “do later, otherwise problems sometimes”, but it is not entirely clear if the relevant issue has been resolved.
      

      Additional Comments: The presented tool offers a GUI interface to the PhysiCell framework for agent-based modeling. As outlined for the paper, this offers significant value to the users since editing a model is now much more accessible. The tool comes with extensive functionality and instructions. Overall, the tool functions as advertised, and will be of great value to the community of PhysiCell users that now have to edit XML files by hand. It is therefore (mostly) publishable as is if some of the issues with installation (mentioned above) can be straightened out. That said, we do think some improvements could make both the tool and the paper more accessible to a larger user audience. Most of these have been mentioned in the other questions, but we will list some additional ones below. Note that many of these are just suggestions, so we will leave it up to the authors if and when they implement them.

      Suggestions for the paper: While the paper nicely outlines design ideas and usage of the tool, there were some points where we felt that the main point did not quite come across, for example: - As mentioned in the question about problem statement and intended audience, adding some information to the paper would make it a more useful resource to users not yet familiar with PhysiCell (see remarks there). - The section “Design and development” describes the development history of the tool. In principle this is a valuable addition, because it illustrates how the project is under ongoing development and has already been improved several times based on feedback of users. However, the amount of information on each previous stage is slightly confusing; it is not entirely clear how this relates to the paper and current tool. If the main point is to showcase that the current tool has been built based on practical user experiences, this would probably come across better if this section was somewhat shorter and focused on the design choices rather than previous versions. If the main point is something else, it should be clarified what the main idea is. – The point of Table 1 was unclear to us – consider removing or explaining the main idea. - Several figures do not have captions (e.g. Figure 1 but also others); it would be helpful to clarify what message the figure should convey. – P4 “adjust the syntax for Windows if necessary” – is it self-explanatory how users should adjust? Consider adding the correct code for windows as well if possible, since users that want to use the GUI tool might not be familiar with command line syntax. - P6 “if you create your own custom C++ code referring directly to cell type ID” – this functionality is never discussed. This might be part of the general PhysiCell functionality, but it would be good to at least provide a link to a resource on how you could do this. - P8 “Only those parameters that display … editing the C++ code” – it was not entirely clear to me what this means, could you clarify? - P13 mentions you can immediately see changes to the model parameters made. This is very useful for prototyping when users want immediate feedback. However, what happens when you try to save output for a simulation where parameters were changed while the simulation was running? Would users be reminded that their current output is not representative? - Discussion: it is good to mention that the tool is already being used. Can you give an indication based on your experience how long it takes new users to learn to navigate the tool? This might be useful information to add in the paper. - The last statement on LLMs seems to come out of nowhere. Consider leaving it out or expanding further on what would be needed to make this work/how feasible this is.

      Further comments on the tool itelf: - The paper mentions that results may not be fully reproducible if multiple threads are used (I assume this is the case even when a random seed is set). In this case, would it make sense to throw a warning the first time a user tries to set a seed with multiple threads, to avoid confusion as to why the results are not reproducible? - Unusable fields are not always greyed out to indicate that they are disabled, which sometimes makes it seem as though the tool is unresponsive. In other places unusable options are set to grey, so it might be good to double-check if this is consistent. - At the initial conditions (IC) page there is no legend; it might be good to add one. - There are some small inconsistencies between the field names mentioned in the paper and those in the tool/screenshots. For example “boundary condition” (p5) should be “dirichlet BC”, “uptake” (p6) should be “uptake rate”. For the latter, the paper mentions that the length scale is 100 micron but this should be visible in the tool as well. - Not all fields have labels, so it is not always clear what the options do (see e.g. drop-downs in Figure 6). – There are a few points in the tool where you have to “enable” a functionality before it works, but this might not always be intuitive. For example, if you upload a file with initial conditions, it can be assumed that you want to use it. There might be good reasons for this in some cases but in general, consider if all these checkpoints are necessary or if this could be simplified. Same goes for the csv files that have to be saved separately instead of through the main “save” button – in the long term it might be worth saving all relevant files when they are updated, or at least throwing a warning that you have to save some of them separately.

    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      In this study, Jellinger et al. performed engram-specific sequencing and identified genes that were selectively regulated in positive/negative engram populations. In addition, they performed chronic activation of the negative engram population over 3 months and observed several effects on fear/anxiety behavior and cellular events such as upregulation of glial cells and decreased GABA levels.

      Strengths:

      They provide useful engram-specific GSEA data and the main concept of the study, linking negative valence/memory encoding to cellular level outcomes including upregulation of glial cells, is interesting and valuable.

      Weaknesses:

      A number of experimental shortcomings make the conclusion of the study largely unsupported. In addition, the observed differences in behavioral experiments are rather small, inconsistent, and the interpretation of the differences is not compelling.

      Major points for improvement:

      (1) Lack of essential control experiments

      With the current set of experiments, it is not certain that the DREADD system they used was potent and stable throughout the 3 months of manipulations. Basic confirmatory experiments (e.g., slice physiology at 1m vs. 3m) to show that the DREADD effects on these vHP are stable would be an essential bottom line to make these manipulation experiments convincing.

      In previous work from our lab performing long-term activation of Gq DREADD receptors in the vHPC, we quantify the presence of Gq receptor expression over 3-, 6- and 9-month timepoints and show that there is no decrease in receptor expression, as measured via fluorescence intensity (Suthard et al., 2023). In this study, we also address that even if our manipulation is only working for 1 month, rather than 3 months, we are observing the long-term effects of this shorter-term stimulation. This is still relevant, and only changes how we interpret these findings, as shorter-term stimulation or disruption of neuronal activity can still have detrimental effects on behavior.

      Furthermore, although the authors use the mCherry vector as a control, they did not have a vehicle/saline control for the hM3Dq AAV. Thus, the long-term effects such as the increase in glial cells could simply be due to the toxicity of DREADD expression, rather than an induced activity of these cells.

      For chemogenetic studies, our experimental rationale utilized a standard approach in the field, which includes one of two control options: 1) active receptor vs. control vector + ligand or 2) active receptor + ligand or saline control. We chose the first option, as this more properly controls for the potential off-target effects of the ligand itself, as shown in other previous work (Xia et al., 2017). This is particularly important for studies using CNO, as many off-target effects have been noted as a limitation (Manvich et al., 2018). We chose to use DCZ as it is closely related to CNO and newer ligands, but comes with added benefits of high specificity, low off-target effects, high potency and brain penetrance (Nagai et al., 2020), but any potential off-target effects of DCZ are yet to be completely investigated as this ligand is very new.

      Evidence of DREADD toxicity has been shown at high titer levels of AAV2/7- CamKIIα-hM4D(Gi)-mCherry in the hippocampus at 5 weeks, as the reviewer pointed out in their above comment (Goossens et al., 2021). Our viral strategy is targeted to a much smaller number of cells using AAV9-DIO-Flex-hM3Dq-mCherry at a lower titer, unlike expression within a much larger population of CaMKII+ excitatory neurons in this study. Additionally, visual comparison of their viral load and expression with ours shows much more intense expression that spans a larger area of the hippocampus (Goossens et al, 2021; Figure 1D), whereas ours is isolated to a smaller region of vHPC (see Figure 1B).

      Further, we attempted to quantify a decrease in neuronal health (Yousef et al., 2017) resulting from DREADD expression via NeuN counts within multiple hippocampal subregions for the 6- and 14-month groups across active Gq receptor and mCherry conditions and did not observe significant decreases in NeuN as a result (Supplemental Figure 1). However, immunohistochemistry of an individual marker may not be sufficient to capture the entire health profile of an individual neuron and future work should consider other markers of cell death or inflammation, which we have added to the Limitations & Future Work section of our Discussion.

      (2) Figure 1 and the rest of the study are disconnected

      The authors used the cFos-tTA system to label positive/negative engram populations, while the TRAP2 system was used for the chronic activation experiments. Although both genetic tools are based on the same IEG Fos, the sensitivity of the tools needs to be validated. In particular, the sensitivity of the TRAP2 system can be arbitrarily altered by the amount of tamoxifen (or 4OHT) and the administration protocols. The authors should at least compare and show the percentage of labeled cells in both methods and discuss that the two experiments target (at least slightly) different populations. In addition, the use of TRAP2 for vHP is relatively new; the authors should confirm that this method actually captures negative engram populations by checking for reactivation of these cells during recall by overlap analysis of Fos staining or by artificial activation.

      We thank the reviewer for their comments and opportunity to discuss the marked differences between TRAP2 and DOX systems. In particular, we agree that while both systems rely on the the Fos promoter to drive an effector of interest, their efficacy and temporal resolution vary substantially depending on genetic cell-type, brain region, temporal parameters of Dox or 4-OHT delivery, subject-by-subject metabolic variability, and threshold to Fos induction given the promoter sequences inherent to each system. For example, recent studies have reported the following:

      - The TRAP2 line labels a subset of endogenously activeCA1 pyramidal cells (e.g. 5-18%) while the DOX system labels 20-40% of CA1 pyramidal cells (DeNardo et al, 2019; Monasterio et al, BioRxiv 2024 ).

      - The temporal windows for each range from hours in TRAP2 to 24-48 hours for DOX (DeNardo et al, 2019; Denny et al, 2014; Liu & Ramirez et al, 2012).

      - The efficacy of “tagging” a population of cells with TRAP2 vs with DOX will constrain the number of possible cells that may overlap with cFos upon re-exposure to a given experience (e.g. see the observed overlaps in vCA1 - BLA circuits (Kim & Cho, 2020), compared to vCA1 in general (Ortega-de San Luis et al, 2023) and valence-specific vCA1 populations (Shpokayte et al, 2022).

      - Tagging vCA1 cells with both the TRAP2 and DOX systems are nonetheless sufficient to drive corresponding behaviors (e.g. vCA1 terminal stimulation drives behavioral changes with the DOX and TRAP2 system (Shpokayte et al, 2022) and vCA1 stimulation of an updated fear-linked ensemble drives light-induced freezing in a neutral context utilizing the TRAP2 and DOX systems (Ortega-de San Luis et al, 2023)).

      Finally, and promisingly, as more studies continue to link the in vivo physiological dynamics of these cell populations tagged using each system (e.g. compare Pettit et al, 2022 with Tanaka et al, 2018) and correlating their activity to behavioral phenotypes, our field is in the prime position to uncover deeper principles governing hippocampus-mediated engrams in the brain. Together, we believe a more comprehensive understanding of these systems is fully warranted, especially in the service of further cataloging cellular similarities and differences within such tagged populations.

      (3)  Interpretation of the behavior data

      In Figures 3a and b, the authors show that the experimental group showed higher anxiety based on time spent in the center/open area. However, there were no differences in distance traveled and center entries, which are often reduced in highly anxious mice. Thus, it is not clear what the exact effect of the manipulation is. The authors may want to visualize the trajectories of the mice's locomotion instead of just showing bar graphs.

      Our findings show that our experimental group displays higher levels of anxiety-like behaviors as measured via time spent in center/open area, while there are no differences in distance traveled or center entries. For distance traveled, our interpretation is in line with complementary research (Jimenez et al, 2018; Kheirbek et al, 2013) that shows no changes in distance traveled/distance traveled in the center coupled with changes in anxiety levels as a result of manipulation within anxiety-related circuits. More broadly, any locomotion-related deficit could cause a change in distance traveled that is unrelated to anxiety-like behaviors alone. For example, a reduction in distance traveled could be coupled with a decrease in time spent in the center, but could also result only from motor or exploratory deficits. We hope that this explanation clarifies our interpretation of the open field and elevated plus maze findings in light of other literature.

      In addition, the data shown in Figure 4b is somewhat surprising - the 14MO control showed more freezing than the 6MO control, which can be interpreted as "better memory in old". As this is highly counterintuitive, the authors may want to discuss this point. The authors stated that "Mice typically display increased freezing behavior as they age, so these effects during remote recall are expected" without any reference. This is nonsense, as just above in Figure 4a, older mice actually show less freezing than young mice. Overall, the behavioral effects are rather small and random. I would suggest that these data be interpreted more carefully.

      In Figure 4B, we present our findings from remote recall and observe increased freezing levels in control mice with age, as mentioned by the reviewer, indicating increased memory. This is in line with previous work from Shoji & Miyakawa, 2019 which has been added as a reference for the quotation described above; we thank the reviewer for pointing this error out. As the reviewer has pointed out, above in Figure 4A, we measured freezing levels across all groups during contextual fear conditioning before the start of chronic stimulation, as this was the session we ‘tagged’ a negative memory in. Although it appears that there may be slightly lower levels of freezing in older (14-month old) mice, our findings do not determine statistical significance for difference between age group, only effects of time and subject which are expected as freezing increases within the session and animals display high levels of variability in freezing levels across many experiments (Figure 4A i-iii). We also find in previous work that control mice receiving 3-, 6- and 9-months of chronic DCZ stimulation in the vHPC with empty vector (mCherry) receptor show an increase in freezing with age (Suthard et al, 2023; Figure 2A ii).

      (4) Lack of citation and discussion of relevant study

      Khalaf et al. 2018 from Gräff lab showed that experimental activation of recall-induced populations leads to fear attenuation. Despite the differences in experimental details, the conceptual discrepancy should be discussed.

      As mentioned by the reviewer, Khalaf et al. 2018 showed that experimental activation of recall-induced populations in the dentate gyrus leads to fear attenuation. Specifically, they pose that this fear attenuation occurs in these ensembles through updating or unlearning of the original memory trace via the engagement, rather than suppression, of an original traumatic experience. Despite the differences in experimental details with our current study and this work, we agree that the conceptual discrepancy should be discussed. First, one major difference is that we are reactivating an ensemble that was tagged during fear memory encoding, while Khalaf et al. are activating a remote recall-induced ensemble that was tagged one month after encoding. Although there is high overlap between the encoding and recall ensembles when mice are exposed to the conditioning context, these ensembles are not identical and may result in different behavioral phenotypes when chronically reactivated. Further, Khalaf et al rely on reactivation of the recall-induced ensemble during extinction to facilitate rapid fear attenuation. This differs from our current work, as their reactivation is occurring during the extinction process in the previously conditioned context, while we are reactivating chronically in the animal’s home cage over the course of a longer time period. It may be necessary that the memory is first reactivated, and thus, more liable to re-contextualization, in the original context compared to an unrelated homecage environment where there are presumably no related cues present. Importantly, this previous work tests the attenuation of fear shortly after an extinction process, while we are not traditionally extinguishing the context with aid of the memory reactivation. Finally, we are testing remote recall (3 months post-conditioning), while they are testing at a shorter time interval (28 days). In line with these ideas, future work may seek to tease out the mechanistic differences between recent and remote memory extinction both in terms of natural memory recall and chronically manipulated memory-bearing cells.

      Reviewer #2 (Public Review):

      Summary:

      Jellinger, Suthard, et al. investigated the transcriptome of positive and negative valence engram cells in the ventral hippocampus, revealing anti- and pro-inflammatory signatures of these respective valences. The authors further reactivated the negative valence engram ensembles to assay the effects of chronic negative memory reactivation in young and old mice. This chronic re-activation resulted in differences in aspects of working memory, and fear memory, and caused morphological changes in glia. Such reactivation-associated changes are putatively linked to GABA changes and behavioral rumination.

      Strengths:

      Much of the content of this manuscript is of benefit to the community, such as the discovery of differential engram transcriptomes dependent on memory valence. The chronic activation of neurons, and the resultant effects on glial cells and behavior, also provide the community with important data. Laudable points of this manuscript include the comprehensiveness of behavioral experiments, as well as the cross-disciplinary approach.

      Weaknesses:

      There are several key claims made that are unsubstantiated by the data, particularly regarding the anthropomorphic framing of "rumination" on a mouse model and the role of GABA. The conclusions and inferences in these areas need to be carefully considered.

      (1) There are many issues regarding the arguments for the behavioural data's human translation as "rumination." There is no definition of rumination provided in the manuscript, nor how rumination is similar/different to intrusive thoughts (which are psychologically distinct but used relatively interchangeably in the manuscript), nor how rumination could be modelled in the rodent. The authors mention that they are attempting to model rumination behaviours by chronically reactivating the negative engram ("To understand if our experimental model of negative rumination..."), but this occurs almost at the very end of the results section, and no concrete evidence from the literature is provided to attempt to link the behavioural results (decreased working memory, increased fear extinction times) to rumination-like behaviours. The arguments in the final paragraph of the Discussion section about human rumination appear to be unrelated to the data presented in the manuscript and contain some uncited statements. Finally, the rumination claims seem to be based largely upon a single data figure that needs to be further developed (Figure 6, see also point 2 below).

      (2) The staining and analysis in Figure 6 are challenging to interpret, and require more evidence to substantiate the conclusions of these results. The histological images are zoomed out, and at this resolution, it appears that only the pyramidal cell layer is being stained. A GABA stain should also label the many sparsely spaced inhibitory interneurons existing across all hippocampal layers, yet this is not apparent here. Moreover, both example images in the treatment group appear to have lower overall fluorescence intensity in both DAPI and GABA. The analysis is also unclear: the authors mention "ROIs" used to measure normalized fluorescence intensity but do not specify what the ROI encapsulates. Presumably, the authors have segmented each DAPI-positive cell body and assessed fluorescence however, this is not explicated nor demonstrated, making the results difficult to interpret.

      Based on the collective discussion from all reviewers on the completeness of our GABA quantification and its implications, we have decided to remove this figure and perform more substantive analysis of this E/I imbalance in future work.

      (3) A smaller point, but more specific detail is needed for how genes were selected for GSEA analysis. As GSEA relies on genes to be specified a priori, to avoid a circular analysis, these genes need to be selected in a blind/unbiased manner to avoid biasing downstream results and conclusions. It's likely the authors have done this, but explicitly noting how genes were selected is an important context for this analysis.

      As mentioned in our Methods section, gene sets were selected based on pre-existing biology and understanding of genes canonically involved in “neurodegeneration” such as those related to apoptotic pathways and neuroinflammation or “neuroprotection” such as brain-derived neurotrophic factor, to name a few. A limitation of this method is that we must avoid making strong claims about the actual function of these up- or down-regulated genes without performing proper knock-in or knock-out studies, but we hope that this provides an unbiased inventory for future experiments to perform causal manipulations.

      Reviewer #3 (Public Review):

      Summary:

      The authors note that negative ruminations can lead to pathological brain states and mood/anxiety dysregulation. They test this idea by using mouse engram-tagging technology to label dentate gyrus ensembles activated during a negative experience (fear conditioning). They show that chronic chemogenetic activation of these ensembles leads to behavioral (increased anxiety, increased fear generalization, reduced fear extinction) and neural (increases in neuroinflammation, microglia, and astrocytes).

      Strengths:

      The question the authors ask here is an intriguing one, and the engram activation approach is a powerful way to address the question. Examination of a wide range of neural and behavioral dependent measures is also a strength.

      Weaknesses:

      The major weakness is that the authors have found a range of changes that are correlates of chronic negative engram reactivation. However, they do not manipulate these outcomes to test whether microglia, astrocytes, or neuroinflammation are causally linked to the dysregulated behaviors.

      Recommendations For The Authors:

      Reviewer #1 (Recommendations For The Authors):

      - Figure 2c should include Month0, the BW before the start of the manipulation.

      Regrettably, we do not have access to the Month 0 body weights at this time as this project changed hands over the course of the past year or so. This is an inherent limitation that we missed during analysis and we pose this as a limitation in the Results section after describing this finding. Therefore, it is possible that over the first month of stimulation (Month 0-1), there may have been a drop in body weight that rebounded by the first measurement at Month 1 that continued to increase normally through Months 2-3, as shown in our Figure 1. Thank you for this note.

      - Figure 6a looks confusing - the background signal in the green channel is very different between control and experimental groups. Were representative images taken with different microscope settings?

      The representative images were taken with the same microscope power settings, but were adjusted in brightness/contrast within FIJI for clarity in the Figure – we apologize that this was misleading in any way and thank the reviewer for their feedback. Further, based on the collective discussion from all reviewers on the completeness of our GABA quantification and its implications, we have decided to remove this figure and perform more substantive analysis of this E/I imbalance in future work.

      - Typo mChe;try

      This typo was fixed

      - "During this contextual... mice in the 6- and 14- month groups..." Isn't it 3- and 11- month respectively at the time of fear conditioning? Throughout the manuscript, this point was written very confusingly.

      Yes, we thank the reviewer for pointing this out. It has been corrected to 3- and 11-month old mice at the timing of fear conditioning and clarified throughout the manuscript where applicable.

      - "GABAergic eYFP fluorescence" Where does the eYFP come from? The methods state that GABA quantification is based on IHC staining.

      Based on the collective discussion from all reviewers on the completeness of our GABA quantification and its implications, we have decided to remove this figure and perform more substantive analysis of this

      E/I imbalance in future work. We discuss this E/I balance not being directly assessed in the Limitations & Future Directions section of our Discussion, noting the importance of detailed quantification of both excitatory and inhibitory markers within the hippocampus.

      Reviewer #2 (Recommendations For The Authors):

      (1) There is a full methods section ("Analysis of RNA-seq data") that mostly describes RNA-seq analysis that seemingly does not appear in the paper. This section should be reviewed.

      We have included this portion of the methods that explain the previous workflow from Shpokayte et al., 2022 where this dataset was generated and this has been noted in the “Analysis of RNA-seq data” section of the methods.

      (2) Figure 6: GABA staining should be more critically analyzed, as discussed above, and validated with another GABA antibody for rigor. From the representative images provided in Figure 6, it looks possibly as though the hM3Dq images were simply not fully in the focal plane when being imaged or were over-washed, as DAPI staining also appears to be lower in these images.

      Based on the collective discussion from all reviewers on the completeness of our GABA quantification and its implications, we have decided to remove this figure and perform more substantive analysis of this E/I imbalance in future work. Specifically, it will be necessary to rigorously investigate both excitatory and inhibitory markers within this region to ensure these claims are substantiated. Thank you for this suggestion.

      (3) The first claim that human GABAergic interneurons cause rumination is uncited. (Page 19, first sentence beginning with: "Evidence from human studies suggests...").

      Based on the collective discussion from all reviewers on the completeness of our GABA quantification and its implications, we have decided to remove this figure and perform more substantive analysis of this E/I imbalance in future work. Apologies for the lack of citation in-text, the proper citation for this finding is Schmitz et al, 2017.

      (4) Gene names throughout the manuscript and figure are written in the wrong format for mice (eg: Page 13, second line: SPP1, TTR, and C1QB1 instead of Spp1, Ttr, C1qb1).

      This was corrected throughout the manuscript.

      (5) Tense on Page 15 third sentence of the second paragraph: "...spatial working memory was assessed...".

      This was corrected throughout the manuscript.

      (6) Supplemental Figure 1 would benefit from normalization of the NeuN+ cell counts. The inclusion of an excitatory and inhibitory neuron marker in this figure might benefit the argument that there is a change in the excitation/inhibition of the hippocampus - as the numbers of excitatory neurons outweigh the numbers of inhibitory neurons that would be assayed here.

      In an effort to normalize the NeuN+ cell counts, for each of our ROIs (6-8 single tiles for each brain region (DG, vCA1, vSub) x 3-5 coronal slices = ~18 single tiles per mouse x 3-4 mice) we captured a 300 x 300 micrometer, single-tile z-stack at 20x magnification. These ROIs were matched for dimensions and brain regions across all groups for each hippocampal subregion quantified. We initially proposed to normalize these NeuN counts over DAPI, but because DAPI includes all nuclei (microglia, oligodendrocytes, astrocytes and neurons), we weren’t sure this was the most optimal tool. We do agree that further quantification of excitatory and inhibitory cell markers would be vital to more concrete interpretation of our findings and we have added this to our Limitations & Future Work section of the Discussion.

      Reviewer #3 (Recommendations For The Authors):

      (1) The DOX tagging window lacks temporal precision. I suggest the authors note this as a limitation.

      We thank the reviewer for noting this, and we have added this limitation to the Methods section with the context of the 24-48 hour DOX window being longer than other methods like TRAP.

      (2) Is there a homeostatic response to chronic engram stimulation? That is, is DCZ as effective in increasing neuronal excitability on day 90 as it is on day 1. This could be addressed with electrophysiology, or with IEG induction. Alternatively, the authors could refer to previous literature-- for example, Xia et al (2017) eLife-- that examined whether there was any blunting of the effects of DREADD ligands after sustained delivery via drinking water. There, of course, may be other papers as well.

      As noted by the reviewer, it is important to determine if DCZ maintains its effects on neuronal excitability throughout the 3 month administration period. To address this, previous work has shown that CNO administration in drinking water over one month consistently inhibited hM4Di+ neurons without altering baseline neuronal excitability as measured by firing rate and potassium currents (Xia et al, 2017). Although this is only for one month, it is administered via the same oral route as our DCZ protocol and suggests that at least for that amount of time we are likely producing consistent effects. In our reply above to Reviewer #1’s comment, we also note that even if DCZ is only having an effect for one month, rather than 3 months, we are still observing enduring changes that resulted from this short-term disturbance.

      (3) Please double check there is no group effect on weight in 6-month-old mice in Figure 2C.

      Two-way RM ANOVA showed no main effect of Group within the 6-month-old control and hM3Dq groups.

      Group: F(1,17) = 1.361, p=0.2594.

      (4) The shock intensity is much higher than is typical for fear conditioning studies in mice. Why was this the case?

      Yes, we do agree that this shock intensity is on the higher side of typical paradigms in mice, however, our lab has utilized 0.75mA to 1.5mA intensity foot shocks for contextual fear conditioning in the past (Suthard & Senne et al, 2023; 2024; Dorst & Senne et al, 2023; Grella et al., 2022; Finkelstein et al., 2022) and we maintained this protocol for internal consistency. However, it would be interesting to systematically investigate how differing intensities of foot shock, subsequent tagging of this ensemble and reactivation would uniquely impact behavioral state acutely and chronically in mice.

      (5) Remote freezing is very low. The authors should comment on this-- perhaps repeated testing has led to some extinction?

      A reviewer above suggested a similar phenomenon may be occuring, specifically fear attenuation as a result of chronic stimulation. They referenced previous work from Khalaf et al. 2018, where they reactivated a recall-induced ensemble, while we reactivated an ensemble tagged during encoding. We expand upon this work in light of our findings within the Limitations & Future Work section of our Discussion. However, we do appreciate the lower levels of freezing observed in remote recall and sought out other literature to understand the typical range of remote freezing levels. One thing that we note is that our remote recall is occurring 3 months after conditioning, which is much longer than typical 14-28 day protocols. However, we find that freezing levels at remote timepoints from 21-45 days results in contextual freezing levels of between 20-50% approximately (Kol et al., 2020), as well as 40-75% approximately in a variety of 28 day remote recall experiments (Lee et al., 2023). This information, together with our current experimental protocol demonstrates a wide range of remote freezing levels that may depend heavily on the foot shock intensity, duration of days after conditioning, and animal variability.

      (6) "mice display increased freezing with age": please add a reference.

      Apologies, we missed the citation for that claim and it has been added in-text and in the references list (Shoji & Miyakawa, 2019).

      (7) Related to the low freezing levels for remote memory, why is generalization minimal? Many studies have shown that there is a time-dependent emergence of generalized fear, yet here this is not seen. Is it linked to extinction (as above)? Or genetic background?

      Previous work has shown that rats receiving multiple foot shocks during conditioning displayed a time-dependent generalization of context memory, while those receiving less shocks did not (Poulos et al., 2016), as the reviewer noted in their comment. In our current study, we observe low levels of generalization in all of our groups compared to freezing levels displayed in the conditioned context at the remote timepoint, in opposition to this time-dependent enhancement of generalization. It is possible that the genetic background of our C57BL/6J mice compared to the Long-Evans rat strain in this previous work accounts for some of this difference. In addition, it is possible that the longer duration of time (3 months) compared to their remote timepoint (28 days) resulted in time-dependent decrease in generalization that decreases with greater durations of time from original conditioning. As noted above, it is indeed plausible that the reactivation of a contextual fear ensemble over time is attenuating freezing levels for both the original and similar contexts (Khalaf et al, 2018). We discuss the differences in our study and this 2018 work more comprehensively above.

      (8) Morphological phenotypes of astrocytes/microglia. Would be great to do some transcriptomic profiling of microglia/astrocytes to couple with the morphological characterization (but appreciate this is beyond the scope of current work).

      We thank the reviewer this suggestion, we agree that would be an incredibly informative future experiment and have added this to our Limitations & Future Experiments section of the Discussion.

      (9) The authors could consider including a limitations section in their discussion which discusses potential future directions for this work:

      - causal experiments.

      - E/I balance is not assessed directly (interestingly, in this regard, expanded engrams are linked to increased generalization [e.g., Ramsaran et al 2023]).

      Thank you for this suggestion, we have added a Limitations & Future Directions section to our Discussion and have expanded upon these suggested points.

      (10) For Figure 10, consider adding an experimental design/timeline.

      We are making the assumption that the reviewer meant Figure 1 instead of Figure 10 here, but note that there is a description of the viral expression duration (D0-D10), followed by an off Dox period of 48 hours (D10-D12), with subsequent engram tagging of a negative (foot shock) or positive (male-to-female exposure) on D12. In our experiments (Shpokayte et al., 2022), Dox was administered for 24 hours (D12-D13), which was followed by sacrificing the animal for cell suspension and sequencing of the positive and negative engram populations. This figure also shows the viral strategy for the Tet-tag system (Figure 1A), as well as representative viral expression in vHPC (Figure 1B). We are happy to add additional experimental design/timeline information to this figure that would be helpful to the reviewer.

    1. Python-based workflow for processing linked-read sequencing datasets, thereby filling the current gap in the field caused by platform-centric genome-specific linked-read data analysis tools.

      Reviewer 3: Dmitrii Meleshko The paper titled "LRTK: A Platform-Agnostic Toolkit for Linked-Read Analysis of Both Human Genomes and Metagenomes" by Yang et al. is dedicated to the development of a unified interface for linked-read data processing.The problem described in the paper indeed exists; each linked-read technology requires complex preprocessing steps that are not straightforward or efficient. The idea of consolidating multiple tools in one place, with some of them modified to handle multiple data types, is commendable. Overall, I am supportive of this paper. My main concern, however, is that the impact of linked-read applications in the paper appears to be exaggerated, and the authors need to provide more context in their presentation. Also, some parts of the paper are vague described. I will elaborate on my concerns in more detail below.X) Linked-read sequencing generates reads with high base quality and extrapolative 64 information on long-range DNA connectedness, which has led to significant 65 advancements in human genome and metagenome research[1-3]. - Citations 1-3 do not really tell about advancements in human genome and metagenome research, these are technologies papers. Similar problem can be found in "Despite the limitations that genome specificity…" paragraph. Authors cited and described several algorithms, that are not really genomic studies. E.g. "stLFR[2] has found application in a customized pipeline that has been developed to first convert its raw reads into a 10x-compatible format, after which Long Ranger is applied for downstream analysis." is not an example of genomic study, but a pipeline description.X) Table S1 does not improve the paper, I would say it does completely the opposite. LongRanger is not a toolkit, it should be considered as read alignment tool that outputs some SVs and haplotypes along the way. So LongRanger vs LRTK comparison does not make sense to me. There are other tools that solve metagenome assembly problem, human assembly problem, call certain classes of SVs etc.x) I think incorporating longranger is important, since its performance is reported to be better than EMA for human samples and it is also more popular than EMA. Is it possible and have you tried doing it?x) I would remove exaggerations such as "myriad" from the text. The scope of linked-reads is pretty limited nowadays. I agree that linked-reads might be useful in metagenomics/transcriptomics and other scenarios that were mentionedin the text, but the number of studies is very limited especially nowadays, and was not really big when 10X platform was on the risex) "LRTK reconstructs long DNA fragments" - when people talk about long fragment reconstruction, they usually mean moleculo-style reconstruction through assembly. This reconstruction resemble "barcode deconvolution", described in Danko et al, and Mak et al. So I would stick to this terminologyx) it is important to note that, Aquila, LinkedSV and VALOR2 are linked-read specific tools, while FreeBayes, Samtools and GATK are short-read tools. Also, provide target SV length for both groups of tools.x) There are some minor problems with Github readme. E.g. "*parameters". Also, I don't understand how to use conversion in real life… E.g. 10X Genomics data often comes as a folder with multiple gzipped R1/R2/I1 files. I don't understand how would I use it in that case.x) Please cite or explain why this is happening (not only when) - "A known concern with stLFR linked-read sequencing is the loss of barcode specificity during analysis."x) I don't understand what is "Length-weighted average (μFL) and unweighted average (WμFL) of DNA 688 fragment lengths." from the figure. One of them is just an average and what about second? Figure looks confusingx) LRTK supports reconstruction of long DNA fragments - this section describes something else. More about statistics and data QCx) LRTK promotes metagenome assembly using barcode specificity - please remove supernova, it was never a metagenomic assembler. Check cloudSPAdes insteadx) "The superior assembly performance we have observed" - superior compared to what? If so, some short-read benchmark should be included.x) "LRTK improves human genome variant phasing using long range information" - What dataset is this? What callset was used for ground truth? Briefly describe how comparisons were done?x) Figures 5F-G together are very confusing.First I don't expect tools like LinkedSV to have high recall (around 1.0) and low precision. Also, figure G is kind of subset of figure F, but results are completely different. Also use explicit notation. E.g. 50-1kbp and 1-10kbp mean completely different thingsx) We curated one benchmarking dataset and two real datasets to demonstrate the 307 performance of LRTK - what do you mean by "curation" herex) Why don't you use Tell-Seq barcode whitelist mentioned here - https://sagescience.com/wpcontent/uploads/2020/10/TELL-Seq-Software-Roadmap-User-Guide-2.pdfx) Tiered alignment approach is vaguely introduced. It is not clear what "n% most closely covered windows." mean, or how do we select a subset of reference genomes for the second phase

    2. benchmarking and three real linked-read data sets from both the human genome and metagenome. We showcase LRTK’s ability to generate comparative performance results from the preceding benchmark study and to report these results in publication-ready HTML document plots. LRTK provides comprehensive and flexible modules along with an easy-to-use

      Reviewer 2: Lauren Mak Summary: This manuscript describes the need for a generalized linked-read (LR) analysis package and showcases the package the authors developed to address this need. Overall, the workflow is welldesigned but there are major gaps in the benchmarking, analysis, and documentation process that need to be addressed before publication.Documentation:The purpose of multiple tool options: While the analysis package is technically sound, one major aspect is left unexplained- why are there so many algorithm options included without guidance as to which one to use? There are clearly performance differences by different algorithms (combinations of 2+ not considered either) on different types of LR sequence.Provenance of ATCC-MSA-1003: Nowhere in the manuscript is the biological and technical composition of the metagenomics control described. It would be helpful to mention that this is specifically a mock gut microbiome sample, as well as the relative abundances of the originating species as well as the absolute amounts of genetic material per species (ex. as measured by genomic coverage) in the actual dataset. As a corollary, there should be standard deviations in any figures that display a summary statistic (ex. Figure 3A- precision, recall, etc.) that seems to be averaged across the species in a sample. This includes Figure 3A and Figure 4A.Dataset details: There is no table indicating the number of reads for each dataset, which would be helpful in interpreting Figures 3 and 4.Open source?: However, there was no Github link provided, only a link to the Conda landing page. Are there thorough instructions provided for the package's installation, input, output, and environment management?Benchmarking:The lack of simulated tests: The above concern (expected performance on idealized datasets) is best addressed with simulated data, which was not done despite the fact that LRSim exists (and apparently the authors have written a tool for stLFR as well previously).Indels: What are the sizes of the indels detected? Why were newer tools, such as PopIns2, Pamir, or Novel-X not tried as well?Analysis:Lines 166-169: Figure 1 panel A1 vs. B1- why do the distribution of estimated fragment sizes from the 10x datasets look so different in metagenomic vs. human samplees, when there is reasonable consistency in TELL-Seq and stLFR datasets?Lines 182-184: Figure 3A- why is LRTK's taxonomic classification quality generally lower than the of the tools? At least in terms of recall, it should perform better as mapping reads to reference genomes should have a lower false negative rate than k-mer-based tools. Also, what is the threshold for having detect a taxon? Is it just any number of reads or is there a minimum bound?Lines 187-188: Figure 3B- at least 15% of each caller's set of variants is unique to the variant, while a maximum of 50% is universal. I'd not interpret that as consistency.Lines 192-193: Are you referring to allelic imbalance as it is popularly used to refer to expression variation between the two haplotypes of a diploid organism? This clearly doesn't apply in the case of bacteria. If this is not what you're referring to, please define and/or cite the applicable definition.Lines 201-208: It's odd that despite the 10x datasets having the largest estimated fragment size, they have some of the smallest genome fractions, NGA50, and NA50. Why is this? Are they just smaller datasets, on average?Miscellaneous:UHGG: Please mention the fact that the UHGG is the default database, as well as whether or not the user will be able to supply their own databases.Line 363: What does {M} refer to?Line 369: What does U mean here? Is this the number of uniquely aligned reads in one of the windows N that a multi-aligned read aligns to?Lines 371-372: What does 'n% most closely covered windows' refer to?Lines 399-405: How are SNVs chosen for MAI analysis from the three available SNV callers?Lines 653-656: Which dataset was used for quality evaluation?Line 665: What do the abbreviations BAF and T stand for?

  6. Jun 2024
    1. Author response:

      The following is the authors’ response to the original reviews.

      Public Reviews:

      Reviewer #1 (Public Review):

      Summary:

      This study reports that IT neurons have biased representations toward low spatial frequency

      (SF) and faster decoding of low SFs than high SFs. High SF-preferred neurons, and low SF-preferred neurons to a lesser degree, perform better category decoding than neurons with other profiles (U and inverted U shaped). SF coding also shows more sparseness than category coding in the earlier phase of the response and less sparseness in the later phase. The results are also contrasted with predictions of various DNN models.

      Strengths:

      The study addressed an important issue on the representations of SF information in a high-level visual area. Data are analyzed with LDA which can effectively reduce the dimensionality of neuronal responses and retain category information.

      We would like to express our sincere gratitude for your insightful and constructive comments which greatly contributed to the refinement of the manuscript. We appreciate the time and effort you dedicated to reviewing our work and providing suggestions. We have carefully considered each of your comments and addressed the suggested revisions accordingly.

      Weaknesses:

      The results are likely compromised by improper stimulus timing and unmatched spatial frequency spectrums of stimuli in different categories.

      The authors used a very brief stimulus duration (35ms), which would degrade the visual system's contrast sensitivity to medium and high SF information disproportionately (see Nachmias, JOSAA, 1967). Therefore, IT neurons in the study could have received more degraded medium and high SF inputs compared to low SF inputs, which may be at least partially responsible for higher firing rates to low SF R1 stimuli (Figure 1c) and poorer recall performance with median and high SF R3-R5 stimuli in LDA decoding. The issue may also to some degree explain the delayed onset of recall to higher SF stimuli (Figure 2a), preferred low SF with an earlier T1 onset (Figure 2b), lower firing rate to high SF during T1 (Figure 2c), somewhat increased firing rate to high SF during T2 (because weaker high SF inputs would lead to later onset, Figure 2d).

      We appreciate your concern regarding the course-to-fine nature of SF processing in the vision hierarchy and the short exposure time of our paradigm. According to your comment, we repeated the analysis of SF representation with 200ms exposure time as illustrated in Appendix 1 - Figure 4. Our recorded data contains the 200ms version of exposure time for all neurons in the main phase. As can be seen, the results are similar to what we found with 33 ms experiments.

      Next, we bring your attention to the following observations:

      (1) According to Figure 2d, the average firing rate of IT neurons for HSF could be higher than LSF in the late response phase. Therefore, the amount of HSF input received by the IT neurons is as much as LSF, however, its impact on the IT response is observable in the later phase of the response. Thus, the LSF preference is because of the temporal advantage of the LSF processing rather than contrast sensitivity.

      (2) According to Figure 3a, 6% of the neurons are HSF-preferred and their firing rate in HSF is comparable to the LSF firing rate in the LSF-preferred group. This analysis is carried out in the early phase of the response (70-170 ms). While most of the neurons prefer LSF, this observation shows that there is an HSF input that excites a small group of neurons. Furthermore, the highest separability index also belongs to the HSF-preferred profile in the early phase of the response which supports the impact of the HSF part of the input.

      (3) Similar LSF-preferred responses are also reported by Chen et al. (2018) (50ms for SC) and Zhang et al. (2023) (3.5 - 4 secs for V2 and V4) for longer duration times.

      Our results suggest that the LSF-preferred nature of the IT responses in terms of firing rate and recall, is not due to the weakness or lack of input source (or information) for HSF but rather to the processing nature of the SF in the vision hierarchy.

      To address this issue in the manuscript:

      Figure Appendix 1 - Figure 4 is added to the manuscript and shows the recall value and onset for R1-R5 with 200ms of exposure time.

      We added the following description to the discussion:

      “To rule out the degraded contrast sensitivity of the visual system to medium and high SF information because of the brief exposure time, we repeated the analysis with 200ms exposure time as illustrated in Appendix 1 - Figure 4 which indicates the same LSF-preferred results. Furthermore, according to Figure 2, the average firing rate of IT neurons for HSF could be higher than LSF in the late response phase. It indicates that the amount of HSF input received by the IT neurons in the later phase is as much as LSF, however, its impact on the IT response is observable in the later phase of the response. Thus, the LSF preference is because of the temporal advantage of the LSF processing rather than contrast sensitivity. Next, according to Figure 3(a), 6\% of the neurons are HSF-preferred and their firing rate in HSF is comparable to the LSF firing rate in the LSF-preferred group. This analysis is carried out in the early phase of the response (70-170ms). While most of the neurons prefer LSF, this observation shows that there is an HSF input that excites a small group of neurons. Additionally, the highest SI belongs to the HSF-preferred profile in the early phase of the response which supports the impact of the HSF part of the input. Similar LSF-preferred responses are also reported by Chen et. al. (2018) (50ms for SC) and Zhang et. al. (2023) (3.5 - 4 secs for V2 and V4). Therefore, our results show that the LSF-preferred nature of the IT responses in terms of firing rate and recall, is not due to the weakness or lack of input source (or information) for HSF but rather to the processing nature of the SF in the IT cortex.”

      Figure 3b shows greater face coding than object coding by high SF and to a lesser degree by low SF neurons. Only the inverted-U-shaped neurons displayed slightly better object coding than face coding. Overall the results give an impression that IT neurons are significantly more capable of coding faces than coding objects, which is inconsistent with the general understanding of the functions of IT neurons. The problem may lie with the selection of stimulus images (Figure 1b). To study SF-related category coding, the images in two categories need to have similar SF spectrums in the Fourier domain. Such efforts are not mentioned in the manuscript, and a look at the images in Figure 1b suggests that such efforts are likely not properly made. The ResNet18 decoding results in Figure 6C, in that IT neurons of different profiles show similar face and object coding, might be closer to reality.

      Because of the limited number of stimuli in our experiments, it is hard to discuss the category selectivity, which needs a higher number of stimuli. To overcome the limited number of stimuli in our experiment, we fixed 60% (nine out of 15 stimuli) while varying the remaining stimuli to reduce the selective bias. To check the coding capability of the IT neurons for face and non-face objects, we evaluated the recall of face vs. non-face classification in intact stimuli (similar to classifiers stated in the manuscript). Results show that at the population level, the recall value for objects is 90.45%, and for faces is 92.45%. However, the difference is not significant (p-value=0.44). On the other hand, we note that a large difference in the SI value does not translate directly to the classification accuracy, rather it illustrates the strength of representation.

      Regarding the SF spectrums, after matching the luminance and contrast of the images we matched the power of the images concerning SF and category. Powers are calculated using the sum of the absolute value of the Fourier transform of the image. Considering all stimuli, the ANOVA analysis shows that various SF bands have similar power (one-way ANOVA, p-value=0.24). Furthermore, comparing the power of faces and images in all SF bands (including intact) and both unscrambled and scrambled images indicates no significant difference between face and object (p-vale > 0.1). Therefore, the result of Figure 3b suggests that IT employs various SF bands for the recognition of various objects.

      Comparing the results of CNNs and IT shows that the CNNs do not capture the complexities of the IT cortex in terms of SF. One of the sources of this difference is because of the behavioral saliency of the face stimulus in the training of the primate visual system.

      To address this issue in the manuscript:

      The following description is added to the discussion:

      “… the decoding performance of category classification (face vs. non-face) in intact stimuli is 94.2%. The recall value for objects vs. scrambled is 90.45%, and for faces vs. scrambled is 92.45% (p-value=0.44), which indicates the high level of generalizability and validity characterizing our results.”

      The following description is added to the method section, SF filtering.

      “Finally, we equalized the stimulus power in all SF bands (intact, R-R5). The SF power among all conditions (all SF bands, face vs. non-face and unscrambled vs. scrambled) does not vary significantly (p-value > 0.1). SF power is calculated as the sum of the square value of the image coefficients in the Fourier domain.”

      Reviewer #2 (Public Review):

      Summary:

      This paper aimed to examine the spatial frequency selectivity of macaque inferotemporal (IT) neurons and its relation to category selectivity. The authors suggest in the present study that some IT neurons show a sensitivity for the spatial frequency of scrambled images. Their report suggests a shift in preferred spatial frequency during the response, from low to high spatial frequencies. This agrees with a coarse-to-fine processing strategy, which is in line with multiple studies in the early visual cortex. In addition, they report that the selectivity for faces and objects, relative to scrambled stimuli, depends on the spatial frequency tuning of the neurons.

      Strengths:

      Previous studies using human fMRI and psychophysics studied the contribution of different spatial frequency bands to object recognition, but as pointed out by the authors little is known about the spatial frequency selectivity of single IT neurons. This study addresses this gap and they show that at least some IT neurons show a sensitivity for spatial frequency and

      interestingly show a tendency for coarse-to-fine processing.

      We extend our sincere appreciation for your thoughtful and constructive feedback on our paper. We are grateful for the time and expertise you invested in reviewing our work. Your detailed suggestions have been instrumental in addressing several key aspects of the paper, contributing to its clarity and scholarly merit. We have carefully considered each of your comments and have made revisions accordingly.

      Weaknesses and requested clarifications:

      (1) It is unclear whether the effects described in this paper reflect a sensitivity to spatial frequency, i.e. in cycles/ deg (depends on the distance from the observer and changes when rescaling the image), or is a sensitivity to cycles /image, largely independent of image scale. How is it related to the well-documented size tolerance of IT neuron selectivity?

      Our stimuli are filtered using cycles/images and knowing the distance of the subject to the monitor, we can calculate the cycles/degrees. To the best of our knowledge, this is also the case for all other SF-related studies. To find the relation of observations to the cycles/image and degree/image, one should keep one of them fixed while changing the other, for example changing the subject's distance to the monitor will change the SF content in terms of cycle/degree. With our current data, we cannot discriminate this effect. To address this issue, we added the following description to the discussion. To address this issue, we added the following description to the discussion:

      “Finally, since our experiment maintains a fixed SF content in terms of both cycles per degree and cycles per image, further experiments are needed to discern whether our observations reflect sensitivity to cycles per degree or cycles per image.”

      (2) The authors' band-pass filtered phase scrambled images of faces and objects. The original images likely differed in their spatial frequency amplitude spectrum and thus it is unclear whether the differing bands contained the same power for the different scrambled images. If not, this could have contributed to the frequency sensitivity of the neurons.

      After equalizing the luminance and contrast of the images, we equilized their power concerning SF and category. The powers were calculated using the sum of the absolute values of the Fourier transform of the images. The results of the ANOVA analysis across all stimuli indicate that various SF bands exhibit similar power (one-way ANOVA, p-value = 0.24). Additionally, a comparison of power between faces and objects in all SF bands (including intact), for both unscrambled and scrambled images, reveals no significant differences (p-value > 0.1). To clarify this point, we have incorporated the following information into the Methods section.

      “Finally, we equalized the stimulus power in all SF bands (intact, R-R5). The SF power among all conditions (all SF bands, face vs. non-face and unscrambled vs. scrambled) does not vary significantly (ANOVA, p-value > 0.1).”

      (3) How strong were the responses to the phase-scrambled images? Phase-scrambled images are expected to be rather ineffective stimuli for IT neurons. How can one extrapolate the effect of the spatial frequency band observed for ineffective stimuli to that for more effective stimuli, like objects or (for some neurons) faces? A distribution should be provided, of the net responses (in spikes/s) to the scrambled stimuli, and this for the early and late windows.

      The sample neuron in Figure 1c is chosen to be a good indicator of the recorded neurons. In the early response phase, the average firing rate to scrambled stimuli is 26.3 spikes/s which is significantly higher than the response in -50 to 50ms which is 23.4. In comparison, the mean response to intact face stimuli is 30.5 spikes/s, while object stimuli elicit an average response of 28.8 spikes/s. Moving to the late phase, T2, the responses to scrambled, face, and object stimuli are 19.5, 19.4, and 22.4 spikes/s, respectively. Moreover, when the classification accuracy for SF exceeds chance levels, it indicates a significant impact of SF bands on the IT response. This raises a direct question about the explicit coding for SF bands in the IT cortex observed for ineffective stimuli and how it relates to complex and effective stimuli, such as faces. To show the strength of neuron responses to the SF bands in scrambled images, We added Appendix 1 - Figure 2 and also added Appendix 1 - Figure 1, according to comment 4, which shows the average and std of the responses to all SF bands. The following description is added to the results section.

      “Considering the strength of responses to scrambled stimuli, the average firing rate in response to scrambled stimuli is 26.3 Hz, which is significantly higher than the response observed between -50 and 50 ms, where it is 23.4 Hz (p-value=3x10-5). In comparison, the mean response to intact face stimuli is 30.5 Hz, while non-face stimuli elicit an average response of 28.8 Hz. The distribution of neuron responses for scrambled, face, and non-face in T1 is illustrated in Appendix 1 - Figure 2.

      […]

      Moreover, the average firing rates of scrambled, face, and non-face stimuli are 19.5 Hz, 19.4 Hz, and 22.4 Hz, respectively. The distribution of neuron responses is illustrated in Appendix 1 Figure 2.”

      (4) The strength of the spatial frequency selectivity is unclear from the presented data. The authors provide the result of a classification analysis, but this is in normalized units so that the reader does not know the classification score in percent correct. Unnormalized data should be provided. Also, it would be informative to provide a summary plot of the spatial frequency selectivity in spikes/s, e.g. by ranking the spatial frequency bands for each neuron based on half of the trials and then plotting the average responses for the obtained ranks for the other half of the trials. Thus, the reader can appreciate the strength of the spatial frequency selectivity, considering trial-to-trial variability. Also, a plot should be provided of the mean response to the stimuli for the two analysis windows of Figure 2c and 2d in spikes/s so one can appreciate the mean response strengths and effect size (see above).

      The normalization of the classification result is just obtained by subtracting the chance level, which is 0.2, from the whole values. Therefore the values could still be interpreted in percent as we did in the results section. To make this clear, we removed the “a.u.” from the figure and we added the following description to the results section.

      “The accuracy value is normalized by subtracting the chance level (0.2).”

      Regarding the selectivity of the neuron, as suggested by your comment, we added a new figure in the appendix section, Appendix 1 - figure 2. This figure shows the strength of SF selectivity, considering trial-to-trial variability. The following description is added to the results section:

      “The strength of SF selectivity, considering the trial-to-trial variability is provided in Appendix 1 Figure 2, by ranking the SF bands for each neuron based on half of the trials and then plotting the average responses for the obtained ranks for the other half of the trials.”

      The firing rates of Figures 2c and 2d are normalized for better illustration since the variation in firing rates is high across neurons, as can be observed in Figure Appendix 1 - Figure 1. Since we seek trends in the response, the absolute values are not important (since the baseline firing rates of neurons are different), but the values relative to the baseline firing rate determine the trend. To address the mean response and the strength of the SF response, the following description is added to the results section.

      “Considering the strength of responses to scrambled stimuli, the average firing rate in response to scrambled stimuli is 26.3 Hz, which is significantly higher than the response observed between -50 and 50 ms, where it is 23.4 Hz (p-value=3x10-5). In comparison, the mean response to intact face stimuli is 30.5 Hz, while non-face stimuli elicit an average response of 28.8 Hz. The distribution of neuron responses for scrambled, face, and non-face in T1 is illustrated in Appendix 1 - Figure 2.

      […]

      Moreover, the average firing rates of scrambled, face, and non-face stimuli are 19.5 Hz, 19.4

      Hz, and 22.4 Hz, respectively. The distribution of neuron responses is illustrated in Appendix 1 Figure 2.”

      Furthermore, we added a figure, Appendix 1 - Figure 3, to illustrate the strength of SF selectivity in our profiles. The following is added to the results section:

      “To check the robustness of the profiles, considering the trial-to-trial variability, the strength of SF selectivity in each profile is provided in Appendix 1 - Figure 3, by forming the profile of each neuron based on half of the trials and then plotting the average SF responses with the other

      half of the trials.”

      (5) It is unclear why such brief stimulus durations were employed. Will the results be similar, in particular the preference for low spatial frequencies, for longer stimulus durations that are more similar to those encountered during natural vision?

      Please refer to the first comment of Reviewer 1.

      (6) The authors report that the spatial frequency band classification accuracy for the population of neurons is not much higher than that of the best neuron (line 151). How does this relate to the SNC analysis, which appears to suggest that many neurons contribute to the spatial frequency selectivity of the population in a non-redundant fashion? Also, the outcome of the analyses should be provided (such as SNC and decoding (e.g. Figure 1D)) in the original units instead of undefined arbitrary units.

      The population accuracy is approximately 5% higher than the best neuron. However, we have no reference to compare the effect size (the value is roughly similar for face vs object while the chance levels are different). However, as stated in Methods, SNC is calculated for two label modes (LSF and HSF) and it can not be directly compared to the best neuron accuracy. Regarding the unit of SNC, it can be interpreted directly to percent by multiplying by a factor of 100. We removed the “a.u.” to prevent misunderstanding and modified the results section for clearance.

      “… SNC score for SF (two labels, LSF (R1 and R2) vs. HSF (R4 and R5)) and category … (average SNC for SF=0.51\%±0.02 and category=0.1\%±0.04 …”

      (7) To me, the results of the analyses of Figure 3c,d, and Figure 4 appear to disagree. The latter figure shows no correlation between category and spatial frequency classification accuracies while Figure 3c,d shows the opposite.

      In Figure 3c,d, following what we observed in Figure 3a,b about the category coding capabilities in the population of neurons based on the profile of the single neurons, we tested a similar idea if the coding capability of single neurons in SF/category could predict the coding capability of population neurons in terms of category/SF. Therefore, both analyses investigate a relation between a characteristic of single neurons and the coding capability of a population of similar neurons. On the other hand, in Figure 4, the idea is to check the characteristics of the coding mechanisms behind SF and category coding. In Figure 4a, we check if there exists any relation between category and SF coding capability within a single neuron activity without the impact of other neurons, to investigate the idea that SF coding may be a byproduct of an object recognition mechanism. In Figure 4b, we investigated the contribution of all neurons in population decision, again to check whether the mechanisms behind the SF and category coding are the same or not. This analysis shows how individual neurons contribute to SF or category coding at the population level. Therefore, the experiments in Figures 3 and 4 are different in the analysis method and what they were designed to investigate and we cannot directly compare the results.

      (8) If I understand correctly, the "main" test included scrambled versions of each of the "responsive" images selected based on the preceding test. Each stimulus was presented 15 times (once in each of the 15 blocks). The LDA classifier was trained to predict the 5 spatial frequency band labels and they used 70% of the trials to train the classifier. Were the trained and tested trials stratified with respect to the different scrambled images? Also, LDA assumes a normal distribution. Was this the case, especially because of the mixture of repetitions of the same scrambled stimulus and different scrambled stimuli?

      In response to your inquiry regarding the stratification of trials, both the training and testing data were representative of the entire spectrum of scrambled images used in our experiment. To address your concern about the assumption of a normal distribution, especially given the mixture of repetitions of the same scrambled stimulus and different stimuli, our analysis of firing rates reveals a slightly left-skewed normal distribution. While there is a deviation from a perfectly normal distribution, we are confident that this skewness does not compromise the robustness of the LDA classifier.

      (9) The LDA classifiers for spatial frequency band (5 labels) and category (2 labels) have different chance and performance levels. Was this taken into account when comparing the SNC between these two classifiers? Details and SNC values should be provided in the original (percent difference) instead of arbitrary units in Figure 5a. Without such details, the results are impossible to evaluate.

      For both SNC and CMI calculations in SF, we considered two labels of HSF (R4 and R5) and LSF (R1 and R2). This was mentioned in the Methods section, after equation (5). According to your comment, to make it clear in the results section, we also added this description to the results section.

      “… illustrates the SNC score for SF (two labels, LSF (R1 and R2) vs. HSF (R4 and R5)) and category (face vs. non-face) … conditioned on the label, SF (LSF (R1 and R2) vs. HSF (R4 and R5)) or category, to assess the information.”

      The value of SNC can also be directly converted to the percent by a factor of 100. To make it clear, we removed “a.u.” from the y-axis.

      (10) Recording locations should be described in IT, since the latter is a large region. Did their recordings include the STS? A/P and M/L coordinate ranges of recorded neurons?

      We appreciate your suggestion for the recording location. Nevertheless, given the complexities associated with neurophysiological recordings and the limitations imposed by our methodologies, we face challenges in precisely localizing every unit if they are located in STS or not. To address your comment, We added Appendix 1 - Figure 5 which shows the SF and category coding capability of neurons along their recorded locations.

      (11) The authors should show in Supplementary Figures the main data for each of the two animals, to ensure the reader that both monkeys showed similar trends.

      We added Appendix 2 which shows the consistency of the main results in the two monkeys.

      (12) The authors found that the deep nets encoded better the spatial frequency bands than the IT units. However, IT units have trial-to-trial response variability and CNN units do not. Did they consider this when comparing IT and CNN classification performance? Also, the number of features differs between IT and CNN units. To me, comparing IT and CNN classification performances is like comparing apples and oranges.

      Deep convolutional neural networks are currently considered the state-of-the-art models of the primate visual pathway. However, as you mentioned and based on our results, they do not yet capture various complexities of the visual ventral stream. Yet studying the similarities and differences between CNN and brain regions, such as the IT cortex, is an active area of research, such as:

      a. Kubilius, Jonas, et al. "Brain-like object recognition with high-performing shallow recurrent ANNs." Advances in neural information processing systems 32 (2019).

      b. Xu, Yaoda, and Maryam Vaziri-Pashkam. "Limits to visual representational correspondence between convolutional neural networks and the human brain." Nature Communications, 12.1 (2021).

      c. Jacob, Georgin, et al. "Qualitative similarities and differences in visual object representations between brains and deep networks." Nature Communications, 12.1 (2021).

      Therefore, we believe comparing IT and CNN, despite all of the differences in terms of their characteristics, can help both fields grow faster, especially in introducing brain-inspired networks.

      (13) The authors should define the separability index in their paper. Since it is the main index to show a relationship between category and spatial frequency tuning, it should be described in detail. Also, results should be provided in the original units instead of undefined arbitrary units. The tuning profiles in Figure 3A should be in spikes/s. Also, it was unclear to me whether the classification of the neurons into the different tuning profiles was based on an ANOVA assessing per neuron whether the effect of the spatial frequency band was significant (as should be done).

      Based on your comment, we added the description of the separability index to the methods section. However, since the separability index is defined as the division of two dispersion matrices, it has no units by nature. The tuning profiles in Figure 3a are normalized for better illustration since the variation in firing rates is high. Since we seek trends in the response, the absolute values are not important. Regarding the SF profile formation, to better present the SF profile assignment, we updated the method section. Furthermore, The strength of responses for scrambled stimuli can be observed in Appendix 1 - Figures 1 and 2.

      (14) As mentioned above, the separability analysis is the main one suggesting an association between category and spatial frequency tuning. However, they compute the separability of each category with respect to the scrambled images. Since faces are a rather homogeneous category I expect that IT neurons have on average a higher separability index for faces than for the more heterogeneous category of objects, at least for neurons responsive to faces and/or objects. The higher separability for faces of the two low- and high-pass spatial frequency neurons could reflect stronger overall responses for these two classes of neurons. Was this the case? This is a critical analysis since it is essential to assess whether it is category versus responsiveness that is associated with the spatial frequency tuning. Also, I do not believe that one can make a strong claim about category selectivity when only 6 faces and 3 objects (and 6 other, variable stimuli; 15 stimuli in total) are employed to assess the responses for these categories (see next main comment). This and the above control analysis can affect the main conclusion and title of the paper.

      We appreciate your concern regarding category selectivity or responsiveness of the SF profiles. First, we note that we used SI since it overcomes the limitations of the accuracy and recall metrics as they are discrete and can be saturated. Using SI, we cannot directly calculate face vs object with SI, since this index only reports one value for the whole discrimination task. Therefore, we have to calculate the SI for face/object vs scrambled to obtain a value per category. However, as you suggested, it raises the question of whether we assess how well the neural responses distinguish between actual images (faces or objects) and their scrambled versions or if we just assess the responsiveness. Based on Figure 3b, since we have face-selective (LSF and HSF preferred profiles), object-selective (inverse U), and the U profile, where SI is the same for both face and object, we believe the SF profile is associated with the category selectivity, otherwise we would have the same face/object recall in all profiles, as we have in the U shape profile.

      To analyze this issue further, we calculated the number of face/object selective neurons in 70-170ms. We found 43 face-selective neurons and 36 object-selective neurons (FDR corrected p-value < 0.05). Therefore, the number of face-selective and object-selective neurons is similar. Next, we check the selectivity of the neurons within each profile. Number of face/object selective neurons is LP=13/3, HP=6/2, IU=3/9, U=14/13, and the remaining belong to the NP group. Results show higher face-selective neurons in LP and HP and a higher number of object-selective neurons in the IU class. The U class contains roughly the same number of face and object-selective neurons. This observation supports the relationship between category selectivity and profiles.

      Next, we examined the average neuron response to the face and object in each profile. The difference between the firing rate of the face and object in none of the profiles was significant (Ranksum with a significance level of 0.05). However, the rates are as follows. The average firing rate (spikes/s) of face/object is LP=36.72/28.77, HP=28.55/25.52, IU=21.55/27.25, U=38.48/36.28. While the differences are not significant, they support the relationship between profiles and categories instead of responsiveness.

      The following description is added to the results section to cover this point of view.

      “To assess whether the SF profiles distinguish category selectivity or merely evaluate the neuron's responsiveness, we quantified the number of face/non-face selective neurons in the 70-170ms time window. Our analysis shows a total of 43 face-selective neurons and 36 non-face-selective neurons (FDR-corrected p-value < 0.05). The results indicate a higher proportion of face-selective neurons in LP and HP, while a greater number of non-face-selective neurons is observed in the IU category (number of face/non-face selective neurons: LP=13/3, HP=6/2, IU=3/9). The U category exhibits a roughly equal distribution of face and non-face-selective neurons (U=14/13). This finding reinforces the connection between category selectivity and the identified profiles. We then analyzed the average neuron response to faces and non-faces within each profile. The difference between the firing rates for faces and non-faces in none of the profiles is significant (face/non-face average firing rate (Hz): LP=36.72/28.77, HP=28.55/25.52, IU=21.55/27.25, U=38.48/36.28, Ranksum with significance level of 0.05). Although the observed differences are not statistically significant, they provide support for the association between profiles and categories rather than mere responsiveness.”

      About the low number of stimuli, please check the next comment.

      (15) For the category decoding, the authors employed intact, unscrambled stimuli. Were these from the main test? If yes, then I am concerned that this represents a too small number of stimuli to assess category selectivity. Only 9 fixed + 6 variable stimuli = 15 were in the main test. How many faces/ objects on average? Was the number of stimuli per category equated for the classification? When possible use the data of the preceding selectivity test which has many more stimuli to compute the category selectivity.

      We used only the main phase recorded data, which contains 15 images in each session. Each image results in 12 stimuli (intact, R1-R5, and phase-scrambled version). Thus, there exists a total of 180 unique stimuli in each session. Increasing the number of images would have increased the recording time. We compensated for this limitation by increasing the diversity of images in each session by picking the most responsive ones from the selectivity phase. On average, 7.54 of the stimuli were face in each session. We added this information to the Methods section. Furthermore, as mentioned in the discussion, for each classification run, the number of samples per category is equalized. We note that we cannot use the selectivity data for analysis, since the SF-related stimuli are filtered in different bands.

      Recommendations For The Authors:

      Reviewer #1 (Recommendations For The Authors):

      I suggest that the authors double-check their results by performing control experiments with longer stimulus duration and SF-spectrum-matched face and object stimuli.

      Thanks for your suggestion, according to your comment, we added Appendix 1 - Figure 3.

      In addition, I had a very difficult time understanding the differences between Figure 3c and Figure 4a. Please rewrite the descriptions to clarify.

      Thanks for your suggestion, we tried to revise the description of these two figures. The following description is added to the results section for Figure 3c.

      “Next, to examine the relation between the SF (category) coding capacity of the single neurons and the category (SF) coding capability of the population level, we calculated the correlation between coding performance at the population level and the coding performance of single neurons within that population (Figure 3 c and d). In other words, we investigated the relation between single and population levels of coding capabilities between SF and category. The SF (or category) coding performance of a sub-population of 20 neurons that have roughly the same single-level coding capability of the category (or SF) is examined.”

      Lines 147-148: The text states that 'The maximum accuracy of a single neuron was 19.08% higher than the chance level'. However, in Figure 4, the decoding accuracies of individual neurons for category and SF range were between 49%-90% and 20%-40%, respectively.

      Please explain the discrepancies.

      The first number is reported according to chance level which is 20%, thus the unnormalized number is 39% which is consistent with the SF accuracy in Figure 4. We added the following description to prevent any misunderstanding.

      “… was 19.08\% higher than the chance level (unnormalized accuracy is 49.08\%, neuron \#193, M2).”

      Lines 264-265: Should 'the alternative for R3 and R4' be 'the alternative for R4 and R5'?

      Thanks for your attention, it's “R4 and R5”. We corrected that mistake.

      Lines 551-562: The labels for SF classification are R1-R5. Is it a binary or a multi-classification task?

      It’s a multi-label classification. We made it clear in the text.

      “… labels were SF bands (R1, R2, ..., R5, multi-label classifier).”

      Figure 4b: Neurons in SF/category decoding exhibit both positive and negative weights. However, in the analysis of sparse neuron weights in Equation 6, only the magnitude of the weights is considered. Is the sign of weight considered too?

      We used the absolute value of the neuron weight to calculate sparseness. We also corrected Equation 6.

      Reviewer #2 (Recommendations For The Authors):

      (1) Line 52: what do the authors mean by coordinate processing in object recognition?

      To avoid any potential misunderstanding, we used the exact phrase in Saneyoshi and Michimata (2015). It is in fact, coordinate relations processing. Coordinate relations specify the metric information of the relative locations of objects.

      (2) About half of the Introduction is a summary of the Results. This can be shortened.

      Thanks for your suggestion.

      (3) Line 134: Peristimulus time histogram instead of Prestimulus time histogram.

      Thanks for your attention. We corrected that.

      (4) Line 162: the authors state that R1 is decoded faster than R5, but the reported statistic is only for R1 versus R2.

      It was a typo, the p-value is only reported for R1 and R5.

      (5) Line 576: which test was used for the asses the statistical significance?

      The test is Wilcoxon signed-rank. We added it to the text.

      (6) How can one present a 35 ms long stimulus with a 60 Hz frame rate (the stimuli were presented on a 60Hz monitor (line 470))? Please correct.

      Thanks for your attention. We corrected that. The time of stimulus presentation is 33ms and the monitor rate is 120Hz.

    1. Reviewer #3 (Public Review):

      [Editors' note: This review contains many criticisms that apply to the whole sub-field of slow/fast gamma oscillations in the hippocampus, as opposed to this particular paper. In the editors' view, these comments are beyond the scope of any single paper. However, they represent a view that, if true, should contextualise the interpretation of this paper and all papers in the sub-field. In doing so, they highlight an ongoing debate within the broader field.]

      Summary:

      The authors aimed to elucidate the role of dynamic gamma modulation in the development of hippocampal theta sequences, utilizing the traditional framework of "two gammas," a slow and a fast rhythm. This framework is currently being challenged, necessitating further analyses to establish and secure the assumed premises before substantiating the claims made in the present article.

      The results are too preliminary and need to integrate contemporary literature. New analyses are required to address these concerns. However, by addressing these issues, it may be possible to produce an impactful manuscript.

      I. Introduction<br /> Within the introduction, multiple broad assertions are conveyed that serve as the premise for the research. However, equally important citations that are not mentioned potentially contradict the ideas that serve as the foundation. Instances of these are described below:

      (1) Are there multiple gammas? The authors launched the study on the premise that two different gamma bands are communicated from CA3 and the entorhinal cortex. However, recent literature suggests otherwise, offering that the slow gamma component may be related to theta harmonics:

      From a review by Etter, Carmichael and Williams (2023)<br /> "Gamma-based coherence has been a prominent model for communication across the hippocampal-entorhinal circuit and has classically focused on slow and fast gamma oscillations originating in CA3 and medial entorhinal cortex, respectively. These two distinct gammas are then hypothesized to be integrated into hippocampal CA1 with theta oscillations on a cycle-to-cycle basis (Colgin et al., 2009; Schomburg et al., 2014). This would suggest that theta oscillations in CA1 could serve to partition temporal windows that enable the integration of inputs from these upstream regions using alternating gamma waves (Vinck et al., 2023). However, these models have largely been based on correlations between shifting CA3 and medial entorhinal cortex to CA1 coherence in theta and gamma bands. In vivo, excitatory inputs from the entorhinal cortex to the dentate gyrus are most coherent in the theta band, while gamma oscillations would be generated locally from presumed local inhibitory inputs (Pernía-Andrade and Jonas, 2014). This predominance of theta over gamma coherence has also been reported between hippocampal CA1 and the medial entorhinal cortex (Zhou et al., 2022). Another potential pitfall in the communication-through-coherence hypothesis is that theta oscillations harmonics could overlap with higher frequency bands (Czurkó et al., 1999; Terrazas et al., 2005), including slow gamma (Petersen and Buzsáki, 2020). The asymmetry of theta oscillations (Belluscio et al., 2012) can lead to harmonics that extend into the slow gamma range (Scheffer-Teixeira and Tort, 2016), which may lead to a misattribution as to the origin of slow-gamma coherence and the degree of spike modulation in the gamma range during movement (Zhou et al., 2019)."

      And from Benjamin Griffiths and Ole Jensen (2023)<br /> "That said, in both rodent and human studies, measurements of 'slow' gamma oscillations may be susceptible to distortion by theta harmonics [53], meaning open questions remain about what can be attributed to 'slow' gamma oscillations and what is attributable to theta."

      This second statement should be heavily considered as it is from one of the original authors who reported the existence of slow gamma.

      Yet another instance from Schomburg, Fernández-Ruiz, Mizuseki, Berényi, Anastassiou, Christof Koch, and Buzsáki (2014):<br /> "Note that modulation from 20-30 Hz may not be related to gamma activity but, instead, reflect timing relationships with non-sinusoidal features of theta waves (Belluscio et al., 2012) and/or the 3rd theta harmonic."

      One of this manuscript's authors is Fernández-Ruiz, a contemporary proponent of the multiple gamma theory. Thus, the modulation to slow gamma offered in the present manuscript may actually be related to theta harmonics.

      With the above emphasis from proponents of the slow/fast gamma theory on disambiguating harmonics from slow gamma, our first suggestion to the authors is that they A) address these statements (citing the work of these authors in their manuscript) and B) demonstrably quantify theta harmonics in relation to slow gamma prior to making assertions of phase relationships (methodological suggestions below). As the frequency of theta harmonics can extend as high as 56 Hz (PMID: 32297752), overlapping with the slow gamma range defined here (25-45 Hz), it will be important to establish an approach that decouples the two phenomena using an approach other than an arbitrary frequency boundary.

      (2) Can gammas be segregated into different lamina of the hippocampus? This idea appears to be foundational in the premise of the research but is also undergoing revision.

      As discussed by Etter et al. above, the initial theory of gamma routing was launched on coherence values. However, the values reported by Colgin et al. (2009) lean more towards incoherence (a value of 0) rather than coherence (1), suggesting a weak to negligible interaction. Nevertheless, this theory is coupled with the idea that the different gamma frequencies are exclusive to the specific lamina of the hippocampus.

      Recently, Deschamps et al. (2024) suggested a broader, more nuanced understanding of gamma oscillations than previously thought, emphasizing their wide range and variability across hippocampal layers. This perspective challenges the traditional dichotomy of gamma sub-bands (e.g., slow vs. medium gamma) and their associated cognitive functions based on a more rigid classification according to frequency and phase relative to the theta rhythm. Moreover, they observed all frequencies across all layers.

      Similarly, the current source density plots from Belluscio et al. (2012) suggest that SG and FG can be observed in both the radiatum and lacunosum-moleculare.

      Therefore, if the initial coherence values are weak to negligible and both slow and fast gamma are observed in all layers of the hippocampus, can the different gammas be exclusively related to either anatomical inputs or psychological functions (as done in the present manuscript)? Do these observations challenge the authors' premise of their research? At the least, please discuss.

      (3) Do place cells, phase precession, and theta sequences require input from afferent regions? It is offered in the introduction that "Fast gamma (~65-100Hz), associated with the input from the medial entorhinal cortex, is thought to rapidly encode ongoing novel information in the context (Fernandez-Ruiz et al., 2021; Kemere, Carr, Karlsson, & Frank, 2013; Zheng et al., 2016)".

      CA1 place fields remain fairly intact following MEC inactivation include Ipshita Zutshi, Manuel Valero, Antonio Fernández-Ruiz , and György Buzsáki (2022)- "CA1 place cells and assemblies persist despite combined mEC and CA3 silencing" and from Hadas E Sloin, Lidor Spivak, Amir Levi, Roni Gattegno, Shirly Someck, Eran Stark (2024) - "These findings are incompatible with precession models based on inheritance, dual-input, spreading activation, inhibition-excitation summation, or somato-dendritic competition. Thus, a precession generator resides locally within CA1."

      These publications, at the least, challenge the inheritance model by which the afferent input controls CA1 place field spike timing. The research premise offered by the authors is couched in the logic of inheritance, when the effect that the authors are observing could be governed by local intrinsic activity (e.g., phase precession and gamma are locally generated, and the attribution to routed input is perhaps erroneous). Certainly, it is worth discussing these manuscripts in the context of the present manuscript.

      II. Results

      (1) Figure 2-<br /> a. There is a bit of a puzzle here that should be discussed. If slow and fast frequencies modulate 25% of neurons, how can these rhythms serve as mechanisms of communication/support psychological functions? For instance, if fast gamma is engaged in rapid encoding (line 72) and slow gamma is related to the integration processing of learned information (line 84), and these are functions of the hippocampus, then why do these rhythms modulate so few cells? Is this to say 75% of CA1 neurons do not listen to CA3 or MEC input?

      b. Figure 2. It is hard to know if the mean vector lengths presented are large or small. Moreover, one can expect to find significance due to chance. For instance, it is challenging to find a frequency in which modulation strength is zero (please see Figure 4 of PMID: 30428340 or Figure 7 of PMID: 31324673).

      i. Please construct the histograms of Mean Vector Length as in the above papers, using 1 Hz filter steps from 1-120Hz and include it as part of Figure 2 (i.e., calculate the mean vector length for the filtered LFP in steps of 1-2 Hz, 2-3 Hz, 3-4 Hz,... etc). This should help the authors portray the amount of modulation these neurons have relative to the theta rhythm and other frequencies. If the theta mean vector length is higher, should it be considered the primary modulatory influence of these neurons (with slow and fast gammas as a minor influence)?

      ii. It is possible to infer a neuron's degree of oscillatory modulation without using the LFP. For instance, one can create an ISI histogram as done in Figure 1 here (https://www.biorxiv.org/content/10.1101/2021.09.20.461152v3.full.pdf+html; "Distinct ground state and activated state modes of firing in forebrain neurons"). The reciprocal of the ISI values would be "instantaneous spike frequency". In favor of the Douchamps et al. (2024) results, the figure of the BioRXiV paper implies that there is a single gamma frequency modulate as there is only a single bump in the ISIs in the 10^-1.5 to 10^-2 range. Therefore, to vet the slow gamma results and the premise of two gammas offered in the introduction, it would be worth including this analysis as part of Figure 2.

      c. There are some things generally concerning about Figure 2.

      i. First, the raw trace does not seem to have clear theta epochs (it is challenging to ascertain the start and end of a theta cycle). Certainly, it would be worth highlighting the relationship between theta and the gammas and picking a nice theta epoch.

      ii. Also, in panel A, there looks to be a declining amplitude relationship between the raw, fast, and slow gamma traces, assuming that the scale bars represent 100uV in all three traces. The raw trace is significantly larger than the fast gamma. However, this relationship does not seem to be the case in panel B (in which both the raw and unfiltered examples of slow and fast gamma appear to be equal; the right panels of B suggest that fast gamma is larger than slow, appearing to contradict the A= 1/f organization of the power spectral density). Please explain as to why this occurs. Including the power spectral density (see below) should resolve some of this.

      iii. Within the example of spiking to phase in the left side of Panel B (fast gamma example)- the neuron appears to fire near the trough twice, near the peak twice, and somewhere in between once. A similar relationship is observed for the slow gamma epoch. One would conclude from these plots that the interaction of the neuron with the two rhythms is the same. However, the mean vector lengths and histograms below these plots suggest a different story in which the neuron is modulated by FG but not SG. Please reconcile this.

      iv. For calculating the MVL, it seems that the number of spikes that the neuron fires would play a significant role. Working towards our next point, there may be a bias of finding a relationship if there are too few spikes (spurious clustering due to sparse data) and/or higher coupling values for higher firing rate cells (cells with higher firing rates will clearly show a relationship), forming a sort of inverse Yerkes-Dodson curve. Also, without understanding the magnitude of the MVL relative to other frequencies, it may be that these values are indeed larger than zero, but not biologically significant.

      - Please provide a scatter plot of Neuron MVL versus the Neuron's Firing Rate for 1) theta (7-9 Hz), 2) slow gamma, and 3) fast gamma, along with their line of best fit.

      - Please run a shuffle control where the LFP trace is shifted by random values between 125-1000ms and recalculate the MVL for theta, slow, and fast gamma. Often, these shuffle controls are done between 100-1000 times (see cross-correlation analyses of Fujisawa, Buzsaki et al.).

      - To establish that firing rate does not play a role in uncovering modulation, it would be worth conducting a spike number control, reducing the number of spikes per cell so that they are all equal before calculating the phase plots/MVL.

      (2) Something that I anticipated to see addressed in the manuscript was the study from Grosmark and Buzsaki (2016): "Cell assembly sequences during learning are "replayed" during hippocampal ripples and contribute to the consolidation of episodic memories. However, neuronal sequences may also reflect preexisting dynamics. We report that sequences of place-cell firing in a novel environment are formed from a combination of the contributions of a rigid, predominantly fast-firing subset of pyramidal neurons with low spatial specificity and limited change across sleep-experience-sleep and a slow-firing plastic subset. Slow-firing cells, rather than fast-firing cells, gained high place specificity during exploration, elevated their association with ripples, and showed increased bursting and temporal coactivation during postexperience sleep. Thus, slow- and fast-firing neurons, although forming a continuous distribution, have different coding and plastic properties."

      My concern is that much of the reported results in the present manuscript appear to recapitulate the observations of Grosmark and Buzsaki, but without accounting for differences in firing rate. A parsimonious alternative explanation for what is observed in the present manuscript is that high firing rate neurons, more integrated into the local network and orchestrating local gamma activity (PING), exhibit more coupling to theta and gamma. In this alternative perspective, it's not something special about how the neurons are entrained to the routed fast gamma, but that the higher firing rate neurons are better able to engage and entrain their local interneurons and, thus modulate local gamma. However, this interpretation challenges the discussion around the importance of fast gamma routed from the MEC.

      a. Please integrate the Grosmark & Buzsaki paper into the discussion.

      b. Also, please provide data that refutes or supports the alternative hypothesis in which the high firing rate cells are just more gamma modulated as they orchestrate local gamma activity through monosynaptic connections with local interneurons (e.g., Marshall et al., 2002, Hippocampal pyramidal cell-interneuron spike transmission is frequency dependent and responsible for place modulation of interneuron discharge). Otherwise, the attribution to a MEC routed fast gamma routing seems tenuous.<br /> c. It is mentioned that fast-spiking interneurons were removed from the analysis. It would be worth including these cells, calculating the MVL in 1 Hz increments as well as the reciprocal of their ISIs (described above).

      (3) Methods - Spectral decomposition and Theta Harmonics.

      a. It is challenging to interpret the exact parameters that the authors used for their multi-taper analysis in the methods (lines 516-526). Tallon-Baudry et al., (1997; Oscillatory γ-Band (30-70 Hz) Activity Induced by a Visual Search Task in Humans) discuss a time-frequency trade-off where frequency resolution changes with different temporal windows of analysis. This trade-off between time and frequency resolution is well known as the uncertainty principle of signal analysis, transcending all decomposition methods. It is not only a function of wavelet or FFT, and multi-tapers do not directly address this. (The multitaper method, by using multiple specially designed tapers -like the Slepian sequences- smooths the spectrum. This smoothing doesn't eliminate leakage but distributes its impact across multiple estimates). Given the brevity of methods and the issues of theta harmonics as offered above, it is worth including some benchmark trace testing for the multi-taper as part of the supplemental figures.

      i. Please spectrally decompose an asymmetric 8 Hz sawtooth wave showing the trace and the related power spectral density using the multiple taper method discussed in the methods.

      ii. Please also do the same for an elliptical oscillation (perfectly symmetrical waves, but also capable of casting harmonics). Matlab code on how to generate this time series is provided below:<br /> A = 1; % Amplitude<br /> T = 1/8; % Period corresponding to 8 Hz frequency<br /> omega = 2*pi/T; % Angular frequency<br /> C = 1; % Wave speed<br /> m = 0.9; % Modulus for the elliptic function (0<br /> x = linspace(0, 2*pi, 1000); % temporal domain<br /> t = 0; % Time instant

      % Calculate B based on frequency and speed<br /> B = sqrt(omega/C);

      % Cnoidal wave equation using the Jacobi elliptic function<br /> u = A .* ellipj(B.*(x - C*t), m).^2;

      % Plotting the cnoidal wave<br /> figure;<br /> plot(x./max(x), u);<br /> title('8 Hz Cnoidal Wave');<br /> xlabel('time (x)');<br /> ylabel('Wave amplitude (u)');<br /> grid on;

      The Symbolic Math Toolbox needs to be installed and accessible in your MATLAB environment to use ellipj. Otherwise, I trust that, rather than plotting a periodic orbit around a circle (sin wave) the authors can trace the movement around an ellipse with significant eccentricity (the distance between the two foci should be twice the distance between the co-vertices).

      iii. Line 522: "The power spectra across running speeds and absolute power spectrum (both results were not shown)...". Given the potential complications of multi-taper discussed above, and as each convolution further removes one from the raw data, it would be the most transparent, simple, and straightforward to provide power spectra using the simple fft.m code in Matlab (We imagine that the authors will agree that the results should be robust against different spectral decomposition methods. Otherwise, it is concerning that the results depend on the algorithm implemented and should be discussed. If gamma transience is a concern, the authors should trigger to 2-second epochs in which slow/fast gamma exceeds 3-7 std. dev. above the mean, comparing those resulting power spectra to 2-second epochs with ripples - also a transient event). The time series should be at least 2 seconds in length (to avoid spectral leakage issues and the issues discussed in Talon-Baudry et al., 1997 above).

      Please show the unmolested power spectra (Y-axis units in mV2/Hz, X-axis units as Hz) as a function of running speed (increments of 5 cm/s) for each animal. I imagine three of these PSDs for 3 of the animals will appear in supplemental methods while one will serve as a nice manuscript figure. With this plot, please highlight the regions that the authors are describing as theta, slow, and fast gamma. Also, any issues should be addressed should there be notable differences in power across animals or tetrodes (issues with locations along proximal-distal CA1 in terms of MEC/LEC input and using a local reference electrode are discussed below).

      iv. Schomberg and colleagues (2014) suggested that the modulation of neurons in the slow gamma range could be related to theta harmonics (see above). Harmonics can often extend in a near infinite as they regress into the 1/f background (contributing to power, but without a peak above the power spectral density slope), making arbitrary frequency limits inappropriate. Therefore, in order to support the analyses and assertions regarding slow gamma, it seems necessary to calculate a "theta harmonic/slow gamma ratio". Aru et al. (2015; Untangling cross-frequency coupling in neuroscience) offer that: " The presence of harmonics in the signal should be tested by a bicoherence analysis and its contribution to CFC should be discussed." Please test both the synthetic signals above and the raw LFP, using temporal windows of greater than 4 seconds (again, the large window optimizes for frequency resolution in the time-frequency trade-off) to calculate the bicoherence. As harmonics are integers of theta coupled to itself and slow gamma is also coupled to theta, a nice illustration and contribution to the field would be a method that uses the bispectrum to isolate and create a "slow gamma/harmonic" ratio.

      (4) I appreciate the inclusion of the histology for the 4 animals. Knerim and colleagues describe a difference in MEC projection along the proximal-distal axis of the CA1 region (https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3866456/)- "There are also differences in their direct projections along the transverse axis of CA1, as the LEC innervates the region of CA1 closer to the subiculum (distal CA1), whereas the MEC innervates the region of CA1 closer to CA2 and CA3 (proximal CA1)" From the histology, it looks like some of the electrodes are in the part of CA1 that would be dominated by LEC input while a few are closer to where the MEC would project.

      a. How do the authors control for these differences in projections? Wouldn't this change whether or not fast gamma is observed in CA1?

      b. I am only aware of one manuscript that describes slow gamma in the LEC which appeared in contrast to fast gamma from the MEC (https://www.science.org/doi/10.1126/science.abf3119). One would surmise that the authors in the present manuscript would have varying levels of fast gamma in their CA1 recordings depending on the location of the electrodes in the Proximal-distal axis, to the extent that some of the more medial tetrodes may need to be excluded (as they should not have fast gamma, rather they should be exclusively dominated by slow gamma). Alternatively, the authors may find that there is equal fast gamma power across the entire proximal-distal axis. However, this would pose a significant challenge to the LEC/slow gamma and MEC/fast gamma routing story of Fernandez-Ruiz et al. and require reconciliation/discussion.

      c. Is there a difference in neuron modulation to these frequencies based on electrode location in CA1?

      (5) Given a comment in the discussion (see below), it will be worth exploring changes in theta, theta harmonic, slow gamma, and fast gamma power with running speed as no changes were observed with theta sequences or lap number versus. Notably, Czurko et al., report an increase in theta and harmonic power with running speed (1999) while Ahmed and Mehta (2012) report a similar effect for gamma.

      a. Please determine if the oscillations change in power and frequency of the rhythms discussed above change with running speed using the same parameters applied in the present manuscript. The specific concern is that how the authors calculate running speed is not sensitive enough to evaluate changes.

      b. It is astounding that animals ran as fast as they did in what appears to be the first lap (Figure 3F), especially as rats' natural proclivity is thigmotaxis and inquisitive exploration in novel environments. Can the authors expand on why they believe their rats ran so quickly on the first lap in a novel environment and how to replicate this? Also, please include the individual values for each animal on the same plot.

      c. Can the authors explain how the statistics on line 169 (F(4,44)) work? Specifically, it is challenging to determine how the degrees of freedom were calculated in this case and throughout if there were only 4 animals (reported in methods) over 5 laps (depicted in Figure 3F. Given line 439, it looks like trials and laps are used synonymously). Four animals over 5 laps should have a DOF of 16.

      (6) Throughout the manuscript, I am concerned about an inflation of statistical power. For example on line 162, F(2,4844). The large degrees of freedom indicate that the sample size was theta sequences or a number of cells. Since multiple observations were obtained from the same animal, the statistical assumption of independence is violated. Therefore, the stats need to be conducted using a nested model as described in Aarts et al. (2014; https://pubmed.ncbi.nlm.nih.gov/24671065/). A statistical consult may be warranted.

      (7) It is stated that one tetrode served as a quiet recording reference. The "quiet" part is an assumption when often, theta and gamma can be volume conducted to the cortex (e.g., Sirota et al., 2008; This is often why laboratories that study hippocampal rhythms use the cerebellum for the differential recording electrode and not an electrode in the corpus callosum). Generally, high frequencies propagate as well as low frequencies in the extracellular milieu (https://www.eneuro.org/content/4/1/ENEURO.0291-16.2016). For transparency, the authors should include a limitation paragraph in their discussion that describes how their local tetrode reference may be inadvertently diminishing and/or distorting the signal that they are trying to isolate. Otherwise, it would be worth hearing an explanation as to how the author's approach avoids this issue.

      Apologetically, this review is already getting long. Moreover, I have substantial concerns that should be resolved prior to delving into the remainder of the analyses. e.g., the analyses related to Figure 3-5 assert that FG cells are important for sequences. However, the relationship to gamma may be secondary to either their relationship to theta or, based on the Grosmark and Buzsaki paper, it may just be a phenomenon coupled to the fast-firing cells (fast-firing cells showing higher gamma modulation due to a local PING dynamic). Moreover, the observation of slow gamma is being challenged as theta harmonics, even by the major proponents of the slow/fast gamma theory. Therefore, the report of slow gamma precession would come as an unsurprising extension should they be revealed to be theta harmonics (however, no control for harmonics was implemented; suggestions were made above). Following these amendments, I would be grateful for the opportunity to provide further feedback.

      III. Discussion.

      a. Line 330- it was offered that fast gamma encodes information while slow gamma integrates in the introduction. However, in a task such as circular track running (from the methods, it appears that there is no new information to be acquired within a trial), one would guess that after the first few laps, slow gamma would be the dominant rhythm. Therefore, one must wonder why there are so few neurons modulated by slow gamma (~3.7%).

      b. Line 375: The authors contend that: "...slow gamma, related to information compression, was also required to modulate fast gamma phase-locked cells during sequence development. We replicated the results of slow gamma phase precession at the ensemble level (Zheng et al., 2016), and furthermore observed it at late development, but not early development, of theta sequences." In relation to the idea that slow gamma may be coupled to - if not a distorted representation of - theta harmonics, it has been observed that there are changes in theta relative to novelty.

      i. A. Jeewajee, C. Lever, S. Burton, J. O'Keefe, and N. Burgess (2008) report a decrease in theta frequency in novel circumstances that disappears with increasing familiarity.

      ii. One could surmise that this change in frequency is associated with alterations in theta harmonics (observed here as slow gamma), challenging the author's interpretation.

      iii. Therefore, the authors have a compelling opportunity to replicate the results of Jeewajee et al., characterizing changes of theta along with the development of slow gamma precession, as the environment becomes familiar. It will become important to demonstrate, using bicoherence as offered by Aru et al., how slow gamma can be disambiguated from theta harmonics. Specifically, we anticipate that the authors will be able to quantify A) theta harmonics (the number, and their respective frequencies and amplitudes), B) the frequency and amplitude of slow gamma, and C) how they can be quantitatively decoupled. Through this, their discussion of oscillatory changes with novelty-familiarity will garner a significant impact.

      c. Broadly, it is interesting that the authors emphasize the gamma frequency throughout the discussion. Given that the power spectral density of the Local Field Potential (LFP) exhibits a log-log relationship between amplitude and frequency, as described by Buzsáki (2005) in "Rhythms of the Brain," and considering that the LFP is primarily generated through synaptic transmembrane currents (Buzsáki et al., 2012), it seems parsimonious to consider that the bulk of synaptic activity occurs at lower frequencies (e.g., theta). Since synaptic transmission represents the most direct form of inter-regional communication, one might wonder why gamma (characterized by lower amplitude rhythms) is esteemed so highly compared to the higher amplitude theta rhythm. Why isn't the theta rhythm, instead, regarded as the primary mode of communication across brain regions? A discussion exploring this question would be beneficial.

    1. Reviewer #2 (Public Review):

      Summary:

      The authors investigate the mechanisms supporting learning to suppress distractors at predictable locations, focusing on proactive suppression mechanisms manifesting before the onset of a distractor. They used EEG and inverted encoding models (IEM). The experimental paradigm alternates between a visual search task and a spatial memory task, followed by a placeholder screen acting as a 'ping' stimulus -i.e., a stimulus to reveal how learned distractor suppression affects hidden priority maps. Behaviorally, their results align with the effects of statistical learning on distractor suppression. Contrary to the proactive suppression hypothesis, which predicts reduced memory-specific tuning of neural representations at the expected distractor location, their IEM results indicate increased tuning at the high-probability distractor location following the placeholder and prior to the onset of the search display.

      Strengths:

      Overall, the manuscript is well-written and clear, and the research question is relevant and timely, given the ongoing debate on the roles of proactive and reactive components in distractor processing. The use of a secondary task and EEG/IEM to provide a direct assessment of hidden priority maps in anticipation of a distractor is, in principle, a clever approach. The study also provides behavioral results supporting prior literature on distractor suppression at high-probability locations.

      Weaknesses:

      (1) At a conceptual level, I understand the debate and opposing views, but I wonder whether it might be more comprehensive to present also the possibility that both proactive and reactive stages contribute to distractor suppression. For instance, anticipatory mechanisms (proactive) may involve expectations and signals that anticipate the expected distractor features, whereas reactive mechanisms contribute to the suppression and disengagement of attention.

      (2) The authors focus on hidden priority maps in pre-distractor time windows, arguing that the results challenge a simple proactive view of distractor suppression. However, they do not provide evidence that reactive mechanisms are at play or related to the pinging effects found in the present paradigm. Is there a relationship between the tuning strength of CTF at the high-probability distractor location and the actual ability to suppress the distractor (e.g., behavioral performance)? Is there a relationship between CTF tuning and post-distractor ERP measures of distractor processing? While these may not be the original research questions, they emerge naturally and I believe should be discussed or noted as limitations.

      (3) How do the authors ensure that the increased tuning (which appears more as a half-split or hemifield effect rather than gradual fine-grained tuning, as shown in Figure 5) is not a byproduct of the dual-task paradigm used, rather than a general characteristic of learned attentional suppression? For example, the additional memory task and the repeated experience with the high-probability distractor at the specific location might have led to longer-lasting and more finely-tuned traces for memory items at that location compared to others.

      (4) It is unclear how IEM was performed on total vs. evoked power, compared to typical approaches of running it on single trials or pseudo-trials.

      (5) Following on point 1. What is the rationale for relating decreased (but not increased) tuning of CTF to proactive suppression? Could it be that proactive suppression requires anticipatory tuning towards the expected feature to implement suppression? In other terms, better 'tuning' does not necessarily imply a higher signal amplitude and could be observable even under signal suppression. The authors should comment on this and clarify.

      Minor:

      (1) In the Word file I reviewed, there are minor formatting issues, such as missing spaces, which should be double-checked.

      (2) Would the authors predict that proactive mechanisms are not involved in other forms of attention learning involving distractor suppression, such as habituation?

      (3) A clear description in the Methods section of how individual CTFs for each location were derived would help in understanding the procedure.

      (4) Why specifically 1024 resampling iterations?

    1. path.read_text()

      ここまではエラーにならなかったのですが、Windows環境ではここで

      UnicodeDecodeError: 'cp932' codec can't decode byte 0x99 in position 205170: illegal multibyte sequence

      というエラーが出ました。

      path.read_text(encoding='utf-8')

      と明示すればエラーは回避できました。訳注つけますか?

    1. You can also use Bash in the Azure Cloud Shell to connect to your VM. You can use Cloud Shell in a web browser, from the Azure portal, or as a terminal in Visual Studio Code using the Azure Account extension. You can also install the Windows Subsystem for Linux to connect to your VM over SSH and use other native Linux tools within a Bash shell.

      the Azure portal streamlines the process so there is minimal overhead during deployment

    1. Reviewer #1 (Public Review):

      Summary:

      Sammons, Masserini et al. examine the connectivity of different types of CA3 pyramidal cells ("thorny" and "athorny"), and how their connectivity putatively contributes to their relative timing in sharp-wave-like activity. First, using patch-clamp recordings, they characterize the degree of connectivity within and between athorny and thorny cells. Based upon these experimental results, they compute a synaptic product matrix, and use this to inform a computational model of CA3 activity. This model finds that this differential connectivity between these populations, augmented by two different types of inhibitory neurons, can account for the relative timing of activity observed in sharp waves in vivo.

      Strengths:

      The patch-clamp experiments are exceptionally thorough and well done. These are very challenging experiments and the authors should be commended for their in-depth characterization of CA3 connectivity.

      Weaknesses:

      (1) The computational elements of this study feel underdeveloped. Whereas the authors do a thorough job experimentally characterizing connections between excitatory neurons, the inhibitory neurons used in the model seem to be effectivity "fit neurons" and appear to have been tuned to produce the emergent properties of CA3 sharp wave-like activity. Although I appreciate the goal was to implicate CA3 connectivity contributions to activity timing, a stronger relationship seems like it could be examined. For example, did the authors try to "break" their model? It would be informative if they attempted different synaptic product matrices (say, the juxtaposition of their experimental product matrix) and see whether experimentally-derived sequential activity could not be elicited. It seems as though this spirit of analysis was examined in Figure 4C, but only insofar as individual connectivity parameters were changed in isolation.

      (2) Additional explanations of how parameters for interneurons were incorporated in the model would be very helpful. As it stands, it is difficult to understand the degree to which the parameters of these neurons are biologically constrained versus used as fit parameters to produce different time windows of activity in types of CA3 pyramidal cells.

    2. Reviewer #3 (Public Review):

      Summary:

      The hippocampal CA3 region is generally considered to be the primary site of initiation of sharp wave ripples-highly synchronous population events involved in learning and memory although the precise mechanism remains elusive. A recent study revealed that CA3 comprises two distinct pyramidal cell populations: thorny cells that receive mossy fiber input from the dentate gyrus, and athorny cells that do not. That study also showed that it is athorny cells in particular that play a key role in sharp wave initiation. In the present work, Sammons, Masserini, and colleagues expand on this by examining the connectivity probabilities among and between thorny and athorny cells. First, using whole-cell patch clamp recordings, they find an asymmetrical connectivity pattern, with athorny cells receiving the most synaptic connections from both athorny and thorny cells, and thorny cells receiving fewer. They then demonstrate in spiking neural network simulations how this asymmetrical connectivity may underlie the preferential role of athorny cells in sharp wave initiation.

      Strengths:

      The authors provide independent validation of some of the findings by Hunt et al. (2018) concerning the distinction between thorny and athorny pyramidal cells in CA3 and advance our understanding of their differential integration in CA3 microcircuits. The properties of excitatory connections among and between thorny and athorny cells described by the authors will be key in understanding CA3 functions including, but not limited to, sharp wave initiation.

      As stated in the paper, the modeling results lend support to the idea that the increased excitatory connectivity towards athorny cells plays a key role in causing them to fire before thorny cells in sharp waves. More generally, the model adds to an expanding pool of models of sharp wave ripples which should prove useful in guiding and interpreting experimental research.

      Weaknesses:

      The mechanism by which athorny cells initiate sharp waves in the model is somewhat confusingly described. As far as I understood, random fluctuations in the activities of A and B neurons provide windows of opportunity for pyramidal cells to fire if they have additionally recovered from adaptive currents. Thorny and athorny pyramidal cells are then set in a winner-takes-all competition which is quickly won by the athorny cells. The main thesis of the paper seems to be that athorny cells win this competition because they receive more inputs both from themselves and from thorny cells, hence, the connectivity "underlies the sequential activation". However, it is also stated that athorny cells activate first due to their lower rheobase and steeper f-I curve, and it is also indicated in the methods that athorny (but not thorny) cells fire in bursts. It seems that it is primarily these features that make them fire first, something which apparently happens even when the A to A connectivity is set to 0-albeit with a very small lag. Perhaps the authors could further clarify the differential role of single cell and network parameters in determining the sequential activation of athorny and thorny cells. Is the role of asymmetric excitatory connectivity only to enhance the initial intrinsic advantage of athorny cells? If so, could this advantage also be enhanced in other ways?

      Although a clear effort has been made to constrain the model with biological data, too many degrees of freedom remain that allow the modeler to make arbitrary decisions. This is not a problem in itself, but perhaps the authors could explain more of their reasoning and expand upon the differences between their modeling choices and those of others. For example, what are the conceptual or practical advantages of using adaptation in pyramidal neurons as opposed to short-term synaptic plasticity as in the model by Hunt et al.? Relatedly, what experimental observations could validate or falsify the proposed mechanisms?

      In the data by Hunt et al., thorny cells have a higher baseline (non-SPW) firing rate, and it is claimed that it is actually stochastic correlations in their firing that are amplified by athorny cells to initiate sharp waves. However, in the current model, the firing of both types of pyramidal cells outside of ripples appears to be essentially zero. Can the model handle more realistic firing rates as described by Hunt et al., or as produced by e.g., walking around an environment tiled with place cells, or would that trigger SPWs continuously?

    1. Reviewer #2 (Public Review):

      In this study, Leighton et al performed remarkable experiments by combining in-vivo patch-clamp recording with two-photon dendritic Ca2+ imaging. The voltage-clamp mode is a major improvement over the pioneer versions of this combinatorial experiment that had led to major breakthroughs in the neuroscience field for visualizing and understanding synaptic input activities in single cells in-vivo (sharp electrodes: Svoboda et al, Nature 1997, Helmchen et al, Nature Neurosci 1999; whole-cell current-clamp: Jia et al, Nature 2010, Chen et al, Nature 2011. I suggest that these papers would be cited). This is because in voltage-clamp mode, despite a full control of membrane voltage in-vivo is not realistic, is nevertheless most effective in preventing back-propagation action potentials, which would severely confound the measurement of individual synaptically-induced Ca2+ influx events. Furthermore, clamping the cell body at a strongly depolarized potential (here the authors did -30mV) also facilitates the detection of synaptically-induced Ca2+ influx. As a result, the authors successfully recorded high-quality Ca2+ imaging data that can be used for precise analysis. To date, even in view of the rapid progress of voltage-sensitive indicators and relevant imaging technologies in the recent years, this very old 'art' of combining single-cell electrophysiology and two-photon imaging (ordinary, raster-scanned, video-rate imaging) of Ca2+ signals still enable measurements of the best-level precision.

      On the other hand, the interpretation of data in this study is a bit narrow-minded and lacks a comprehensive picture. Some suggestions to improve the manuscript are as follows:

      (1) The authors made a segregation of 'spine synapse' and 'shaft synapse' based solely on the two-photon images in-vivo. However, caution shall be taken here, because the optical resolution under in-vivo imaging conditions like this cannot reliably tell apart whether a bright spot within or partially overlapping a segment of dendrite is a spine on top (or below) of it. Therefore, what the authors consider as a 'shaft synapse' (by detecting Ca2+ hotspots) has an unknown probability to be just a spine on top or below the dendrite. If there is other imaging data of higher axial resolution to validate or calibrate, the authors shall take some further considerations or analysis to check the consistency of their data, as the authors do need such a segregation between spine and shaft synapses to show how they evolve over the brain development stages.<br /> (2) The use of terminology 'bursts of spontaneous inputs' for describing voltage-clamp data seems improper. Conventionally, 'burst' refers to suprathreshold spike firing events, but here, the authors use 'burst' to refer to inward synaptic currents collected at the cell body. It is obvious that not every excitatory synaptic input (or ensemble of inputs) activation will lead to spike firing under naturalistic conditions, therefore, these two concepts are not equivalent. It is recommended to use 'barrage of inputs' instead of 'burst of inputs'. Imagine a full picture of the entire dendritic tree, the fact that the authors could always capture spontaneous Ca2+ events here and there within a few pieces of dendrites within an arbitrary field-of-view suggest that, the whole dendritic tree must have many more such events going on as a barrage while the author's patch electrode picks up the summed current flow from the whole dendritic tree.<br /> (3) Following the above issue, an analysis of the temporal correlation between synaptic (not segregating 'spine' or 'shaft') Ca2+ events and EPSCs is absent. Again, the authors drew arbitrary time windows to clump the events for statistical analysis. However, the demonstrated example data already show that the onset times of individual synaptic Ca2+ events do not necessarily align with the beginning of a 'barrage' inward current event.<br /> (4) The authors claim that "these observations indicate that the activity patterns investigated here are not or only slightly affected by low-level anesthesia". It would be nice to show some of the recordings in this work without any anesthesia to support this claim.<br /> (5) I suggest the authors should provide the number of cells and mice recorded in the figure legends.<br /> (6) Instead of showing only cartoon illustrations of dendrites in Figure 3-6, I suggest showing the two-photon images as well together with the cartoon.

      The authors have addressed most of my issues, but I miss the responses to my points 5 and 6. I have no additional comments.